Zeng, Yuni, Mao, Hua, Peng, Dezhong and Yi, Zhang (2019) Spectrogram based multi-task audio classification. Multimedia Tools and Applications, 78 (3). pp. 3705-3722. ISSN 1380-7501
|
Text
Zeng et al - Spectrogram based multi-task audio classification AAM.pdf - Accepted Version Download (2MB) | Preview |
Abstract
Audio classification is regarded as a great challenge in pattern recognition. Although audio classification tasks are always treated as independent tasks, tasks are essentially related to each other such as speakers’ accent and speakers’ identification. In this paper, we propose a Deep Neural Network (DNN)-based multi-task model that exploits such relationships and deals with multiple audio classification tasks simultaneously. We term our model as the gated Residual Networks (GResNets) model since it integrates Deep Residual Networks (ResNets) with a gate mechanism, which extract better representations between tasks compared with Convolutional Neural Networks (CNNs). Specifically, two multiplied convolutional layers are used to replace two feed-forward convolution layers in the ResNets. We tested our model on multiple audio classification tasks and found that our multi-task model achieves higher accuracy than task-specific models which train the models separately.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | Multi-task learning, Convolutional neural networks, Deep residual networks, Audio classification |
Subjects: | G400 Computer Science |
Department: | Faculties > Engineering and Environment > Computer and Information Sciences |
Depositing User: | Paul Burns |
Date Deposited: | 12 Jun 2019 16:57 |
Last Modified: | 01 Aug 2021 11:31 |
URI: | http://nrl.northumbria.ac.uk/id/eprint/39657 |
Downloads
Downloads per month over past year