PANNs: Large-Scale Pretrained Audio Neural Networks for Audio Pattern Recognition

23 Aug 2020  ·  Kong Qiuqiang, Cao Yin, Iqbal Turab, Wang Yuxuan, Wang Wenwu, Plumbley Mark D. ·

Audio pattern recognition is an important research topic in the machine learning area, and includes several tasks such as audio tagging, acoustic scene classification, music classification, speech emotion classification and sound event detection. Recently, neural networks have been applied to tackle audio pattern recognition problems. However, previous systems are built on specific datasets with limited durations. Recently, in computer vision and natural language processing, systems pretrained on large-scale datasets have generalized well to several tasks. However, there is limited research on pretraining systems on large-scale datasets for audio pattern recognition. In this paper, we propose pretrained audio neural networks (PANNs) trained on the large-scale AudioSet dataset. These PANNs are transferred to other audio related tasks. We investigate the performance and computational complexity of PANNs modeled by a variety of convolutional neural networks. We propose an architecture called Wavegram-Logmel-CNN using both log-mel spectrogram and waveform as input feature. Our best PANN system achieves a state-of-the-art mean average precision (mAP) of 0.439 on AudioSet tagging, outperforming the best previous system of 0.392. We transfer PANNs to six audio pattern recognition tasks, and demonstrate state-of-the-art performance in several of those tasks. We have released the source code and pretrained models of PANNs: https://github.com/qiuqiangkong/audioset_tagging_cnn.

PDF Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Audio Classification AudioSet PANNs-CNN14 (Single) Test mAP 0.431 # 20
AUC 0.973 # 5
d-prime 2.732 # 5
Audio Tagging AudioSet CNN14 mean average precision 0.431 # 8

Methods


No methods listed for this paper. Add relevant methods here