Fully Convolutional Speech Recognition

Current state-of-the-art speech recognition systems build on recurrent neural networks for acoustic and/or language modeling, and rely on feature extraction pipelines to extract mel-filterbanks or cepstral coefficients. In this paper we present an alternative approach based solely on convolutional neural networks, leveraging recent advances in acoustic models from the raw waveform and language modeling. This fully convolutional approach is trained end-to-end to predict characters from the raw waveform, removing the feature extraction step altogether. An external convolutional language model is used to decode words. On Wall Street Journal, our model matches the current state-of-the-art. On Librispeech, we report state-of-the-art performance among end-to-end models, including Deep Speech 2 trained with 12 times more acoustic data and significantly more linguistic data.

PDF Abstract

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Speech Recognition LibriSpeech test-clean Convolutional Speech Recognition Word Error Rate (WER) 3.26 # 40
Speech Recognition LibriSpeech test-other Convolutional Speech Recognition Word Error Rate (WER) 10.47 # 41
Speech Recognition WSJ dev93 Convolutional Speech Recognition Word Error Rate (WER) 6.8 # 4
Speech Recognition WSJ eval92 Convolutional Speech Recognition Word Error Rate (WER) 3.5 # 10
Speech Recognition WSJ eval93 Convolutional Speech Recognition Word Error Rate (WER) 6.8 # 3

Methods


No methods listed for this paper. Add relevant methods here