Command Palette
Search for a command to run...
Neil Zeghidour; Qiantong Xu; Vitaliy Liptchinsky; Nicolas Usunier; Gabriel Synnaeve; Ronan Collobert

Abstract
Current state-of-the-art speech recognition systems build on recurrent neural networks for acoustic and/or language modeling, and rely on feature extraction pipelines to extract mel-filterbanks or cepstral coefficients. In this paper we present an alternative approach based solely on convolutional neural networks, leveraging recent advances in acoustic models from the raw waveform and language modeling. This fully convolutional approach is trained end-to-end to predict characters from the raw waveform, removing the feature extraction step altogether. An external convolutional language model is used to decode words. On Wall Street Journal, our model matches the current state-of-the-art. On Librispeech, we report state-of-the-art performance among end-to-end models, including Deep Speech 2 trained with 12 times more acoustic data and significantly more linguistic data.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| speech-recognition-on-librispeech-test-clean | Convolutional Speech Recognition | Word Error Rate (WER): 3.26 |
| speech-recognition-on-librispeech-test-other | Convolutional Speech Recognition | Word Error Rate (WER): 10.47 |
| speech-recognition-on-wsj-dev93 | Convolutional Speech Recognition | Word Error Rate (WER): 6.8 |
| speech-recognition-on-wsj-eval92 | Convolutional Speech Recognition | Word Error Rate (WER): 3.5 |
| speech-recognition-on-wsj-eval93 | Convolutional Speech Recognition | Word Error Rate (WER): 6.8 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.