HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Neural Network Language Modeling with Letter-based Features and Importance Sampling

{Xie Chen Sanjeev Khudanpur Ke Li Jian Wang Yiming Wang Daniel Povey Hainan Xu Shiyin Kang}

Abstract

In this paper we describe an extension of the Kaldi software toolkit to support neural-based language modeling, intendedfor use in automatic speech recognition (ASR) and related tasks. We combine the use of subword features (letter n-grams) and one-hot encoding of frequent words so that the models can handle large vocabularies containing infrequentwords. We propose a new objective function that allows for training of unnormalized probabilities. An importance sampling based method is supported to speed up training when the vocabulary is large. Experimental results on five corpora show that Kaldi-RNNLM rivals other recurrent neural network language model toolkits both on performance and training speed.

Benchmarks

BenchmarkMethodologyMetrics
speech-recognition-on-librispeech-test-cleantdnn + chain + rnnlm rescoring
Word Error Rate (WER): 3.06
speech-recognition-on-librispeech-test-othertdnn + chain + rnnlm rescoring
Word Error Rate (WER): 7.63

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Neural Network Language Modeling with Letter-based Features and Importance Sampling | Papers | HyperAI