HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Two-stage Textual Knowledge Distillation for End-to-End Spoken Language Understanding

Seongbin Kim Gyuwan Kim Seongjin Shin Sangmin Lee

Two-stage Textual Knowledge Distillation for End-to-End Spoken Language Understanding

Abstract

End-to-end approaches open a new way for more accurate and efficient spoken language understanding (SLU) systems by alleviating the drawbacks of traditional pipeline systems. Previous works exploit textual information for an SLU model via pre-training with automatic speech recognition or fine-tuning with knowledge distillation. To utilize textual information more effectively, this work proposes a two-stage textual knowledge distillation method that matches utterance-level representations and predicted logits of two modalities during pre-training and fine-tuning, sequentially. We use vq-wav2vec BERT as a speech encoder because it captures general and rich features. Furthermore, we improve the performance, especially in a low-resource scenario, with data augmentation methods by randomly masking spans of discrete audio tokens and contextualized hidden representations. Consequently, we push the state-of-the-art on the Fluent Speech Commands, achieving 99.7% test accuracy in the full dataset setting and 99.5% in the 10% subset setting. Throughout the ablation studies, we empirically verify that all used methods are crucial to the final performance, providing the best practice for spoken language understanding. Code is available at https://github.com/clovaai/textual-kd-slu.

Code Repositories

clovaai/textual-kd-slu
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
spoken-language-understanding-on-fluenttextual-kd-slu
Accuracy (%): 99.7

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Two-stage Textual Knowledge Distillation for End-to-End Spoken Language Understanding | Papers | HyperAI