Command Palette
Search for a command to run...
Jeremy Howard; Sebastian Ruder

Abstract
Inductive transfer learning has greatly impacted computer vision, but existing approaches in NLP still require task-specific modifications and training from scratch. We propose Universal Language Model Fine-tuning (ULMFiT), an effective transfer learning method that can be applied to any task in NLP, and introduce techniques that are key for fine-tuning a language model. Our method significantly outperforms the state-of-the-art on six text classification tasks, reducing the error by 18-24% on the majority of datasets. Furthermore, with only 100 labeled examples, it matches the performance of training from scratch on 100x more data. We open-source our pretrained models and code.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| sentiment-analysis-on-imdb | ULMFiT | Accuracy: 95.4 |
| sentiment-analysis-on-yelp-binary | ULMFiT | Error: 2.16 |
| sentiment-analysis-on-yelp-fine-grained | ULMFiT | Error: 29.98 |
| text-classification-on-ag-news | ULMFiT | Error: 5.01 |
| text-classification-on-dbpedia | ULMFiT | Error: 0.80 |
| text-classification-on-trec-6 | ULMFiT | Error: 3.6 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.