HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Cross-lingual Language Model Pretraining

Guillaume Lample; Alexis Conneau

Cross-lingual Language Model Pretraining

Abstract

Recent studies have demonstrated the efficiency of generative pretraining for English natural language understanding. In this work, we extend this approach to multiple languages and show the effectiveness of cross-lingual pretraining. We propose two methods to learn cross-lingual language models (XLMs): one unsupervised that only relies on monolingual data, and one supervised that leverages parallel data with a new cross-lingual language model objective. We obtain state-of-the-art results on cross-lingual classification, unsupervised and supervised machine translation. On XNLI, our approach pushes the state of the art by an absolute gain of 4.9% accuracy. On unsupervised machine translation, we obtain 34.3 BLEU on WMT'16 German-English, improving the previous state of the art by more than 9 BLEU. On supervised machine translation, we obtain a new state of the art of 38.5 BLEU on WMT'16 Romanian-English, outperforming the previous best approach by more than 4 BLEU. Our code and pretrained models will be made publicly available.

Code Repositories

1-punchMan/CLTS
pytorch
Mentioned in GitHub
JunnYu/xlm_paddle
paddle
Mentioned in GitHub
facebookresearch/UnsupervisedMT
pytorch
Mentioned in GitHub
fshdnc/enfi-XLM
pytorch
Mentioned in GitHub
Tikquuss/meta_XLM
pytorch
Mentioned in GitHub
facebookresearch/MLQA
Mentioned in GitHub
huggingface/transformers
pytorch
Mentioned in GitHub
facebookresearch/XLM
pytorch
Mentioned in GitHub
Somefive/XNLI
pytorch
Mentioned in GitHub
samwisegamjeee/pytorch-transformers
pytorch
Mentioned in GitHub
feyzaakyurek/XLM-LwLL
pytorch
Mentioned in GitHub
kheeong/XLM_OWN
pytorch
Mentioned in GitHub
deterministic-algorithms-lab/Large-XLM
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
machine-translation-on-wmt2016-romanianMLM pretraining
BLEU score: 35.3
natural-language-inference-on-xnli-frenchXLM (MLM+TLM)
Accuracy: 80.2
unsupervised-machine-translation-on-wmt2014-1MLM pretraining for encoder and decoder
BLEU: 33.3
unsupervised-machine-translation-on-wmt2014-2MLM pretraining for encoder and decoder
BLEU: 33.4
unsupervised-machine-translation-on-wmt2016MLM pretraining for encoder and decoder
BLEU: 26.4
unsupervised-machine-translation-on-wmt2016-1MLM pretraining for encoder and decoder
BLEU: 34.3
unsupervised-machine-translation-on-wmt2016-2MLM pretraining for encoder and decoder
BLEU: 33.3
unsupervised-machine-translation-on-wmt2016-3MLM pretraining for encoder and decoder
BLEU: 31.8
unsupervised-machine-translation-on-wmt2016-5MLM pretraining for encoder and decoder
BLEU: 33.3

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp