HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension

Mike Lewis Yinhan Liu Naman Goyal Marjan Ghazvininejad Abdelrahman Mohamed Omer Levy Ves Stoyanov Luke Zettlemoyer

BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension

Abstract

We present BART, a denoising autoencoder for pretraining sequence-to-sequence models. BART is trained by (1) corrupting text with an arbitrary noising function, and (2) learning a model to reconstruct the original text. It uses a standard Tranformer-based neural machine translation architecture which, despite its simplicity, can be seen as generalizing BERT (due to the bidirectional encoder), GPT (with the left-to-right decoder), and many other more recent pretraining schemes. We evaluate a number of noising approaches, finding the best performance by both randomly shuffling the order of the original sentences and using a novel in-filling scheme, where spans of text are replaced with a single mask token. BART is particularly effective when fine tuned for text generation but also works well for comprehension tasks. It matches the performance of RoBERTa with comparable training resources on GLUE and SQuAD, achieves new state-of-the-art results on a range of abstractive dialogue, question answering, and summarization tasks, with gains of up to 6 ROUGE. BART also provides a 1.1 BLEU increase over a back-translation system for machine translation, with only target language pretraining. We also report ablation experiments that replicate other pretraining schemes within the BART framework, to better measure which factors most influence end-task performance.

Code Repositories

shijx12/kqapro_baselines
pytorch
Mentioned in GitHub
W4ngatang/qags
pytorch
Mentioned in GitHub
tangg555/sabart
pytorch
Mentioned in GitHub
awalther/scibart
pytorch
Mentioned in GitHub
jiacheng-xu/text-sum-uncertainty
pytorch
Mentioned in GitHub
chakravarthi-v/Polaroid-1
pytorch
Mentioned in GitHub
facebookresearch/GENRE
pytorch
Mentioned in GitHub
mcao610/Factual-Error-Correction
pytorch
Mentioned in GitHub
microsoft/fastseq
pytorch
Mentioned in GitHub
jongwooko/nash-pruning-official
pytorch
Mentioned in GitHub
vgaraujov/seq2seq-spanish-plms
pytorch
Mentioned in GitHub
xieyxclack/factual_coco
pytorch
Mentioned in GitHub
nlmatics/llmsherpa
Mentioned in GitHub
tanyuqian/aspect-based-summarization
pytorch
Mentioned in GitHub
shmsw25/bart-closed-book-qa
pytorch
Mentioned in GitHub
thefonseca/factorsum
pytorch
Mentioned in GitHub
zhdbwe/Paper-DailyReading
tf
Mentioned in GitHub
KushGrandhi/Polaroid
pytorch
Mentioned in GitHub
john-bradshaw/rxn-lm
pytorch
Mentioned in GitHub
allenai/scientific-claim-generation
pytorch
Mentioned in GitHub
vinayak19th/Brevis-2.0
Mentioned in GitHub
udnet96/BART-various-finetune
pytorch
Mentioned in GitHub
huggingface/transformers
pytorch
Mentioned in GitHub
dawn0815/UniSA
pytorch
Mentioned in GitHub
facebookresearch/bart_ls
pytorch
Mentioned in GitHub
skt-ai/kobart
Mentioned in GitHub
qywu/memformers
pytorch
Mentioned in GitHub
i2r-simmc/i2r-simmc-2020
pytorch
Mentioned in GitHub
huangxt39/BART_on_COVID_dialogue
pytorch
Mentioned in GitHub
microsoft/Table-Pretraining
pytorch
Mentioned in GitHub
maanvithag/thinkai
Mentioned in GitHub
timrozday/spl-indications-bart
pytorch
Mentioned in GitHub
HHousen/TransformerSum
pytorch
Mentioned in GitHub
wyu97/Easy-use-BART
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
abstractive-text-summarization-on-cnn-dailyBART
ROUGE-1: 44.16
ROUGE-2: 21.28
ROUGE-L: 40.90
open-domain-question-answering-on-eli5BART
Rouge-1: 30.6
Rouge-2: 6.2
Rouge-L: 24.3
question-answering-on-squad11-devBART Base (with text infilling)
F1: 90.8
text-summarization-on-x-sumBART
ROUGE-1: 45.14
ROUGE-2: 22.27
ROUGE-3: 37.25

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension | Papers | HyperAI