HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

RoBERTa: A Robustly Optimized BERT Pretraining Approach

Yinhan Liu; Myle Ott; Naman Goyal; Jingfei Du; Mandar Joshi; Danqi Chen; Omer Levy; Mike Lewis; Luke Zettlemoyer; Veselin Stoyanov

RoBERTa: A Robustly Optimized BERT Pretraining Approach

Abstract

Language model pretraining has led to significant performance gains but careful comparison between different approaches is challenging. Training is computationally expensive, often done on private datasets of different sizes, and, as we will show, hyperparameter choices have significant impact on the final results. We present a replication study of BERT pretraining (Devlin et al., 2019) that carefully measures the impact of many key hyperparameters and training data size. We find that BERT was significantly undertrained, and can match or exceed the performance of every model published after it. Our best model achieves state-of-the-art results on GLUE, RACE and SQuAD. These results highlight the importance of previously overlooked design choices, and raise questions about the source of recently reported improvements. We release our models and code.

Code Repositories

hkuds/easyrec
pytorch
Mentioned in GitHub
SindhuMadi/FakeNewsDetection
Mentioned in GitHub
expertailab/spaceqa
pytorch
Mentioned in GitHub
awslabs/mlm-scoring
mxnet
Mentioned in GitHub
common-english/bert-all
pytorch
Mentioned in GitHub
pytorch/fairseq
Official
pytorch
benywon/ReCO
pytorch
Mentioned in GitHub
UnknownGenie/altered-BERT-KPE
pytorch
Mentioned in GitHub
knuddj1/op_text
pytorch
Mentioned in GitHub
znhy1024/protoco
pytorch
Mentioned in GitHub
CalumPerrio/WNUT-2020
pytorch
Mentioned in GitHub
zfj1998/CodeBert-Code2Text
pytorch
Mentioned in GitHub
simon-benigeri/narrative-generation
pytorch
Mentioned in GitHub
dig-team/hanna-benchmark-asg
pytorch
Mentioned in GitHub
flexible-fl/flex-nlp
Mentioned in GitHub
salesforce/codet5
pytorch
Mentioned in GitHub
musixmatchresearch/umberto
pytorch
Mentioned in GitHub
facebookresearch/anli
pytorch
Mentioned in GitHub
nguyenvulebinh/vietnamese-roberta
pytorch
Mentioned in GitHub
viethoang1512/kpa
pytorch
Mentioned in GitHub
knuddy/op_text
pytorch
Mentioned in GitHub
wzzzd/LM_NER
pytorch
Mentioned in GitHub
sdadas/polish-roberta
pytorch
Mentioned in GitHub
Tencent/TurboTransformers
pytorch
Mentioned in GitHub
abdumaa/hiqualprop
pytorch
Mentioned in GitHub
devhemza/BERTweet_sentiment_analysis
pytorch
Mentioned in GitHub
eternityyw/tram-benchmark
Mentioned in GitHub
huggingface/transformers
pytorch
Mentioned in GitHub
oneflow-inc/libai
Mentioned in GitHub
bfopengradient/NLP_ROBERTA
Mentioned in GitHub
clovaai/textual-kd-slu
pytorch
Mentioned in GitHub
xiaoqian19940510/text-classification-
pytorch
Mentioned in GitHub
aistairc/kirt_bert_on_abci
pytorch
Mentioned in GitHub
pisalore/roberta_results
pytorch
Mentioned in GitHub
bcaitech1/p2-klue-Heeseok-Jeong
pytorch
Mentioned in GitHub
mthcom/hscore-dataset-pruning
pytorch
Mentioned in GitHub
lashoun/hanna-benchmark-asg
pytorch
Mentioned in GitHub
kaushaltrivedi/fast-bert
pytorch
Mentioned in GitHub
octanove/shiba
pytorch
Mentioned in GitHub
traviscoan/cards
Mentioned in GitHub
utterworks/fast-bert
pytorch
Mentioned in GitHub
zaradana/Fast_BERT
pytorch
Mentioned in GitHub
IndicoDataSolutions/finetune
tf
Mentioned in GitHub
brightmart/roberta_zh
tf
Mentioned in GitHub
few-shot-NER-benchmark/BaselineCode
pytorch
Mentioned in GitHub
ibm/vira-intent-discovery
Mentioned in GitHub
blawok/named-entity-recognition
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
common-sense-reasoning-on-commonsenseqaRoBERTa-Large 355M
Accuracy: 72.1
common-sense-reasoning-on-swagRoBERTa
Test: 89.9
document-image-classification-on-rvl-cdipRoberta base
Accuracy: 90.06
Parameters: 125M
linguistic-acceptability-on-colaRoBERTa (ensemble)
Accuracy: 67.8%
multi-task-language-understanding-on-mmluRoBERTa-base 125M (fine-tuned)
Average (%): 27.9
natural-language-inference-on-anli-testRoBERTa (Large)
A1: 72.4
A2: 49.8
A3: 44.4
natural-language-inference-on-multinliRoBERTa
Matched: 90.8
natural-language-inference-on-multinliRoBERTa (ensemble)
Mismatched: 90.2
natural-language-inference-on-qnliRoBERTa (ensemble)
Accuracy: 98.9%
natural-language-inference-on-rteRoBERTa
Accuracy: 88.2%
natural-language-inference-on-rteRoBERTa (ensemble)
Accuracy: 88.2%
natural-language-inference-on-wnliRoBERTa (ensemble)
Accuracy: 89
question-answering-on-piqaRoBERTa-Large 355M
Accuracy: 79.4
question-answering-on-quora-question-pairsRoBERTa (ensemble)
Accuracy: 90.2%
question-answering-on-social-iqaRoBERTa-Large 355M (fine-tuned)
Accuracy: 76.7
question-answering-on-squad20RoBERTa (single model)
EM: 86.820
F1: 89.795
question-answering-on-squad20-devRoBERTa (no data aug)
EM: 86.5
F1: 89.4
reading-comprehension-on-raceRoBERTa
Accuracy: 83.2
Accuracy (High): 81.3
Accuracy (Middle): 86.5
semantic-textual-similarity-on-mrpcRoBERTa (ensemble)
Accuracy: 92.3%
semantic-textual-similarity-on-sts-benchmarkRoBERTa
Pearson Correlation: 0.922
sentiment-analysis-on-sst-2-binaryRoBERTa (ensemble)
Accuracy: 96.7
stock-market-prediction-on-astockRoBERTa WWM Ext (News+Factors)
Accuray: 62.49
F1-score: 62.54
Precision: 62.59
Recall: 62.51
stock-market-prediction-on-astockRoBERTa WWM Ext (News)
Accuray: 61.34
F1-score: 61.48
Precision: 61.97
Recall: 61.32
task-1-grouping-on-ocwRoBERTa (LARGE)
# Correct Groups: 29 ± 3
# Solved Walls: 0 ± 0
Adjusted Mutual Information (AMI): 9.4 ± .4
Adjusted Rand Index (ARI): 8.4 ± .3
Fowlkes Mallows Score (FMS): 26.7 ± .2
Wasserstein Distance (WD): 88.4 ± .4
text-classification-on-arxiv-10RoBERTa
Accuracy: 0.779
type-prediction-on-manytypes4typescriptRoBERTa
Average Accuracy: 59.84
Average F1: 57.54
Average Precision: 57.45
Average Recall: 57.62

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
RoBERTa: A Robustly Optimized BERT Pretraining Approach | Papers | HyperAI