4 个月前

作为少样本学习者的蕴含关系

作为少样本学习者的蕴含关系

摘要

大型预训练语言模型(LMs)在少量样本学习方面展现了卓越的能力。然而,它们的成功很大程度上依赖于模型参数的扩展,这使得训练和部署变得极具挑战性。本文提出了一种新的方法,称为EFL,该方法可以将小型语言模型转化为更优秀的少量样本学习者。该方法的核心思想是将潜在的自然语言处理任务重新表述为蕴含任务,然后使用最少8个样例对模型进行微调。我们进一步展示了所提出的方法可以:(i) 自然地与基于无监督对比学习的数据增强方法结合;(ii) 轻松扩展到多语言少量样本学习。系统评估了18个标准自然语言处理任务的结果表明,该方法在各种现有的最先进(SOTA)少量样本学习方法基础上性能提升了12%,并且在仅有500倍于大型模型如GPT-3的规模下仍能取得具有竞争力的少量样本学习效果。

基准测试

基准方法指标
linguistic-acceptability-on-colaRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 86.4%
natural-language-inference-on-qnliRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 94.5%
natural-language-inference-on-rteRoBERTa-large 355M + EFL + UCA
Accuracy: 87.2%
natural-language-inference-on-rteRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 90.5%
natural-language-inference-on-snliNeural Tree Indexers for Text Understanding
% Test Accuracy: 93.1
Parameters: 355
natural-language-inference-on-snliEFL (Entailment as Few-shot Learner) + RoBERTa-large
% Test Accuracy: 93.1
% Train Accuracy: ?
Parameters: 355m
paraphrase-identification-on-quora-questionRoBERTa-large 355M + Entailment as Few-shot Learner
F1: 89.2
question-answering-on-boolqRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 86.0
semantic-textual-similarity-on-mrpcRoBERTa-large 355M + Entailment as Few-shot Learner
F1: 91.0
semantic-textual-similarity-on-sts-benchmarkRoBERTa-large 355M + Entailment as Few-shot Learner
Pearson Correlation: 0.918
sentiment-analysis-on-crRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 92.5
sentiment-analysis-on-imdbRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 96.1
sentiment-analysis-on-mpqaRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 90.8
sentiment-analysis-on-mrRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 92.5
sentiment-analysis-on-sst-2-binaryRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 96.9
subjectivity-analysis-on-subjRoBERTa-large 355M + Entailment as Few-shot Learner
Accuracy: 97.1

用 AI 构建 AI

从想法到上线——通过免费 AI 协同编程、开箱即用的环境和市场最优价格的 GPU 加速您的 AI 开发

AI 协同编程
即用型 GPU
最优价格
立即开始

Hyper Newsletters

订阅我们的最新资讯
我们会在北京时间 每周一的上午九点 向您的邮箱投递本周内的最新更新
邮件发送服务由 MailChimp 提供
作为少样本学习者的蕴含关系 | 论文 | HyperAI超神经