HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Data Augmentation for BERT Fine-Tuning in Open-Domain Question Answering

Wei Yang; Yuqing Xie; Luchen Tan; Kun Xiong; Ming Li; Jimmy Lin

Data Augmentation for BERT Fine-Tuning in Open-Domain Question Answering

Abstract

Recently, a simple combination of passage retrieval using off-the-shelf IR techniques and a BERT reader was found to be very effective for question answering directly on Wikipedia, yielding a large improvement over the previous state of the art on a standard benchmark dataset. In this paper, we present a data augmentation technique using distant supervision that exploits positive as well as negative examples. We apply a stage-wise approach to fine tuning BERT on multiple datasets, starting with data that is "furthest" from the test data and ending with the "closest". Experimental results show large gains in effectiveness over previous approaches on English QA datasets, and we establish new baselines on two recent Chinese QA datasets.

Benchmarks

BenchmarkMethodologyMetrics
open-domain-question-answering-on-squad1-1BERTserini
EM: 50.2

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Data Augmentation for BERT Fine-Tuning in Open-Domain Question Answering | Papers | HyperAI