Command Palette
Search for a command to run...
Jinhyuk Lee Mujeen Sung Jaewoo Kang Danqi Chen

Abstract
Open-domain question answering can be reformulated as a phrase retrieval problem, without the need for processing documents on-demand during inference (Seo et al., 2019). However, current phrase retrieval models heavily depend on sparse representations and still underperform retriever-reader approaches. In this work, we show for the first time that we can learn dense representations of phrases alone that achieve much stronger performance in open-domain QA. We present an effective method to learn phrase representations from the supervision of reading comprehension tasks, coupled with novel negative sampling methods. We also propose a query-side fine-tuning strategy, which can support transfer learning and reduce the discrepancy between training and inference. On five popular open-domain QA datasets, our model DensePhrases improves over previous phrase retrieval models by 15%-25% absolute accuracy and matches the performance of state-of-the-art retriever-reader models. Our model is easy to parallelize due to pure dense representations and processes more than 10 questions per second on CPUs. Finally, we directly use our pre-indexed dense phrase representations for two slot filling tasks, showing the promise of utilizing DensePhrases as a dense knowledge base for downstream tasks.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| question-answering-on-natural-questions-long | DensePhrases | EM: 71.9 F1: 79.6 |
| question-answering-on-squad11-dev | DensePhrases | EM: 78.3 F1: 86.3 |
| slot-filling-on-kilt-t-rex | DensePhrases | Accuracy: 53.9 F1: 61.74 KILT-AC: 27.84 KILT-F1: 32.34 R-Prec: 37.62 Recall@5: 40.07 |
| slot-filling-on-kilt-zero-shot-re | DensePhrases | Accuracy: 47.42 F1: 54.75 KILT-AC: 41.34 KILT-F1: 46.79 R-Prec: 57.43 Recall@5: 60.47 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.