Command Palette
Search for a command to run...
Minjoon Seo; Aniruddha Kembhavi; Ali Farhadi; Hannaneh Hajishirzi

Abstract
Machine comprehension (MC), answering a query about a given context paragraph, requires modeling complex interactions between the context and the query. Recently, attention mechanisms have been successfully extended to MC. Typically these methods use attention to focus on a small portion of the context and summarize it with a fixed-size vector, couple attentions temporally, and/or often form a uni-directional attention. In this paper we introduce the Bi-Directional Attention Flow (BIDAF) network, a multi-stage hierarchical process that represents the context at different levels of granularity and uses bi-directional attention flow mechanism to obtain a query-aware context representation without early summarization. Our experimental evaluations show that our model achieves the state-of-the-art results in Stanford Question Answering Dataset (SQuAD) and CNN/DailyMail cloze test.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| open-domain-question-answering-on-quasar | BiDAF | EM (Quasar-T): 25.9 F1 (Quasar-T): 28.5 |
| question-answering-on-cnn-daily-mail | BiDAF | CNN: 76.9 Daily Mail: 79.6 |
| question-answering-on-ms-marco | BiDaF Baseline | BLEU-1: 10.64 Rouge-L: 23.96 |
| question-answering-on-narrativeqa | BiDAF | BLEU-1: 33.45 BLEU-4: 15.69 METEOR: 15.68 Rouge-L: 36.74 |
| question-answering-on-squad11 | BiDAF (single model) | EM: 67.974 F1: 77.323 |
| question-answering-on-squad11 | BiDAF (ensemble) | EM: 73.744 F1: 81.525 |
| question-answering-on-squad11-dev | BIDAF (single) | EM: 67.7 F1: 77.3 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.