Command Palette
Search for a command to run...
Xinyi Wang Jason Weston Michael Auli Yacine Jernite

Abstract
Neural sequence to sequence models are well established for applications which can be cast as mapping a single input sequence into a single output sequence. In this work, we focus on cases where generation is conditioned on both a short query and a long context, such as abstractive question answering or document-level translation. We modify the standard sequence-to-sequence approach to make better use of both the query and the context by expanding the conditioning mechanism to intertwine query and context attention. We also introduce a simple and efficient data augmentation method for the proposed model. Experiments on three different tasks show that both changes lead to consistent improvements.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| open-domain-question-answering-on-eli5 | Multi-Inrerleave | Rouge-1: 23.32 Rouge-2: 4.79 Rouge-L: 14.63 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.