HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Improving Conditioning in Context-Aware Sequence to Sequence Models

Xinyi Wang Jason Weston Michael Auli Yacine Jernite

Improving Conditioning in Context-Aware Sequence to Sequence Models

Abstract

Neural sequence to sequence models are well established for applications which can be cast as mapping a single input sequence into a single output sequence. In this work, we focus on cases where generation is conditioned on both a short query and a long context, such as abstractive question answering or document-level translation. We modify the standard sequence-to-sequence approach to make better use of both the query and the context by expanding the conditioning mechanism to intertwine query and context attention. We also introduce a simple and efficient data augmentation method for the proposed model. Experiments on three different tasks show that both changes lead to consistent improvements.

Benchmarks

BenchmarkMethodologyMetrics
open-domain-question-answering-on-eli5Multi-Inrerleave
Rouge-1: 23.32
Rouge-2: 4.79
Rouge-L: 14.63

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Improving Conditioning in Context-Aware Sequence to Sequence Models | Papers | HyperAI