HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Context-Aware Transformer Pre-Training for Answer Sentence Selection

Luca Di Liello Siddhant Garg Alessandro Moschitti

Context-Aware Transformer Pre-Training for Answer Sentence Selection

Abstract

Answer Sentence Selection (AS2) is a core component for building an accurate Question Answering pipeline. AS2 models rank a set of candidate sentences based on how likely they answer a given question. The state of the art in AS2 exploits pre-trained transformers by transferring them on large annotated datasets, while using local contextual information around the candidate sentence. In this paper, we propose three pre-training objectives designed to mimic the downstream fine-tuning task of contextual AS2. This allows for specializing LMs when fine-tuning for contextual AS2. Our experiments on three public and two large-scale industrial datasets show that our pre-training approaches (applied to RoBERTa and ELECTRA) can improve baseline contextual AS2 accuracy by up to 8% on some datasets.

Benchmarks

BenchmarkMethodologyMetrics
question-answering-on-trecqaContextual DeBERTa-V3-Large + SSP
MAP: 0.919
MRR: 0.945

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Context-Aware Transformer Pre-Training for Answer Sentence Selection | Papers | HyperAI