HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Pre-training Transformer Models with Sentence-Level Objectives for Answer Sentence Selection

Luca Di Liello Siddhant Garg Luca Soldaini Alessandro Moschitti

Pre-training Transformer Models with Sentence-Level Objectives for Answer Sentence Selection

Abstract

An important task for designing QA systems is answer sentence selection (AS2): selecting the sentence containing (or constituting) the answer to a question from a set of retrieved relevant documents. In this paper, we propose three novel sentence-level transformer pre-training objectives that incorporate paragraph-level semantics within and across documents, to improve the performance of transformers for AS2, and mitigate the requirement of large labeled datasets. Specifically, the model is tasked to predict whether: (i) two sentences are extracted from the same paragraph, (ii) a given sentence is extracted from a given paragraph, and (iii) two paragraphs are extracted from the same document. Our experiments on three public and one industrial AS2 datasets demonstrate the empirical superiority of our pre-trained transformers over baseline models such as RoBERTa and ELECTRA for AS2.

Benchmarks

BenchmarkMethodologyMetrics
answer-selection-on-asnqELECTRA-Base + SSP
MAP: 0.697
MRR: 0.757
answer-selection-on-asnqDeBERTa-V3-Large + SSP
MAP: 0.743
MRR: 0.800
question-answering-on-trecqaDeBERTa-V3-Large + SSP
MAP: 0.923
MRR: 0.946
question-answering-on-trecqaRoBERTa-Base + PSD
MAP: 0.903
MRR: 0.951
question-answering-on-wikiqaDeBERTa-Large + SSP
MAP: 0.901
MRR: 0.914
question-answering-on-wikiqaRoBERTa-Base + SSP
MAP: 0.887
MRR: 0.899
question-answering-on-wikiqaDeBERTa-V3-Large + ALL
MAP: 0.909
MRR: 0.920

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Pre-training Transformer Models with Sentence-Level Objectives for Answer Sentence Selection | Papers | HyperAI