Command Palette
Search for a command to run...
Pre-training Transformer Models with Sentence-Level Objectives for Answer Sentence Selection
Luca Di Liello Siddhant Garg Luca Soldaini Alessandro Moschitti

Abstract
An important task for designing QA systems is answer sentence selection (AS2): selecting the sentence containing (or constituting) the answer to a question from a set of retrieved relevant documents. In this paper, we propose three novel sentence-level transformer pre-training objectives that incorporate paragraph-level semantics within and across documents, to improve the performance of transformers for AS2, and mitigate the requirement of large labeled datasets. Specifically, the model is tasked to predict whether: (i) two sentences are extracted from the same paragraph, (ii) a given sentence is extracted from a given paragraph, and (iii) two paragraphs are extracted from the same document. Our experiments on three public and one industrial AS2 datasets demonstrate the empirical superiority of our pre-trained transformers over baseline models such as RoBERTa and ELECTRA for AS2.
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| answer-selection-on-asnq | ELECTRA-Base + SSP | MAP: 0.697 MRR: 0.757 |
| answer-selection-on-asnq | DeBERTa-V3-Large + SSP | MAP: 0.743 MRR: 0.800 |
| question-answering-on-trecqa | DeBERTa-V3-Large + SSP | MAP: 0.923 MRR: 0.946 |
| question-answering-on-trecqa | RoBERTa-Base + PSD | MAP: 0.903 MRR: 0.951 |
| question-answering-on-wikiqa | DeBERTa-Large + SSP | MAP: 0.901 MRR: 0.914 |
| question-answering-on-wikiqa | RoBERTa-Base + SSP | MAP: 0.887 MRR: 0.899 |
| question-answering-on-wikiqa | DeBERTa-V3-Large + ALL | MAP: 0.909 MRR: 0.920 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.