HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-based Question Answering

Changmao Li Jinho D. Choi

Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-based Question Answering

Abstract

We introduce a novel approach to transformers that learns hierarchical representations in multiparty dialogue. First, three language modeling tasks are used to pre-train the transformers, token- and utterance-level language modeling and utterance order prediction, that learn both token and utterance embeddings for better understanding in dialogue contexts. Then, multi-task learning between the utterance prediction and the token span prediction is applied to fine-tune for span-based question answering (QA). Our approach is evaluated on the FriendsQA dataset and shows improvements of 3.8% and 1.4% over the two state-of-the-art transformer models, BERT and RoBERTa, respectively.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
question-answering-on-friendsqaLi and Choi - BERT
EM: 46.8
F1: 63.1
question-answering-on-friendsqaLi and Choi - RoBERTa
EM: 53.5
F1: 69.6

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Transformers to Learn Hierarchical Contexts in Multiparty Dialogue for Span-based Question Answering | Papers | HyperAI