Command Palette
Search for a command to run...
An Effective Domain Adaptive Post-Training Method for BERT in Response Selection
Taesun Whang; Dongyub Lee; Chanhee Lee; Kisu Yang; Dongsuk Oh; HeuiSeok Lim

Abstract
We focus on multi-turn response selection in a retrieval-based dialog system. In this paper, we utilize the powerful pre-trained language model Bi-directional Encoder Representations from Transformer (BERT) for a multi-turn dialog system and propose a highly effective post-training method on domain-specific corpus. Although BERT is easily adopted to various NLP tasks and outperforms previous baselines of each task, it still has limitations if a task corpus is too focused on a certain domain. Post-training on domain-specific corpus (e.g., Ubuntu Corpus) helps the model to train contextualized representations and words that do not appear in general corpus (e.g., English Wikipedia). Experimental results show that our approach achieves new state-of-the-art on two response selection benchmarks (i.e., Ubuntu Corpus V1, Advising Corpus) performance improvement by 5.9% and 6% on R@1.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| conversational-response-selection-on-douban-1 | BERT | MAP: 0.591 MRR: 0.633 P@1: 0.454 R10@1: 0.280 R10@2: 0.470 R10@5: 0.828 |
| conversational-response-selection-on-rrs | BERT | MAP: 0.625 MRR: 0.639 P@1: 0.453 R10@1: 0.404 R10@2: 0.606 R10@5: 0.875 |
| conversational-response-selection-on-rrs-1 | BERT | NDCG@3: 0.625 NDCG@5: 0.714 |
| conversational-response-selection-on-ubuntu-1 | BERT-VFT | R10@1: 0.855 R10@2: 0.928 R10@5: 0.985 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.