Command Palette
Search for a command to run...
Runqi Yang; Jianhai Zhang; Xing Gao; Feng Ji; Haiqing Chen

Abstract
In this paper, we present a fast and strong neural approach for general purpose text matching applications. We explore what is sufficient to build a fast and well-performed text matching model and propose to keep three key features available for inter-sequence alignment: original point-wise features, previous aligned features, and contextual features while simplifying all the remaining components. We conduct experiments on four well-studied benchmark datasets across tasks of natural language inference, paraphrase identification and answer selection. The performance of our model is on par with the state-of-the-art on all datasets with much fewer parameters and the inference speed is at least 6 times faster compared with similarly performed ones.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| natural-language-inference-on-scitail | RE2 | Accuracy: 86.0 |
| natural-language-inference-on-snli | RE2 | % Test Accuracy: 88.9 % Train Accuracy: 94.0 Parameters: 2.8m |
| paraphrase-identification-on-quora-question | RE2 | Accuracy: 89.2 |
| question-answering-on-quora-question-pairs | RE2 | Accuracy: 89.2 % |
| question-answering-on-wikiqa | RE2 | MAP: 0.7452 MRR: 0.7618 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.