Command Palette
Search for a command to run...
Pengcheng He; Xiaodong Liu; Weizhu Chen; Jianfeng Gao

Abstract
This paper proposes a hybrid neural network (HNN) model for commonsense reasoning. An HNN consists of two component models, a masked language model and a semantic similarity model, which share a BERT-based contextual encoder but use different model-specific input and output layers. HNN obtains new state-of-the-art results on three classic commonsense reasoning tasks, pushing the WNLI benchmark to 89%, the Winograd Schema Challenge (WSC) benchmark to 75.1%, and the PDP60 benchmark to 90.0%. An ablation study shows that language models and semantic similarity models are complementary approaches to commonsense reasoning, and HNN effectively combines the strengths of both. The code and pre-trained models will be publicly available at https://github.com/namisan/mt-dnn.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| coreference-resolution-on-winograd-schema | HNN | Accuracy: 75.1 |
| natural-language-inference-on-wnli | HNNensemble | Accuracy: 89 |
| natural-language-inference-on-wnli | HNN | Accuracy: 83.6 |
| natural-language-understanding-on-pdp60 | HNN | Accuracy: 90 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.