HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Attention Boosted Sequential Inference Model

Guanyu Li; Pengfei Zhang; Caiyan Jia

Attention Boosted Sequential Inference Model

Abstract

Attention mechanism has been proven effective on natural language processing. This paper proposes an attention boosted natural language inference model named aESIM by adding word attention and adaptive direction-oriented attention mechanisms to the traditional Bi-LSTM layer of natural language inference models, e.g. ESIM. This makes the inference model aESIM has the ability to effectively learn the representation of words and model the local subsentential inference between pairs of premise and hypothesis. The empirical studies on the SNLI, MultiNLI and Quora benchmarks manifest that aESIM is superior to the original ESIM model.

Benchmarks

BenchmarkMethodologyMetrics
natural-language-inference-on-multinliaESIM
Matched: 73.9
Mismatched: 73.9
natural-language-inference-on-quora-questionaESIM
Accuracy: 88.01
natural-language-inference-on-snliaESIM
% Test Accuracy: 88.1

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Attention Boosted Sequential Inference Model | Papers | HyperAI