Command Palette
Search for a command to run...
Tsendsuren Munkhdalai; Hong Yu

Abstract
We present a memory augmented neural network for natural language understanding: Neural Semantic Encoders. NSE is equipped with a novel memory update rule and has a variable sized encoding memory that evolves over time and maintains the understanding of input sequences through read}, compose and write operations. NSE can also access multiple and shared memories. In this paper, we demonstrated the effectiveness and the flexibility of NSE on five different natural language tasks: natural language inference, question answering, sentence classification, document sentiment analysis and machine translation where NSE achieved state-of-the-art performance when evaluated on publically available benchmarks. For example, our shared-memory model showed an encouraging result on neural machine translation, improving an attention-based baseline by approximately 1.0 BLEU.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| machine-translation-on-wmt2014-english-german | NSE-NSE | BLEU score: 17.9 Hardware Burden: Operations per network pass: |
| natural-language-inference-on-snli | 300D NSE encoders | % Test Accuracy: 84.6 % Train Accuracy: 86.2 Parameters: 3.0m |
| natural-language-inference-on-snli | 300D MMA-NSE encoders with attention | % Test Accuracy: 85.4 % Train Accuracy: 86.9 Parameters: 3.2m |
| question-answering-on-wikiqa | MMA-NSE attention | MAP: 0.6811 MRR: 0.6993 |
| sentiment-analysis-on-sst-2-binary | Neural Semantic Encoder | Accuracy: 89.7 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.