HyperAIHyperAI

Command Palette

Search for a command to run...

5 months ago

Linguistic Knowledge as Memory for Recurrent Neural Networks

Bhuwan Dhingra; Zhilin Yang; William W. Cohen; Ruslan Salakhutdinov

Linguistic Knowledge as Memory for Recurrent Neural Networks

Abstract

Training recurrent neural networks to model long term dependencies is difficult. Hence, we propose to use external linguistic knowledge as an explicit signal to inform the model which memories it should utilize. Specifically, external knowledge is used to augment a sequence with typed edges between arbitrarily distant elements, and the resulting graph is decomposed into directed acyclic subgraphs. We introduce a model that encodes such graphs as explicit memory in recurrent neural networks, and use it to model coreference relations in text. We apply our model to several text comprehension tasks and achieve new state-of-the-art results on all considered benchmarks, including CNN, bAbi, and LAMBADA. On the bAbi QA tasks, our model solves 15 out of the 20 tasks with only 1000 training examples per task. Analysis of the learned representations further demonstrates the ability of our model to encode fine-grained entity information across a document.

Benchmarks

BenchmarkMethodologyMetrics
question-answering-on-cnn-daily-mailGA+MAGE (32)
CNN: 78.6

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Linguistic Knowledge as Memory for Recurrent Neural Networks | Papers | HyperAI