HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Single Headed Attention RNN: Stop Thinking With Your Head

Stephen Merity

Single Headed Attention RNN: Stop Thinking With Your Head

Abstract

The leading approaches in language modeling are all obsessed with TV shows of my youth - namely Transformers and Sesame Street. Transformers this, Transformers that, and over here a bonfire worth of GPU-TPU-neuromorphic wafer scale silicon. We opt for the lazy path of old and proven techniques with a fancy crypto inspired acronym: the Single Headed Attention RNN (SHA-RNN). The author's lone goal is to show that the entire field might have evolved a different direction if we had instead been obsessed with a slightly different acronym and slightly different result. We take a previously strong language model based only on boring LSTMs and get it to within a stone's throw of a stone's throw of state-of-the-art byte level language model results on enwik8. This work has undergone no intensive hyperparameter optimization and lived entirely on a commodity desktop machine that made the author's small studio apartment far too warm in the midst of a San Franciscan summer. The final results are achievable in plus or minus 24 hours on a single GPU as the author is impatient. The attention mechanism is also readily extended to large contexts with minimal computation. Take that Sesame Street.

Code Repositories

Smerity/sha-rnn
Official
pytorch
Mentioned in GitHub
Tobias-K93/media-bias-prediction
pytorch
Mentioned in GitHub
alisafaya/SHA-RNN.jl
pytorch
Mentioned in GitHub
floleuerer/fastai_ulmfit
Mentioned in GitHub
saattrupdan/scholarly
tf
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
language-modelling-on-enwiki8SHA-LSTM (4 layers, h=1024, no attention head)
Bit per Character (BPC): 1.33
Number of params: 51M
language-modelling-on-enwiki8SHA-RNN (4 layers, h=1024, single attention head)
Bit per Character (BPC): 1.076
Number of params: 52M
language-modelling-on-enwiki8SHA-RNN (4 layers, h=1024, attention head per layer)
Bit per Character (BPC): 1.068
Number of params: 54M

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Single Headed Attention RNN: Stop Thinking With Your Head | Papers | HyperAI