HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

GPU Kernels for Block-Sparse Weights

{Alec Radford and Diederik P. Kingma Scott Gray}

Abstract

We’re releasing highly optimized GPU kernels for an underexplored class of neural network architectures: networks with block-sparse weights. The kernels allow for efficient evaluation and differentiation of linear layers, including convolutional layers, with flexibly configurable block-sparsity patterns in the weight matrix. We find that depending on the sparsity, these kernels can run orders of magnitude faster than the best available alternatives such as cuBLAS. Using the kernels we improve upon the state-of-the-art in text sentiment analysis and generative modeling of text and images. By releasing our kernels in the open we aim to spur furtheradvancement in model and algorithm design.

Benchmarks

BenchmarkMethodologyMetrics
sentiment-analysis-on-crBlock-sparse LSTM
Accuracy: 92.2
sentiment-analysis-on-imdbBlock-sparse LSTM
Accuracy: 94.99
sentiment-analysis-on-sst-2-binaryBlock-sparse LSTM
Accuracy: 93.2
sentiment-analysis-on-yelp-binaryBlock-sparse LSTM
Error: 3.27

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
GPU Kernels for Block-Sparse Weights | Papers | HyperAI