HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

Learning Implicitly Recurrent CNNs Through Parameter Sharing

Pedro Savarese; Michael Maire

Learning Implicitly Recurrent CNNs Through Parameter Sharing

Abstract

We introduce a parameter sharing scheme, in which different layers of a convolutional neural network (CNN) are defined by a learned linear combination of parameter tensors from a global bank of templates. Restricting the number of templates yields a flexible hybridization of traditional CNNs and recurrent networks. Compared to traditional CNNs, we demonstrate substantial parameter savings on standard image classification tasks, while maintaining accuracy. Our simple parameter sharing scheme, though defined via soft weights, in practice often yields trained networks with near strict recurrent structure; with negligible side effects, they convert into networks with actual loops. Training these networks thus implicitly involves discovery of suitable recurrent architectures. Though considering only the design aspect of recurrent links, our trained networks achieve accuracy competitive with those built using state-of-the-art neural architecture search (NAS) procedures. Our hybridization of recurrent and convolutional networks may also represent a beneficial architectural bias. Specifically, on synthetic tasks which are algorithmic in nature, our hybrid networks both train faster and extrapolate better to test examples outside the span of the training set.

Code Repositories

lolemacs/soft-sharing
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
architecture-search-on-cifar-10-imageSoft Parameter Sharing
Params: 33.5
Percentage error: 2.53
image-classification-on-cifar-10Shared WRN
Percentage correct: 97.47
image-classification-on-cifar-100Shared WRN
Percentage correct: 82.57
neural-architecture-search-on-cifar-10Soft Parameter Sharing
Search Time (GPU days): 0.7
Top-1 Error Rate: 2.53%

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Learning Implicitly Recurrent CNNs Through Parameter Sharing | Papers | HyperAI