HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware

Han Cai; Ligeng Zhu; Song Han

ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware

Abstract

Neural architecture search (NAS) has a great impact by automatically designing effective neural network architectures. However, the prohibitive computational demand of conventional NAS algorithms (e.g. $10^4$ GPU hours) makes it difficult to \emph{directly} search the architectures on large-scale tasks (e.g. ImageNet). Differentiable NAS can reduce the cost of GPU hours via a continuous representation of network architecture but suffers from the high GPU memory consumption issue (grow linearly w.r.t. candidate set size). As a result, they need to utilize~\emph{proxy} tasks, such as training on a smaller dataset, or learning with only a few blocks, or training just for a few epochs. These architectures optimized on proxy tasks are not guaranteed to be optimal on the target task. In this paper, we present \emph{ProxylessNAS} that can \emph{directly} learn the architectures for large-scale target tasks and target hardware platforms. We address the high memory consumption issue of differentiable NAS and reduce the computational cost (GPU hours and GPU memory) to the same level of regular training while still allowing a large candidate set. Experiments on CIFAR-10 and ImageNet demonstrate the effectiveness of directness and specialization. On CIFAR-10, our model achieves 2.08\% test error with only 5.7M parameters, better than the previous state-of-the-art architecture AmoebaNet-B, while using 6$\times$ fewer parameters. On ImageNet, our model achieves 3.1\% better top-1 accuracy than MobileNetV2, while being 1.2$\times$ faster with measured GPU latency. We also apply ProxylessNAS to specialize neural architectures for hardware with direct hardware metrics (e.g. latency) and provide insights for efficient CNN architecture design.

Code Repositories

mit-han-lab/ProxylessNAS
pytorch
Mentioned in GitHub
mit-han-lab/haq-release
pytorch
Mentioned in GitHub
schoolboy-ju/Proxyless-NAS
pytorch
Mentioned in GitHub
ZTao-z/ProxylessNAS
pytorch
Mentioned in GitHub
mit-han-lab/once-for-all
pytorch
Mentioned in GitHub
NVlabs/unas
pytorch
Mentioned in GitHub
mit-han-lab/haq
pytorch
Mentioned in GitHub
mit-han-lab/amc
pytorch
Mentioned in GitHub
ito-rafael/once-for-all-2
pytorch
Mentioned in GitHub
osmr/imgclsmob
mxnet
Mentioned in GitHub
seulkiyeom/once-for-all
pytorch
Mentioned in GitHub
dapeter/nas-for-kws
pytorch
Mentioned in GitHub
AhmadQasim/proxylessnas-dense
pytorch
Mentioned in GitHub
MIT-HAN-LAB/ProxylessNAS
Official
pytorch
Mentioned in GitHub
songhan/SqueezeNet-Residual
Mentioned in GitHub
songhan/DSD
pytorch
Mentioned in GitHub
lukesin/nas-for-kws-2
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
architecture-search-on-cifar-10-imageProxyless-G + c/o
Params: 5.7M
Percentage error: 2.08
image-classification-on-cifar-10Proxyless-G + c/o
Percentage correct: 97.92
image-classification-on-imagenetProxyless
Number of params: 4.0M
Top 1 Accuracy: 74.6%
neural-architecture-search-on-imagenetProxylesNAS
Accuracy: 75.1
MACs: 581M
Params: 5.1M
Top-1 Error Rate: 24.9

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware | Papers | HyperAI