HyperAIHyperAI

Command Palette

Search for a command to run...

4 months ago

sharpDARTS: Faster and More Accurate Differentiable Architecture Search

Andrew Hundt; Varun Jain; Gregory D. Hager

sharpDARTS: Faster and More Accurate Differentiable Architecture Search

Abstract

Neural Architecture Search (NAS) has been a source of dramatic improvements in neural network design, with recent results meeting or exceeding the performance of hand-tuned architectures. However, our understanding of how to represent the search space for neural net architectures and how to search that space efficiently are both still in their infancy. We have performed an in-depth analysis to identify limitations in a widely used search space and a recent architecture search method, Differentiable Architecture Search (DARTS). These findings led us to introduce novel network blocks with a more general, balanced, and consistent design; a better-optimized Cosine Power Annealing learning rate schedule; and other improvements. Our resulting sharpDARTS search is 50% faster with a 20-30% relative improvement in final model error on CIFAR-10 when compared to DARTS. Our best single model run has 1.93% (1.98+/-0.07) validation error on CIFAR-10 and 5.5% error (5.8+/-0.3) on the recently released CIFAR-10.1 test set. To our knowledge, both are state of the art for models of similar size. This model also generalizes competitively to ImageNet at 25.1% top-1 (7.8% top-5) error. We found improvements for existing search spaces but does DARTS generalize to new domains? We propose Differentiable Hyperparameter Grid Search and the HyperCuboid search space, which are representations designed to leverage DARTS for more general parameter optimization. Here we find that DARTS fails to generalize when compared against a human's one shot choice of models. We look back to the DARTS and sharpDARTS search spaces to understand why, and an ablation study reveals an unusual generalization gap. We finally propose Max-W regularization to solve this problem, which proves significantly better than the handmade design. Code will be made available.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
architecture-search-on-cifar-10-imageSharpSepConvDARTS
FLOPS: 579M
Params: 3.6M
Percentage error: 1.98
neural-architecture-search-on-cifar-10sharpDARTS
FLOPS: 357M
Parameters: 1.98M
Search Time (GPU days): 1.8
Top-1 Error Rate: 2.29%
neural-architecture-search-on-cifar-10SharpSepConvDARTS
FLOPS: 579M
Parameters: 3.6M
Search Time (GPU days): 0.8
Top-1 Error Rate: 1.98%
neural-architecture-search-on-imagenetSharpSepConvDARTS
Accuracy: 74.1
MACs: 573M
Params: 4.9M
Top-1 Error Rate: 25.1
neural-architecture-search-on-imagenetsharpDARTS
Accuracy: 76.0
MACs: 950M
Params: 8.3M
Top-1 Error Rate: 24.0

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
sharpDARTS: Faster and More Accurate Differentiable Architecture Search | Papers | HyperAI