HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

MicroNets: Neural Network Architectures for Deploying TinyML Applications on Commodity Microcontrollers

Colby Banbury Chuteng Zhou Igor Fedorov Ramon Matas Navarro Urmish Thakker Dibakar Gope Vijay Janapa Reddi Matthew Mattina Paul N. Whatmough

MicroNets: Neural Network Architectures for Deploying TinyML Applications on Commodity Microcontrollers

Abstract

Executing machine learning workloads locally on resource constrained microcontrollers (MCUs) promises to drastically expand the application space of IoT. However, so-called TinyML presents severe technical challenges, as deep neural network inference demands a large compute and memory budget. To address this challenge, neural architecture search (NAS) promises to help design accurate ML models that meet the tight MCU memory, latency and energy constraints. A key component of NAS algorithms is their latency/energy model, i.e., the mapping from a given neural network architecture to its inference latency/energy on an MCU. In this paper, we observe an intriguing property of NAS search spaces for MCU model design: on average, model latency varies linearly with model operation (op) count under a uniform prior over models in the search space. Exploiting this insight, we employ differentiable NAS (DNAS) to search for models with low memory usage and low op count, where op count is treated as a viable proxy to latency. Experimental results validate our methodology, yielding our MicroNet models, which we deploy on MCUs using Tensorflow Lite Micro, a standard open-source NN inference runtime widely used in the TinyML community. MicroNets demonstrate state-of-the-art results for all three TinyMLperf industry-standard benchmark tasks: visual wake words, audio keyword spotting, and anomaly detection. Models and training scripts can be found at github.com/ARM-software/ML-zoo.

Code Repositories

Benchmarks

BenchmarkMethodologyMetrics
keyword-spotting-on-google-speech-commandsMicroNet-KWS-L
Google Speech Commands V2 12: 95.3
keyword-spotting-on-google-speech-commands-v2-1MicroNet-KWS-L
Accuracy: 95.3
Latency (STM32F746ZG): 0.610128

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
MicroNets: Neural Network Architectures for Deploying TinyML Applications on Commodity Microcontrollers | Papers | HyperAI