Command Palette
Search for a command to run...
MnasFPN: Learning Latency-aware Pyramid Architecture for Object Detection on Mobile Devices
Bo Chen Golnaz Ghiasi Hanxiao Liu Tsung-Yi Lin Dmitry Kalenichenko Hartwig Adams Quoc V. Le

Abstract
Despite the blooming success of architecture search for vision tasks in resource-constrained environments, the design of on-device object detection architectures have mostly been manual. The few automated search efforts are either centered around non-mobile-friendly search spaces or not guided by on-device latency. We propose MnasFPN, a mobile-friendly search space for the detection head, and combine it with latency-aware architecture search to produce efficient object detection models. The learned MnasFPN head, when paired with MobileNetV2 body, outperforms MobileNetV3+SSDLite by 1.8 mAP at similar latency on Pixel. It is also both 1.0 mAP more accurate and 10% faster than NAS-FPNLite. Ablation studies show that the majority of the performance gain comes from innovations in the search space. Further explorations reveal an interesting coupling between the search space design and the search algorithm, and that the complexity of MnasFPN search space may be at a local optimum.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| object-detection-on-coco | MnasFPN (MNASNet-B1) | Hardware Burden: Operations per network pass: box mAP: 24.6 |
| object-detection-on-coco | MnasFPN (MobileNetV2) | Hardware Burden: Operations per network pass: box mAP: 26.1 |
| object-detection-on-coco | MnasFPN x0.7 (MobileNetV2) | Hardware Burden: Operations per network pass: box mAP: 23.8 |
| object-detection-on-coco | MnasFPN (MobileNetV3) | Hardware Burden: Operations per network pass: box mAP: 25.5 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.