Command Palette
Search for a command to run...
Rethinking Out-of-distribution (OOD) Detection: Masked Image Modeling is All You Need
Jingyao Li Pengguang Chen Shaozuo Yu Zexin He Shu Liu Jiaya Jia

Abstract
The core of out-of-distribution (OOD) detection is to learn the in-distribution (ID) representation, which is distinguishable from OOD samples. Previous work applied recognition-based methods to learn the ID features, which tend to learn shortcuts instead of comprehensive representations. In this work, we find surprisingly that simply using reconstruction-based methods could boost the performance of OOD detection significantly. We deeply explore the main contributors of OOD detection and find that reconstruction-based pretext tasks have the potential to provide a generally applicable and efficacious prior, which benefits the model in learning intrinsic data distributions of the ID dataset. Specifically, we take Masked Image Modeling as a pretext task for our OOD detection framework (MOOD). Without bells and whistles, MOOD outperforms previous SOTA of one-class OOD detection by 5.7%, multi-class OOD detection by 3.0%, and near-distribution OOD detection by 2.1%. It even defeats the 10-shot-per-class outlier exposure OOD detection, although we do not include any OOD samples for our detection
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| out-of-distribution-detection-on-imagenet-1k-10 | MOOD | AUROC: 91.3 |
| out-of-distribution-detection-on-imagenet-1k-12 | MOOD | AUROC: 89.1 |
| out-of-distribution-detection-on-imagenet-1k-3 | MOOD | AUROC: 86.9 |
| out-of-distribution-detection-on-imagenet-1k-8 | MOOD | AUROC: 89.8 |
| out-of-distribution-detection-on-imagenet-1k-9 | MOOD | AUROC: 88.5 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.