Command Palette
Search for a command to run...
Stanislav Fort Jie Ren Balaji Lakshminarayanan

Abstract
Near out-of-distribution detection (OOD) is a major challenge for deep neural networks. We demonstrate that large-scale pre-trained transformers can significantly improve the state-of-the-art (SOTA) on a range of near OOD tasks across different data modalities. For instance, on CIFAR-100 vs CIFAR-10 OOD detection, we improve the AUROC from 85% (current SOTA) to more than 96% using Vision Transformers pre-trained on ImageNet-21k. On a challenging genomics OOD detection benchmark, we improve the AUROC from 66% to 77% using transformers and unsupervised pre-training. To further improve performance, we explore the few-shot outlier exposure setting where a few examples from outlier classes may be available; we show that pre-trained transformers are particularly well-suited for outlier exposure, and that the AUROC of OOD detection on CIFAR-100 vs CIFAR-10 can be improved to 98.7% with just 1 image per OOD class, and 99.46% with 10 images per OOD class. For multi-modal image-text pre-trained transformers such as CLIP, we explore a new way of using just the names of outlier classes as a sole source of information without any accompanying images, and show that this outperforms previous SOTA on standard vision OOD benchmark tasks.
Code Repositories
Benchmarks
| Benchmark | Methodology | Metrics |
|---|---|---|
| out-of-distribution-detection-on-cifar-10-vs | R+ViT finetuned on CIFAR-10 | AUPR: 97.75 AUROC: 98.52 |
| out-of-distribution-detection-on-cifar-10-vs | ViT finetuned on CIFAR-10 | AUPR: 97.68 AUROC: 98.42 |
| out-of-distribution-detection-on-cifar-10-vs | MLP-Mixer finetuned on CIFAR-10 | AUPR: 96.28 AUROC: 97.85 |
| out-of-distribution-detection-on-cifar-100-vs | Ensemble of ViTs | AUROC: 98.11 |
| out-of-distribution-detection-on-cifar-100-vs | ViT_B-16 finetuned on CIFAR-100 | AUPR: 91.89 AUROC: 95.53 |
| out-of-distribution-detection-on-cifar-100-vs | MLP-Mixer_B-16 finetuned on CIFAR-100 | AUPR: 90.22 AUROC: 95.31 |
| out-of-distribution-detection-on-cifar-100-vs | ViT-L_16 finetuned on CIFAR-100 | AUROC: 97.98 |
| out-of-distribution-detection-on-cifar-100-vs | R50+ViT_B-16 finetuned on CIFAR-100 | AUPR: 92.08 AUROC: 96.23 |
| out-of-distribution-detection-on-cifar-100-vs | CLIP using class name words describing the two distributions | AUROC: 94.68 |
Build AI with AI
From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.