HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

An Effective Baseline for Robustness to Distributional Shift

Sunil Thulasidasan Sushil Thapa Sayera Dhaubhadel Gopinath Chennupati Tanmoy Bhattacharya Jeff Bilmes

An Effective Baseline for Robustness to Distributional Shift

Abstract

Refraining from confidently predicting when faced with categories of inputs different from those seen during training is an important requirement for the safe deployment of deep learning systems. While simple to state, this has been a particularly challenging problem in deep learning, where models often end up making overconfident predictions in such situations. In this work we present a simple, but highly effective approach to deal with out-of-distribution detection that uses the principle of abstention: when encountering a sample from an unseen class, the desired behavior is to abstain from predicting. Our approach uses a network with an extra abstention class and is trained on a dataset that is augmented with an uncurated set that consists of a large number of out-of-distribution (OoD) samples that are assigned the label of the abstention class; the model is then trained to learn an effective discriminator between in and out-of-distribution samples. We compare this relatively simple approach against a wide variety of more complex methods that have been proposed both for out-of-distribution detection as well as uncertainty modeling in deep learning, and empirically demonstrate its effectiveness on a wide variety of of benchmarks and deep architectures for image recognition and text classification, often outperforming existing approaches by significant margins. Given the simplicity and effectiveness of this method, we propose that this approach be used as a new additional baseline for future work in this domain.

Code Repositories

Sushil-Thapa/Abstention-OoD
Official
pytorch
Mentioned in GitHub

Benchmarks

BenchmarkMethodologyMetrics
out-of-distribution-detection-on-202-Layered GRU
AUROC: 99.6
FPR95: 1.78
out-of-distribution-detection-on-cifar-10Wide ResNet 40x2
AUROC: 99.9
FPR95: 2.0
out-of-distribution-detection-on-cifar-100Wide ResNet 40x2
AUROC: 97.7
FPR95: 23.4
out-of-distribution-detection-on-sst2-Layered GRU
AUROC: 99.7
FPR95: 20.9
out-of-distribution-detection-on-trec-news2-Layered GRU
AUROC: 99.9
FPR95: 4.7

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
An Effective Baseline for Robustness to Distributional Shift | Papers | HyperAI