Deep Anomaly Detection with Outlier Exposure

It is important to detect anomalous inputs when deploying machine learning systems. The use of larger and more complex inputs in deep learning magnifies the difficulty of distinguishing between anomalous and in-distribution examples. At the same time, diverse image and text data are available in enormous quantities. We propose leveraging these data to improve deep anomaly detection by training anomaly detectors against an auxiliary dataset of outliers, an approach we call Outlier Exposure (OE). This enables anomaly detectors to generalize and detect unseen anomalies. In extensive experiments on natural language processing and small- and large-scale vision tasks, we find that Outlier Exposure significantly improves detection performance. We also observe that cutting-edge generative models trained on CIFAR-10 may assign higher likelihoods to SVHN images than to CIFAR-10 images; we use OE to mitigate this issue. We also analyze the flexibility and robustness of Outlier Exposure, and identify characteristics of the auxiliary dataset that improve performance.

PDF Abstract ICLR 2019 PDF ICLR 2019 Abstract

Results from the Paper


Ranked #3 on Out-of-Distribution Detection on CIFAR-100 (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Out-of-Distribution Detection CIFAR-10 WRN 40-2 + OE FPR95 9.50 # 5
AUROC 97.8 # 8
Out-of-Distribution Detection CIFAR-10 WRN 40-2 (MSP Baseline) FPR95 34.94 # 7
AUROC 97.8 # 8
Out-of-Distribution Detection CIFAR-100 WRN 40-2 + OE FPR95 38.50 # 3
Out-of-Distribution Detection CIFAR-100 WRN 40-2 (MSP Baseline) FPR95 62.66 # 4
Out-of-Distribution Detection CIFAR-100 vs SVHN OE AUROC 86.9 # 5
Out-of-Distribution Detection CIFAR-10 vs CIFAR-100 WRN 40-2 + OE AUPR 76.2 # 6
AUROC 93.3 # 9

Methods