Input complexity and out-of-distribution detection with likelihood-based generative models

Likelihood-based generative models are a promising resource to detect out-of-distribution (OOD) inputs which could compromise the robustness or reliability of a machine learning system. However, likelihoods derived from such models have been shown to be problematic for detecting certain types of inputs that significantly differ from training data. In this paper, we pose that this problem is due to the excessive influence that input complexity has in generative models' likelihoods. We report a set of experiments supporting this hypothesis, and use an estimate of input complexity to derive an efficient and parameter-free OOD score, which can be seen as a likelihood-ratio, akin to Bayesian model comparison. We find such score to perform comparably to, or even better than, existing OOD detection approaches under a wide range of data sets, models, model sizes, and complexity estimates.

PDF Abstract ICLR 2020 PDF ICLR 2020 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Anomaly Detection Unlabeled CIFAR-10 vs CIFAR-100 Input Complexity (Glow) AUROC 73.6 # 10
Anomaly Detection Unlabeled CIFAR-10 vs CIFAR-100 Likelihood (PixelCNN++) AUROC 52.6 # 13
Anomaly Detection Unlabeled CIFAR-10 vs CIFAR-100 Likelihood (Glow) AUROC 58.2 # 11
Anomaly Detection Unlabeled CIFAR-10 vs CIFAR-100 Input Complexity (PixelCNN++) AUROC 53.5 # 12

Methods


No methods listed for this paper. Add relevant methods here