Measuring the Interpretability of Unsupervised Representations via Quantized Reversed Probing

ICLR 2022  ·  Iro Laina, Yuki M Asano, Andrea Vedaldi ·

Self-supervised visual representation learning has attracted significant research interest. While the most common way to evaluate self-supervised representations is through transfer to various downstream tasks, we instead investigate the problem of measuring their interpretability, i.e. understanding the semantics encoded in the raw representations. We formulate the latter as estimating the mutual information between the representation and a space of manually labelled concepts. To quantify this we introduce a decoding bottleneck: information must be captured by simple predictors, mapping concepts to clusters of data formed in representation space. This approach, which we call reverse linear probing, provides a single number sensitive to the semanticity of the representation. This measure is also able to detect when the representation correlates with combinations of labelled concepts (e.g. "red apple") instead of just individual attributes ("red" and "apple" separately). Finally, we also suggest that supervised classifiers can be used to automatically label large datasets with a rich space of attributes. We use these insights to evaluate a large number of self-supervised representations, ranking them by interpretability, and highlight the differences that emerge compared to the standard evaluation with linear probes.

PDF Abstract

Results from the Paper


Ranked #90 on Image Classification on ObjectNet (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Benchmark
Image Classification ObjectNet SeLa(v2) (reverse linear probing) Top-5 Accuracy 48.83 # 24
Top-1 Accuracy 20.61 # 90
Image Classification ObjectNet DeepCluster(v2) (reverse linear probing) Top-5 Accuracy 46.81 # 29
Top-1 Accuracy 19.73 # 92
Image Classification ObjectNet OBoW (reverse linear probing) Top-5 Accuracy 31.72 # 32
Top-1 Accuracy 12.23 # 103
Image Classification ObjectNet SwAV (reverse linear probing) Top-5 Accuracy 43.64 # 30
Top-1 Accuracy 17.71 # 95
Image Classification ObjectNet MoCHi (reverse linear probing) Top-5 Accuracy 31.71 # 33
Top-1 Accuracy 12.64 # 102
Image Classification ObjectNet MoCo(v2) (reverse linear probing) Top-5 Accuracy 31.45 # 34
Top-1 Accuracy 12.67 # 101

Methods


No methods listed for this paper. Add relevant methods here