Self-Supervised Aggregation of Diverse Experts for Test-Agnostic Long-Tailed Recognition

20 Jul 2021  ·  Yifan Zhang, Bryan Hooi, Lanqing Hong, Jiashi Feng ·

Existing long-tailed recognition methods, aiming to train class-balanced models from long-tailed data, generally assume the models would be evaluated on the uniform test class distribution. However, practical test class distributions often violate this assumption (e.g., being either long-tailed or even inversely long-tailed), which may lead existing methods to fail in real applications. In this paper, we study a more practical yet challenging task, called test-agnostic long-tailed recognition, where the training class distribution is long-tailed while the test class distribution is agnostic and not necessarily uniform. In addition to the issue of class imbalance, this task poses another challenge: the class distribution shift between the training and test data is unknown. To tackle this task, we propose a novel approach, called Self-supervised Aggregation of Diverse Experts, which consists of two strategies: (i) a new skill-diverse expert learning strategy that trains multiple experts from a single and stationary long-tailed dataset to separately handle different class distributions; (ii) a novel test-time expert aggregation strategy that leverages self-supervision to aggregate the learned multiple experts for handling unknown test class distributions. We theoretically show that our self-supervised strategy has a provable ability to simulate test-agnostic class distributions. Promising empirical results demonstrate the effectiveness of our method on both vanilla and test-agnostic long-tailed recognition. Code is available at \url{https://github.com/Vanint/SADE-AgnosticLT}.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Long-tail Learning CIFAR-100-LT (ρ=10) TADE Error Rate 36.4 # 16
Long-tail Learning CIFAR-100-LT (ρ=100) TADE Error Rate 50.2 # 28
Long-tail Learning CIFAR-100-LT (ρ=50) TADE Error Rate 46.1 # 17
Long-tail Learning CIFAR-10-LT (ρ=10) RIDE Error Rate 10.3 # 22
Long-tail Learning CIFAR-10-LT (ρ=10) TADE Error Rate 9.2 # 15
Long-tail Learning CIFAR-10-LT (ρ=100) TADE Error Rate 16.2 # 12
Long-tail Learning ImageNet-LT TADE(ResNeXt101-32x4d) Top-1 Accuracy 61.4 # 12
Long-tail Learning ImageNet-LT TADE(ResNeXt-50) Top-1 Accuracy 58.8 # 16
Image Classification iNaturalist 2018 TADE (ResNet-50) Top-1 Accuracy 72.9% # 28
Long-tail Learning iNaturalist 2018 TADE Top-1 Accuracy 72.9% # 20
Long-tail Learning iNaturalist 2018 TADE(ResNet-152) Top-1 Accuracy 77% # 7
Long-tail Learning Places-LT TADE Top-1 Accuracy 41.3 # 13
Top 1 Accuracy 40.9 # 1

Methods


No methods listed for this paper. Add relevant methods here