TokenMixup: Efficient Attention-guided Token-level Data Augmentation for Transformers

14 Oct 2022  ·  Hyeong Kyu Choi, Joonmyung Choi, Hyunwoo J. Kim ·

Mixup is a commonly adopted data augmentation technique for image classification. Recent advances in mixup methods primarily focus on mixing based on saliency. However, many saliency detectors require intense computation and are especially burdensome for parameter-heavy transformer models. To this end, we propose TokenMixup, an efficient attention-guided token-level data augmentation method that aims to maximize the saliency of a mixed set of tokens. TokenMixup provides x15 faster saliency-aware data augmentation compared to gradient-based methods. Moreover, we introduce a variant of TokenMixup which mixes tokens within a single instance, thereby enabling multi-scale feature augmentation. Experiments show that our methods significantly improve the baseline models' performance on CIFAR and ImageNet-1K, while being more efficient than previous methods. We also reach state-of-the-art performance on CIFAR-100 among from-scratch transformer models. Code is available at https://github.com/mlvlab/TokenMixup.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Image Classification CIFAR-10 CCT-7/3x1+VTM Percentage correct 97.78 # 65
Image Classification CIFAR-100 CCT-7/3x1+HTM+VTM Percentage correct 83.57 # 86
Image Classification ImageNet ViT-B/16-224+HTM Top 1 Accuracy 82.37% # 499

Methods