MaxViT: Multi-Axis Vision Transformer

4 Apr 2022  ยท  Zhengzhong Tu, Hossein Talebi, Han Zhang, Feng Yang, Peyman Milanfar, Alan Bovik, Yinxiao Li ยท

Transformers have recently gained significant attention in the computer vision community. However, the lack of scalability of self-attention mechanisms with respect to image size has limited their wide adoption in state-of-the-art vision backbones. In this paper we introduce an efficient and scalable attention model we call multi-axis attention, which consists of two aspects: blocked local and dilated global attention. These design choices allow global-local spatial interactions on arbitrary input resolutions with only linear complexity. We also present a new architectural element by effectively blending our proposed attention model with convolutions, and accordingly propose a simple hierarchical vision backbone, dubbed MaxViT, by simply repeating the basic building block over multiple stages. Notably, MaxViT is able to ''see'' globally throughout the entire network, even in earlier, high-resolution stages. We demonstrate the effectiveness of our model on a broad spectrum of vision tasks. On image classification, MaxViT achieves state-of-the-art performance under various settings: without extra data, MaxViT attains 86.5% ImageNet-1K top-1 accuracy; with ImageNet-21K pre-training, our model achieves 88.7% top-1 accuracy. For downstream tasks, MaxViT as a backbone delivers favorable performance on object detection as well as visual aesthetic assessment. We also show that our proposed model expresses strong generative modeling capability on ImageNet, demonstrating the superior potential of MaxViT blocks as a universal vision module. The source code and trained models will be available at https://github.com/google-research/maxvit.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Object Detection COCO 2017 MaxViT-B AP 53.4 # 1
AP50 72.9 # 1
AP75 58.1 # 1
APM 45.7 # 1
APM50 70.3 # 1
APM75 50 # 1
Object Detection COCO 2017 MaxViT-S AP 53.1 # 2
AP50 72.5 # 2
AP75 58.1 # 1
APM 45.4 # 2
APM50 69.8 # 2
APM75 49.5 # 2
Object Detection COCO 2017 MaxViT-T AP 52.1 # 3
AP50 71.9 # 3
AP75 56.8 # 3
APM 44.6 # 3
APM50 69.1 # 3
APM75 48.4 # 3
Image Classification ImageNet MaxViT-B (384res, JFT) Top 1 Accuracy 88.69% # 41
Image Classification ImageNet MaxViT-XL (512res, 21K) Top 1 Accuracy 88.7% # 40
Image Classification ImageNet MaxViT-B (512res, JFT) Top 1 Accuracy 88.82% # 37
Image Classification ImageNet MaxViT-L (512res, JFT) Top 1 Accuracy 89.41% # 29
Image Classification ImageNet MaxViT-XL (384res, JFT) Top 1 Accuracy 89.41% # 29
Image Classification ImageNet MaxViT-XL (512res, JFT) Top 1 Accuracy 89.53% # 27
Image Classification ImageNet MaxViT-B (384res, 21K) Top 1 Accuracy 88.24% # 64
Image Classification ImageNet MaxViT-L (384res, 21K) Top 1 Accuracy 88.32% # 61
Image Classification ImageNet MaxViT-B (512res, 21K) Top 1 Accuracy 88.38% # 57
Image Classification ImageNet MaxViT-L (512res, 21K) Top 1 Accuracy 88.46% # 54
Image Classification ImageNet MaxViT-XL (384res, 21K) Top 1 Accuracy 88.51% # 49
Image Classification ImageNet MaxViT-L (384res, JFT) Top 1 Accuracy 89.12% # 32
Image Classification ImageNet MaxViT-T(512res) Top 1 Accuracy 85.72% # 199
Image Classification ImageNet MaxViT-L (384res) Top 1 Accuracy 86.4% # 143
Image Classification ImageNet MaxViT-T (384res) Top 1 Accuracy 85.24% # 237
Image Classification ImageNet MaxViT-S (224res) Top 1 Accuracy 84.45% # 298
Image Classification ImageNet MaxViT-B (224res) Top 1 Accuracy 84.95% # 263
Image Classification ImageNet MaxViT-T (224res) Top 1 Accuracy 83.62% # 376
Image Classification ImageNet MaxViT-L (512res) Top 1 Accuracy 86.7% # 126
Image Classification ImageNet MaxViT-S (512res) Top 1 Accuracy 86.19% # 169
Image Classification ImageNet MaxViT-B (384res) Top 1 Accuracy 86.34% # 152

Methods


No methods listed for this paper. Add relevant methods here