Shuffle Transformer: Rethinking Spatial Shuffle for Vision Transformer

7 Jun 2021  ·  Zilong Huang, Youcheng Ben, Guozhong Luo, Pei Cheng, Gang Yu, Bin Fu ·

Very recently, Window-based Transformers, which computed self-attention within non-overlapping local windows, demonstrated promising results on image classification, semantic segmentation, and object detection. However, less study has been devoted to the cross-window connection which is the key element to improve the representation ability. In this work, we revisit the spatial shuffle as an efficient way to build connections among windows. As a result, we propose a new vision transformer, named Shuffle Transformer, which is highly efficient and easy to implement by modifying two lines of code. Furthermore, the depth-wise convolution is introduced to complement the spatial shuffle for enhancing neighbor-window connections. The proposed architectures achieve excellent performance on a wide range of visual tasks including image-level classification, object detection, and semantic segmentation. Code will be released for reproduction.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Semantic Segmentation ADE20K UperNet Shuffle-T Validation mIoU 47.6 # 152
Semantic Segmentation ADE20K UperNet Shuffle-B Validation mIoU 50.5 # 104
Semantic Segmentation ADE20K val UperNet Shuffle-B mIoU 50.5 # 45
Semantic Segmentation ADE20K val UperNet Shuffle-S mIoU 49.6 # 53
Semantic Segmentation ADE20K val UperNet Shuffle-T mIoU 47.6 # 61

Methods