MAR: Masked Autoencoders for Efficient Action Recognition

24 Jul 2022  ·  Zhiwu Qing, Shiwei Zhang, Ziyuan Huang, Xiang Wang, Yuehuan Wang, Yiliang Lv, Changxin Gao, Nong Sang ·

Standard approaches for video recognition usually operate on the full input videos, which is inefficient due to the widely present spatio-temporal redundancy in videos. Recent progress in masked video modelling, i.e., VideoMAE, has shown the ability of vanilla Vision Transformers (ViT) to complement spatio-temporal contexts given only limited visual contents. Inspired by this, we propose propose Masked Action Recognition (MAR), which reduces the redundant computation by discarding a proportion of patches and operating only on a part of the videos. MAR contains the following two indispensable components: cell running masking and bridging classifier. Specifically, to enable the ViT to perceive the details beyond the visible patches easily, cell running masking is presented to preserve the spatio-temporal correlations in videos, which ensures the patches at the same spatial location can be observed in turn for easy reconstructions. Additionally, we notice that, although the partially observed features can reconstruct semantically explicit invisible patches, they fail to achieve accurate classification. To address this, a bridging classifier is proposed to bridge the semantic gap between the ViT encoded features for reconstruction and the features specialized for classification. Our proposed MAR reduces the computational cost of ViT by 53% and extensive experiments show that MAR consistently outperforms existing ViT models with a notable margin. Especially, we found a ViT-Large trained by MAR outperforms the ViT-Huge trained by a standard training scheme by convincing margins on both Kinetics-400 and Something-Something v2 datasets, while our computation overhead of ViT-Large is only 14.5% of ViT-Huge.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Action Classification Kinetics-400 MAR (50% mask, ViT-L, 16x4) Acc@1 85.3 # 49
Acc@5 96.3 # 38
Action Classification Kinetics-400 MAR (75% mask, ViT-B, 16x4) Acc@1 79.4 # 103
Acc@5 93.7 # 86
Action Classification Kinetics-400 MAR (50% mask, ViT-B, 16x4) Acc@1 81.0 # 82
Acc@5 94.4 # 70
Action Classification Kinetics-400 MAR (75% mask, ViT-L, 16x4) Acc@1 83.9 # 57
Acc@5 96.0 # 41
Action Recognition Something-Something V2 MAR (50% mask, ViT-L, 16x4) Top-1 Accuracy 74.7 # 12
Top-5 Accuracy 94.9 # 7
Parameters 311 # 14
GFLOPs 276x6 # 6
Action Recognition Something-Something V2 MAR (75% mask, ViT-B, 16x4) Top-1 Accuracy 69.5 # 44
Top-5 Accuracy 91.9 # 31
Parameters 94 # 21
GFLOPs 41x6 # 6
Action Recognition Something-Something V2 MAR (50% mask, ViT-B, 16x4) Top-1 Accuracy 71.0 # 32
Top-5 Accuracy 92.8 # 20
Parameters 94 # 21
GFLOPs 86x6 # 6
Action Recognition Something-Something V2 MAR (75% mask, ViT-L, 16x4) Top-1 Accuracy 73.8 # 16
Top-5 Accuracy 94.4 # 10
Parameters 311 # 14
GFLOPs 131x6 # 6

Methods


No methods listed for this paper. Add relevant methods here