MixFormer: End-to-End Tracking with Iterative Mixed Attention

6 Feb 2023  ยท  Yutao Cui, Cheng Jiang, Gangshan Wu, LiMin Wang ยท

Visual object tracking often employs a multi-stage pipeline of feature extraction, target information integration, and bounding box estimation. To simplify this pipeline and unify the process of feature extraction and target information integration, in this paper, we present a compact tracking framework, termed as MixFormer, built upon transformers. Our core design is to utilize the flexibility of attention operations, and propose a Mixed Attention Module (MAM) for simultaneous feature extraction and target information integration. This synchronous modeling scheme allows to extract target-specific discriminative features and perform extensive communication between target and search area. Based on MAM, we build our MixFormer trackers simply by stacking multiple MAMs and placing a localization head on top. Specifically, we instantiate two types of MixFormer trackers, a hierarchical tracker MixCvT, and a non-hierarchical tracker MixViT. For these two trackers, we investigate a series of pre-training methods and uncover the different behaviors between supervised pre-training and self-supervised pre-training in our MixFormer trackers. We also extend the masked pre-training to our MixFormer trackers and design the competitive TrackMAE pre-training technique. Finally, to handle multiple target templates during online tracking, we devise an asymmetric attention scheme in MAM to reduce computational cost, and propose an effective score prediction module to select high-quality templates. Our MixFormer trackers set a new state-of-the-art performance on seven tracking benchmarks, including LaSOT, TrackingNet, VOT2020, GOT-10k, OTB100 and UAV123. In particular, our MixViT-L achieves AUC score of 73.3% on LaSOT, 86.1% on TrackingNet, EAO of 0.584 on VOT2020, and AO of 75.7% on GOT-10k. Code and trained models are publicly available at https://github.com/MCG-NJU/MixFormer.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Visual Object Tracking GOT-10k MixViT-L(ConvMAE) Average Overlap 75.7 # 7
Success Rate 0.5 85.3 # 7
Success Rate 0.75 75.1 # 3
Visual Object Tracking LaSOT MixViT-L(ConvMAE) AUC 73.3 # 3
Normalized Precision 82.8 # 1
Precision 80.3 # 2
Visual Object Tracking TrackingNet MixViT-L(ConvMAE) Precision 86.0 # 3
Normalized Precision 90.3 # 2
Accuracy 86.1 # 1

Methods


No methods listed for this paper. Add relevant methods here