GRAtt-VIS: Gated Residual Attention for Auto Rectifying Video Instance Segmentation

Recent trends in Video Instance Segmentation (VIS) have seen a growing reliance on online methods to model complex and lengthy video sequences. However, the degradation of representation and noise accumulation of the online methods, especially during occlusion and abrupt changes, pose substantial challenges. Transformer-based query propagation provides promising directions at the cost of quadratic memory attention. However, they are susceptible to the degradation of instance features due to the above-mentioned challenges and suffer from cascading effects. The detection and rectification of such errors remain largely underexplored. To this end, we introduce \textbf{GRAtt-VIS}, \textbf{G}ated \textbf{R}esidual \textbf{Att}ention for \textbf{V}ideo \textbf{I}nstance \textbf{S}egmentation. Firstly, we leverage a Gumbel-Softmax-based gate to detect possible errors in the current frame. Next, based on the gate activation, we rectify degraded features from its past representation. Such a residual configuration alleviates the need for dedicated memory and provides a continuous stream of relevant instance features. Secondly, we propose a novel inter-instance interaction using gate activation as a mask for self-attention. This masking strategy dynamically restricts the unrepresentative instance queries in the self-attention and preserves vital information for long-term tracking. We refer to this novel combination of Gated Residual Connection and Masked Self-Attention as \textbf{GRAtt} block, which can easily be integrated into the existing propagation-based framework. Further, GRAtt blocks significantly reduce the attention overhead and simplify dynamic temporal modeling. GRAtt-VIS achieves state-of-the-art performance on YouTube-VIS and the highly challenging OVIS dataset, significantly improving over previous methods. Code is available at \url{https://github.com/Tanveer81/GRAttVIS}.

PDF Abstract

Results from the Paper


Ranked #4 on Video Instance Segmentation on YouTube-VIS 2021 (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Video Instance Segmentation OVIS validation GRAtt-VIS (ResNet-50) mask AP 36.2 # 21
AP50 60.8 # 19
AP75 36.8 # 21
AR1 16.8 # 14
AR10 40.1 # 17
Video Instance Segmentation OVIS validation GRAtt-VIS (Swin-L) mask AP 45.7 # 9
AP50 69.1 # 9
AP75 47.8 # 8
AR1 19.2 # 6
AR10 49.4 # 8
Video Instance Segmentation YouTube-VIS 2021 GRAtt-VIS (Swin-L) mask AP 60.3 # 4
AP50 81.3 # 7
AP75 67.1 # 6
AR10 64.5 # 7
AR1 48.8 # 3
Video Instance Segmentation YouTube-VIS 2021 GRAtt-VIS (ResNet-50) mask AP 48.9 # 19
AP50 69.2 # 21
AP75 53.1 # 20
AR10 56.0 # 18
AR1 41.8 # 19

Methods