How Much Temporal Long-Term Context is Needed for Action Segmentation?

ICCV 2023  ยท  Emad Bahrami, Gianpiero Francesca, Juergen Gall ยท

Modeling long-term context in videos is crucial for many fine-grained tasks including temporal action segmentation. An interesting question that is still open is how much long-term temporal context is needed for optimal performance. While transformers can model the long-term context of a video, this becomes computationally prohibitive for long videos. Recent works on temporal action segmentation thus combine temporal convolutional networks with self-attentions that are computed only for a local temporal window. While these approaches show good results, their performance is limited by their inability to capture the full context of a video. In this work, we try to answer how much long-term temporal context is required for temporal action segmentation by introducing a transformer-based model that leverages sparse attention to capture the full context of a video. We compare our model with the current state of the art on three datasets for temporal action segmentation, namely 50Salads, Breakfast, and Assembly101. Our experiments show that modeling the full context of a video is necessary to obtain the best performance for temporal action segmentation.

PDF Abstract ICCV 2023 PDF ICCV 2023 Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Action Segmentation 50 Salads LTContext F1@10% 89.4 # 4
Edit 83.2 # 7
Acc 87.7 # 6
F1@25% 87.7 # 6
F1@50% 82.0 # 5
Action Segmentation Assembly101 LTContext MoF 41.2 # 1
F1@10% 33.9 # 1
F1@25% 30.0 # 1
F1@50% 22.6 # 1
Edit 30.4 # 5
Action Segmentation Breakfast LTContext F1@10% 77.6 # 7
F1@50% 60.1 # 7
Acc 74.2 # 9
Edit 77.0 # 7
F1@25% 72.6 # 7

Methods


No methods listed for this paper. Add relevant methods here