Hear Me Out: Fusional Approaches for Audio Augmented Temporal Action Localization

27 Jun 2021  Â·  Anurag Bagchi, Jazib Mahmood, Dolton Fernandes, Ravi Kiran Sarvadevabhatla ·

State of the art architectures for untrimmed video Temporal Action Localization (TAL) have only considered RGB and Flow modalities, leaving the information-rich audio modality totally unexploited. Audio fusion has been explored for the related but arguably easier problem of trimmed (clip-level) action recognition. However, TAL poses a unique set of challenges. In this paper, we propose simple but effective fusion-based approaches for TAL. To the best of our knowledge, our work is the first to jointly consider audio and video modalities for supervised TAL. We experimentally show that our schemes consistently improve performance for state of the art video-only TAL approaches. Specifically, they help achieve new state of the art performance on large-scale benchmark datasets - ActivityNet-1.3 (54.34 mAP@0.5) and THUMOS14 (57.18 mAP@0.5). Our experiments include ablations involving multiple fusion schemes, modality combinations and TAL architectures. Our code, models and associated data are available at https://github.com/skelemoa/tal-hmo.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Temporal Action Localization ActivityNet-1.3 AVFusion mAP IOU@0.5 54.34 # 10
mAP 36.82 # 11
mAP IOU@0.75 37.66 # 6
mAP IOU@0.95 8.93 # 9
Temporal Action Localization THUMOS'14 AVFusion mAP IOU@0.5 57.18 # 1
Temporal Action Localization THUMOS’14 AVFusion mAP IOU@0.5 57.1 # 13
mAP IOU@0.3 70.1 # 12
mAP IOU@0.4 64.9 # 14
mAP IOU@0.6 45.4 # 17
mAP IOU@0.7 28.8 # 21
Avg mAP (0.3:0.7) 53.3 # 20

Methods


No methods listed for this paper. Add relevant methods here