EgoVLPv2: Egocentric Video-Language Pre-training with Fusion in the Backbone

Video-language pre-training (VLP) has become increasingly important due to its ability to generalize to various vision and language tasks. However, existing egocentric VLP frameworks utilize separate video and language encoders and learn task-specific cross-modal information only during fine-tuning, limiting the development of a unified system. In this work, we introduce the second generation of egocentric video-language pre-training (EgoVLPv2), a significant improvement from the previous generation, by incorporating cross-modal fusion directly into the video and language backbones. EgoVLPv2 learns strong video-text representation during pre-training and reuses the cross-modal attention modules to support different downstream tasks in a flexible and efficient manner, reducing fine-tuning costs. Moreover, our proposed fusion in the backbone strategy is more lightweight and compute-efficient than stacking additional fusion-specific layers. Extensive experiments on a wide range of VL tasks demonstrate the effectiveness of EgoVLPv2 by achieving consistent state-of-the-art performance over strong baselines across all downstream. Our project page can be found at https://shramanpramanick.github.io/EgoVLPv2/.

PDF Abstract ICCV 2023 PDF ICCV 2023 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Action Recognition Charades-Ego EgoVLPv2 mAP 34.1 # 2
Moment Queries Ego4D EgoVLPV2 Avg mAP (0.1-0.5) 12.23 # 4
Natural Language Queries Ego4D EgoVLPv2 R@1 IoU=0.3 12.95 # 3
R@5 IoU=0.3 23.80 # 2
R@1 IoU=0.5 7.91 # 4
R@5 IoU=0.5 16.11 # 2
Question Answering EgoTaskQA EgoVLPv2 Direct 46.26 # 1
Multi-Instance Retrieval EPIC-KITCHENS-100 EgoVLPv2 mAP (Avg) 47.3 # 4
nDCG (Avg) 61.9 # 3
Multi-Instance Retrieval EPIC-KITCHENS-100 EgoVLPv2 (Zero-shot) mAP (Avg) 26.7 # 11
nDCG (Avg) 29.1 # 11
Video Summarization Query-Focused Video Summarization Dataset EgoVLPv2 F1 (avg) 52.08 # 1

Methods


No methods listed for this paper. Add relevant methods here