RTQ: Rethinking Video-language Understanding Based on Image-text Model

1 Dec 2023  ·  Xiao Wang, Yaoyu Li, Tian Gan, Zheng Zhang, Jingjing Lv, Liqiang Nie ·

Recent advancements in video-language understanding have been established on the foundation of image-text models, resulting in promising outcomes due to the shared knowledge between images and videos. However, video-language understanding presents unique challenges due to the inclusion of highly complex semantic details, which result in information redundancy, temporal dependency, and scene complexity. Current techniques have only partially tackled these issues, and our quantitative analysis indicates that some of these methods are complementary. In light of this, we propose a novel framework called RTQ (Refine, Temporal model, and Query), which addresses these challenges simultaneously. The approach involves refining redundant information within frames, modeling temporal relations among frames, and querying task-specific information from the videos. Remarkably, our model demonstrates outstanding performance even in the absence of video-language pre-training, and the results are comparable with or superior to those achieved by state-of-the-art pre-training methods. Code is available at https://github.com/SCZwangxiao/RTQ-MM2023.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Video Retrieval ActivityNet RTQ text-to-video R@1 53.5 # 12
text-to-video R@5 81.4 # 7
text-to-video R@10 91.9 # 7
Video Retrieval DiDeMo RTQ text-to-video R@1 57.6 # 10
text-to-video R@5 84.1 # 7
text-to-video R@10 89.9 # 7
Video Captioning MSR-VTT RTQ CIDEr 69.3 # 9
ROUGE-L 66.1 # 6
BLEU-4 49.6 # 8
Video Retrieval MSR-VTT-1kA RTQ text-to-video R@1 53.4 # 9
text-to-video R@5 76.1 # 13
text-to-video R@10 84.4 # 15
Video Captioning MSVD RTQ CIDEr 123.4 # 9
BLEU-4 66.9 # 6
ROUGE-L 82.2 # 4
Video Question Answering NExT-QA RTQ Accuracy 63.2 # 10

Methods


No methods listed for this paper. Add relevant methods here