Video-Text as Game Players: Hierarchical Banzhaf Interaction for Cross-Modal Representation Learning

Contrastive learning-based video-language representation learning approaches, e.g., CLIP, have achieved outstanding performance, which pursue semantic interaction upon pre-defined video-text pairs. To clarify this coarse-grained global interaction and move a step further, we have to encounter challenging shell-breaking interactions for fine-grained cross-modal learning. In this paper, we creatively model video-text as game players with multivariate cooperative game theory to wisely handle the uncertainty during fine-grained semantic interaction with diverse granularity, flexible combination, and vague intensity. Concretely, we propose Hierarchical Banzhaf Interaction (HBI) to value possible correspondence between video frames and text words for sensitive and explainable cross-modal contrast. To efficiently realize the cooperative game of multiple video frames and multiple text words, the proposed method clusters the original video frames (text words) and computes the Banzhaf Interaction between the merged tokens. By stacking token merge modules, we achieve cooperative games at different semantic levels. Extensive experiments on commonly used text-video retrieval and video-question answering benchmarks with superior performances justify the efficacy of our HBI. More encouragingly, it can also serve as a visualization tool to promote the understanding of cross-modal interaction, which have a far-reaching impact on the community. Project page is available at https://jpthu17.github.io/HBI/.

PDF Abstract CVPR 2023 PDF CVPR 2023 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Video Retrieval ActivityNet HBI text-to-video R@1 42.2 # 21
text-to-video R@5 73.0 # 18
text-to-video R@10 84.6 # 16
text-to-video Median Rank 2.0 # 5
text-to-video Mean Rank 6.6 # 8
video-to-text R@1 42.4 # 12
video-to-text R@5 73.0 # 10
video-to-text R@10 86.0 # 7
video-to-text Median Rank 2.0 # 2
video-to-text Mean Rank 6.5 # 7
Video Retrieval DiDeMo HBI text-to-video R@1 46.9 # 28
text-to-video R@5 74.9 # 26
text-to-video R@10 82.7 # 25
text-to-video Median Rank 2.0 # 9
text-to-video Mean Rank 12.1 # 5
video-to-text R@1 46.2 # 12
video-to-text R@10 82.7 # 9
video-to-text Median Rank 2.0 # 5
video-to-text Mean Rank 8.7 # 4
video-to-text R@5 73.0 # 9
Video Retrieval MSR-VTT-1kA HBI text-to-video Mean Rank 12.0 # 7
text-to-video R@1 48.6 # 22
text-to-video R@5 74.6 # 20
text-to-video R@10 83.4 # 21
text-to-video Median Rank 2.0 # 10
video-to-text R@1 46.8 # 17
video-to-text R@5 74.3 # 12
video-to-text R@10 84.3 # 11
video-to-text Median Rank 2.0 # 7
video-to-text Mean Rank 8.9 # 12
Visual Question Answering (VQA) MSRVTT-QA HBI Accuracy 0.462 # 11
Video Question Answering MSRVTT-QA HBI Accuracy 46.2 # 8

Methods