CLIP-ViP: Adapting Pre-trained Image-Text Model to Video-Language Representation Alignment

14 Sep 2022  ·  Hongwei Xue, Yuchong Sun, Bei Liu, Jianlong Fu, Ruihua Song, Houqiang Li, Jiebo Luo ·

The pre-trained image-text models, like CLIP, have demonstrated the strong power of vision-language representation learned from a large scale of web-collected image-text data. In light of the well-learned visual features, some existing works transfer image representation to video domain and achieve good results. However, how to utilize image-language pre-trained model (e.g., CLIP) for video-language pre-training (post-pretraining) is still under explored. In this paper, we investigate two questions: 1) what are the factors hindering post-pretraining CLIP to further improve the performance on video-language tasks? and 2) how to mitigate the impact of these factors? Through a series of comparative experiments and analyses, we find that the data scale and domain gap between language sources have great impacts. Motivated by these, we propose a Omnisource Cross-modal Learning method equipped with a Video Proxy mechanism on the basis of CLIP, namely CLIP-ViP. Extensive results show that our approach improves the performance of CLIP on video-text retrieval by a large margin. Our model also achieves SOTA results on a variety of datasets, including MSR-VTT, DiDeMo, LSMDC, and ActivityNet. We will release our code and pre-trained CLIP-ViP models at https://github.com/microsoft/XPretrain/tree/main/CLIP-ViP.

PDF Abstract

Results from the Paper


Ranked #2 on Video Retrieval on MSR-VTT-1kA (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Video Retrieval ActivityNet CLIP-ViP text-to-video R@1 61.4 # 8
text-to-video R@5 85.7 # 5
text-to-video R@10 92.6 # 6
text-to-video Median Rank 1 # 1
Video Retrieval DiDeMo CLIP-ViP text-to-video R@1 55.3 # 14
text-to-video R@5 82 # 7
text-to-video R@10 89.3 # 8
text-to-video Median Rank 1 # 1
Video Retrieval LSMDC CLIP-ViP text-to-video R@1 30.7 # 9
text-to-video R@5 51.4 # 6
text-to-video R@10 60.6 # 6
text-to-video Median Rank 5 # 2
Video Retrieval MSR-VTT-1kA CLIP-ViP text-to-video R@1 57.7 # 2
text-to-video R@5 80.5 # 2
text-to-video R@10 88.2 # 3
text-to-video Median Rank 1.0 # 1

Methods