The foundation models have recently shown excellent performance on a variety of downstream tasks in computer vision. However, most existing vision foundation models simply focus on image-level pretraining and adpation, which are limited for dynamic and complex video-level understanding tasks. To fill the gap, we present general video foundation models, InternVideo, by taking advantage of both generative and discriminative self-supervised video learning. Specifically, InternVideo efficiently explores masked video modeling and video-language contrastive learning as the pretraining objectives, and selectively coordinates video representations of these two complementary frameworks in a learnable manner to boost various video applications. Without bells and whistles, InternVideo achieves state-of-the-art performance on 39 video datasets from extensive tasks including video action recognition/detection, video-language alignment, and open-world video applications. Especially, our methods can obtain 91.1% and 77.2% top-1 accuracy on the challenging Kinetics-400 and Something-Something V2 benchmarks, respectively. All of these results effectively show the generality of our InternVideo for video understanding. The code will be released at https://github.com/OpenGVLab/InternVideo.
Paper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Video Question Answering | 2 | 7.69% |
Video Understanding | 2 | 7.69% |
Denoising | 1 | 3.85% |
Language Modelling | 1 | 3.85% |
Large Language Model | 1 | 3.85% |
Sentence | 1 | 3.85% |
Speech Recognition | 1 | 3.85% |
Question Answering | 1 | 3.85% |
Text Retrieval | 1 | 3.85% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |