MDMMT-2: Multidomain Multimodal Transformer for Video Retrieval, One More Step Towards Generalization

14 Mar 2022  ยท  Alexander Kunitsyn, Maksim Kalashnikov, Maksim Dzabraev, Andrei Ivaniuta ยท

In this work we present a new State-of-The-Art on the text-to-video retrieval task on MSR-VTT, LSMDC, MSVD, YouCook2 and TGIF obtained by a single model. Three different data sources are combined: weakly-supervised videos, crowd-labeled text-image pairs and text-video pairs. A careful analysis of available pre-trained networks helps to choose the best prior-knowledge ones. We introduce three-stage training procedure that provides high transfer knowledge efficiency and allows to use noisy datasets during training without prior knowledge degradation. Additionally, double positional encoding is used for better fusion of different modalities and a simple method for non-square inputs processing is suggested.

PDF Abstract

Results from the Paper


 Ranked #1 on Video Retrieval on TGIF (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Video Retrieval LSMDC MDMMT-2 text-to-video R@1 26.9 # 13
text-to-video R@5 46.7 # 9
text-to-video R@10 55.9 # 9
text-to-video Median Rank 6.7 # 4
text-to-video Mean Rank 48.0 # 5
Video Retrieval MSR-VTT MDMMT-2 text-to-video R@1 33.7 # 18
text-to-video R@5 60.5 # 17
text-to-video R@10 70.8 # 16
text-to-video Mean Rank 37.8 # 1
text-to-video Median Rank 3.0 # 1
Video Retrieval MSVD MDMMT-2 text-to-video R@1 56.8 # 7
text-to-video R@5 83.1 # 6
text-to-video R@10 89.2 # 5
text-to-video Median Rank 1.0 # 1
text-to-video Mean Rank 8.8 # 7
Video Retrieval TGIF MDMMT-2 text-to-video R@1 25.5 # 1
text-to-video R@5 46.1 # 1
text-to-video R@10 55.7 # 1
text-to-video Mean Rank 94.1 # 1
text-to-video Median Rank 7.0 # 1
Video Retrieval YouCook2 MDMMT-2 text-to-video Median Rank 3.0 # 1
text-to-video R@1 32.0 # 4
text-to-video R@10 74.8 # 3
text-to-video R@5 64.0 # 2
text-to-video Mean Rank 12.7 # 1

Methods


No methods listed for this paper. Add relevant methods here