UNITER: UNiversal Image-TExt Representation Learning

Joint image-text embedding is the bedrock for most Vision-and-Language (V+L) tasks, where multimodality inputs are simultaneously processed for joint visual and textual understanding. In this paper, we introduce UNITER, a UNiversal Image-TExt Representation, learned through large-scale pre-training over four image-text datasets (COCO, Visual Genome, Conceptual Captions, and SBU Captions), which can power heterogeneous downstream V+L tasks with joint multimodal embeddings. We design four pre-training tasks: Masked Language Modeling (MLM), Masked Region Modeling (MRM, with three variants), Image-Text Matching (ITM), and Word-Region Alignment (WRA). Different from previous work that applies joint random masking to both modalities, we use conditional masking on pre-training tasks (i.e., masked language/region modeling is conditioned on full observation of image/text). In addition to ITM for global image-text alignment, we also propose WRA via the use of Optimal Transport (OT) to explicitly encourage fine-grained alignment between words and image regions during pre-training. Comprehensive analysis shows that both conditional masking and OT-based WRA contribute to better pre-training. We also conduct a thorough ablation study to find an optimal combination of pre-training tasks. Extensive experiments show that UNITER achieves new state of the art across six V+L tasks (over nine datasets), including Visual Question Answering, Image-Text Retrieval, Referring Expression Comprehension, Visual Commonsense Reasoning, Visual Entailment, and NLVR$^2$. Code is available at https://github.com/ChenRocks/UNITER.

PDF Abstract ECCV 2020 PDF ECCV 2020 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Zero-Shot Cross-Modal Retrieval Flickr30k UNITER Image-to-text R@1 80.7 # 8
Image-to-text R@5 95.7 # 8
Image-to-text R@10 98.0 # 7
Text-to-image R@1 66.2 # 8
Text-to-image R@5 88.4 # 8
Text-to-image R@10 92.9 # 7
Visual Reasoning NLVR2 Test UNITER (Large) Accuracy 79.5 # 6
Referring Expression Comprehension RefCOCO UNITER-L Val 81.41 # 5
Test A 87.04 # 5
Test B 74.17 # 6
Visual Entailment SNLI-VE test UNITER (Large) Accuracy 78.98 # 5
Visual Entailment SNLI-VE val UNITER Accuracy 78.98 # 7
Visual Question Answering VCR (Q-AR) test UNITER (Large) Accuracy 62.8 # 2
Visual Question Answering VCR (QA-R) test UNITER-large (ensemble of 10 models) Accuracy 83.4 # 2
Visual Question Answering VCR (QA-R) test UNITER (Large) Accuracy 80.8 # 3
Visual Question Answering VCR (Q-A) test UNITER-large (10 ensemble) Accuracy 79.8 # 2
Visual Question Answering VCR (Q-A) test UNITER (Large) Accuracy 77.3 # 3
Visual Question Answering VQA v2 test-dev UNITER (Large) Accuracy 73.24 # 9
Visual Question Answering VQA v2 test-std UNITER (Large) overall 73.4 # 24