Similarity Reasoning and Filtration for Image-Text Matching

5 Jan 2021  ·  Haiwen Diao, Ying Zhang, Lin Ma, Huchuan Lu ·

Image-text matching plays a critical role in bridging the vision and language, and great progress has been made by exploiting the global alignment between image and sentence, or local alignments between regions and words. However, how to make the most of these alignments to infer more accurate matching scores is still underexplored. In this paper, we propose a novel Similarity Graph Reasoning and Attention Filtration (SGRAF) network for image-text matching. Specifically, the vector-based similarity representations are firstly learned to characterize the local and global alignments in a more comprehensive manner, and then the Similarity Graph Reasoning (SGR) module relying on one graph convolutional neural network is introduced to infer relation-aware similarities with both the local and global alignments. The Similarity Attention Filtration (SAF) module is further developed to integrate these alignments effectively by selectively attending on the significant and representative alignments and meanwhile casting aside the interferences of non-meaningful alignments. We demonstrate the superiority of the proposed method with achieving state-of-the-art performances on the Flickr30K and MSCOCO datasets, and the good interpretability of SGR and SAF modules with extensive qualitative experiments and analyses.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Cross-Modal Retrieval COCO 2014 SGRAF Image-to-text R@1 57.8 # 24
Image-to-text R@10 91.6 # 22
Image-to-text R@5 84.9 # 23
Text-to-image R@1 41.9 # 27
Text-to-image R@10 81.3 # 25
Text-to-image R@5 70.7 # 26
Cross-Modal Retrieval Flickr30k SGRAF Image-to-text R@1 77.8 # 15
Image-to-text R@10 97.4 # 14
Image-to-text R@5 94.1 # 15
Text-to-image R@1 58.5 # 16
Text-to-image R@10 88.8 # 16
Text-to-image R@5 83.0 # 15
Image Retrieval Flickr30K 1K test SGRAF R@1 58.5 # 3
R@10 88.8 # 5
R@5 83.0 # 5

Methods


No methods listed for this paper. Add relevant methods here