Multi-modal Text Recognition Networks: Interactive Enhancements between Visual and Semantic Features

30 Nov 2021  ·  Byeonghu Na, Yoonsik Kim, Sungrae Park ·

Linguistic knowledge has brought great benefits to scene text recognition by providing semantics to refine character sequences. However, since linguistic knowledge has been applied individually on the output sequence, previous methods have not fully utilized the semantics to understand visual clues for text recognition. This paper introduces a novel method, called Multi-modAl Text Recognition Network (MATRN), that enables interactions between visual and semantic features for better recognition performances. Specifically, MATRN identifies visual and semantic feature pairs and encodes spatial information into semantic features. Based on the spatial encoding, visual and semantic features are enhanced by referring to related features in the other modality. Furthermore, MATRN stimulates combining semantic features into visual features by hiding visual clues related to the character in the training phase. Our experiments demonstrate that MATRN achieves state-of-the-art performances on seven benchmarks with large margins, while naive combinations of two modalities show less-effective improvements. Further ablative studies prove the effectiveness of our proposed components. Our implementation is available at https://github.com/wp03052/MATRN.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Scene Text Recognition CUTE80 MATRN Accuracy 93.5 # 12
Scene Text Recognition ICDAR2013 MATRN Accuracy 97.9 # 9
Scene Text Recognition ICDAR2015 MATRN Accuracy 86.6 # 10
Scene Text Recognition IIIT5k MATRN Accuracy 96.6 # 13
Scene Text Recognition SVT MATRN Accuracy 95 # 13
Scene Text Recognition SVTP MATRN Accuracy 90.6 # 11

Methods


No methods listed for this paper. Add relevant methods here