RpBERT: A Text-image Relation Propagation-based BERT Model for Multimodal NER

5 Feb 2021  ·  Lin Sun, Jiquan Wang, Kai Zhang, Yindu Su, Fangsheng Weng ·

Recently multimodal named entity recognition (MNER) has utilized images to improve the accuracy of NER in tweets. However, most of the multimodal methods use attention mechanisms to extract visual clues regardless of whether the text and image are relevant. Practically, the irrelevant text-image pairs account for a large proportion in tweets. The visual clues that are unrelated to the texts will exert uncertain or even negative effects on multimodal model learning. In this paper, we introduce a method of text-image relation propagation into the multimodal BERT model. We integrate soft or hard gates to select visual clues and propose a multitask algorithm to train on the MNER datasets. In the experiments, we deeply analyze the changes in visual attention before and after the use of text-image relation propagation. Our model achieves state-of-the-art performance on the MNER datasets.

PDF Abstract

Datasets


  Add Datasets introduced or used in this paper
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Multi-modal Named Entity Recognition SNAP (MNER) RpBERT F1 87.80 # 6
Multi-modal Named Entity Recognition Twitter-15 RpBERT F1 74.90 # 6

Methods