Improving Relation Extraction through Syntax-induced Pre-training with Dependency Masking

Findings (ACL) 2022  ·  Yuanhe Tian, Yan Song, Fei Xia ·

Relation extraction (RE) is an important natural language processing task that predicts the relation between two given entities, where a good understanding of the contextual information is essential to achieve an outstanding model performance. Among different types of contextual information, the auto-generated syntactic information (namely, word dependencies) has shown its effectiveness for the task. However, most existing studies require modifications to the existing baseline architectures (e.g., adding new components, such as GCN, on the top of an encoder) to leverage the syntactic information. To offer an alternative solution, we propose to leverage syntactic information to improve RE by training a syntax-induced encoder on auto-parsed data through dependency masking. Specifically, the syntax-induced encoder is trained by recovering the masked dependency connections and types in first, second, and third orders, which significantly differs from existing studies that train language models or word embeddings by predicting the context words along the dependency paths. Experimental results on two English benchmark datasets, namely, ACE2005EN and SemEval 2010 Task 8 datasets, demonstrate the effectiveness of our approach for RE, where our approach outperforms strong baselines and achieve state-of-the-art results on both datasets.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Relation Extraction SemEval-2010 Task-8 RE-DMP + XLNet F1 89.90 # 11

Methods


No methods listed for this paper. Add relevant methods here