Improved Relation Extraction with Feature-Rich Compositional Embedding Models

EMNLP 2015  ·  Matthew R. Gormley, Mo Yu, Mark Dredze ·

Compositional embedding models build a representation (or embedding) for a linguistic structure based on its component word embeddings. We propose a Feature-rich Compositional Embedding Model (FCM) for relation extraction that is expressive, generalizes to new domains, and is easy-to-implement. The key idea is to combine both (unlexicalized) hand-crafted features with learned word embeddings. The model is able to directly tackle the difficulties met by traditional compositional embeddings models, such as handling arbitrary types of sentence annotations and utilizing global information for composition. We test the proposed model on two relation extraction tasks, and demonstrate that our model outperforms both previous compositional models and traditional feature rich models on the ACE 2005 relation extraction task, and the SemEval 2010 relation classification task. The combination of our model and a log-linear classifier with hand-crafted features gives state-of-the-art results.

PDF Abstract EMNLP 2015 PDF EMNLP 2015 Abstract

Results from the Paper


 Ranked #1 on Relation Extraction on ACE 2005 (Cross Sentence metric)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Relation Extraction ACE 2005 FCM Relation classification F1 58.2 # 6
Cross Sentence No # 1

Methods


No methods listed for this paper. Add relevant methods here