Coreference Resolution

197 papers with code • 11 benchmarks • 41 datasets

Coreference resolution is the task of clustering mentions in text that refer to the same underlying real world entities.

Example:

               +-----------+
               |           |
I voted for Obama because he was most aligned with my values", she said.
 |                                                 |            |
 +-------------------------------------------------+------------+

"I", "my", and "she" belong to the same cluster and "Obama" and "he" belong to the same cluster.

Libraries

Use these libraries to find Coreference Resolution models and implementations

Most implemented papers

Deep contextualized word representations

flairNLP/flair NAACL 2018

We introduce a new type of deep contextualized word representation that models both (1) complex characteristics of word use (e. g., syntax and semantics), and (2) how these uses vary across linguistic contexts (i. e., to model polysemy).

Language Models are Few-Shot Learners

openai/gpt-3 NeurIPS 2020

By contrast, humans can generally perform a new language task from only a few examples or from simple instructions - something which current NLP systems still largely struggle to do.

DeBERTa: Decoding-enhanced BERT with Disentangled Attention

microsoft/DeBERTa ICLR 2021

Recent progress in pre-trained neural language models has significantly improved the performance of many natural language processing (NLP) tasks.

Higher-order Coreference Resolution with Coarse-to-fine Inference

kentonl/e2e-coref NAACL 2018

We introduce a fully differentiable approximation to higher-order inference for coreference resolution.

SpanBERT: Improving Pre-training by Representing and Predicting Spans

facebookresearch/SpanBERT TACL 2020

We present SpanBERT, a pre-training method that is designed to better represent and predict spans of text.

Stanza: A Python Natural Language Processing Toolkit for Many Human Languages

stanfordnlp/stanza ACL 2020

We introduce Stanza, an open-source Python natural language processing toolkit supporting 66 human languages.

End-to-end Neural Coreference Resolution

kentonl/e2e-coref EMNLP 2017

We introduce the first end-to-end coreference resolution model and show that it significantly outperforms all previous work without using a syntactic parser or hand-engineered mention detector.

Mind the GAP: A Balanced Corpus of Gendered Ambiguous Pronouns

google-research-datasets/gap-coreference TACL 2018

Coreference resolution is an important task for natural language understanding, and the resolution of ambiguous pronouns a longstanding challenge.

PaLM: Scaling Language Modeling with Pathways

lucidrains/CoCa-pytorch Google Research 2022

To further our understanding of the impact of scale on few-shot learning, we trained a 540-billion parameter, densely activated, Transformer language model, which we call Pathways Language Model PaLM.

Multi-Task Identification of Entities, Relations, and Coreference for Scientific Knowledge Graph Construction

luanyi/DyGIE EMNLP 2018

We introduce a multi-task setup of identifying and classifying entities, relations, and coreference clusters in scientific articles.