no code implementations • BioNLP (ACL) 2022 • Liyan Tang, Shravan Kooragayalu, Yanshan Wang, Ying Ding, Greg Durrett, Justin F. Rousseau, Yifan Peng
Generating a summary from findings has been recently explored (Zhang et al., 2018, 2020) in note types such as radiology reports that typically have short length.
no code implementations • NAACL (TeachingNLP) 2021 • Greg Durrett, Jifan Chen, Shrey Desai, Tanya Goyal, Lucas Kabela, Yasumasa Onoe, Jiacheng Xu
We present a series of programming assignments, adaptable to a range of experience levels from advanced undergraduate to PhD, to teach students design and implementation of modern NLP systems.
1 code implementation • Findings (EMNLP) 2021 • Jifan Chen, Eunsol Choi, Greg Durrett
To build robust question answering systems, we need the ability to verify whether answers to questions are truly correct, not just “good enough” in the context of imperfect QA datasets.
no code implementations • 15 Apr 2025 • Juan Diego Rodriguez, Wenxuan Ding, Katrin Erk, Greg Durrett
Although large language models (LLMs) have become generally more capable and accurate across many tasks, some fundamental sources of unreliability remain in their behavior.
1 code implementation • 1 Apr 2025 • Melanie Subbiah, Akankshya Mishra, Grace Kim, Liyan Tang, Greg Durrett, Kathleen McKeown
This task is generally treated as a binary judgment of whether the claim is supported or unsupported in relation to the source.
2 code implementations • 7 Feb 2025 • Amitayush Thakur, George Tsoukalas, Greg Durrett, Swarat Chaudhuri
We address this weakness with a multilingual proof framework, ${\rm P{\small ROOF}W{\small ALA}}$, that allows a standardized form of interaction between neural theorem-provers and two established ITPs (Coq and Lean).
no code implementations • 9 Jan 2025 • Xi Ye, Fangcong Yin, Yinghui He, Joie Zhang, Howard Yen, Tianyu Gao, Greg Durrett, Danqi Chen
LongProc consists of six diverse procedural generation tasks, such as extracting structured information from HTML pages into a TSV format and executing complex search procedures to create travel plans.
no code implementations • 29 Oct 2024 • Xinyu Zhao, Fangcong Yin, Greg Durrett
To defray the cost of pretraining LLMs over long contexts, recent work takes an approach of synthetic context extension: fine-tuning LLMs with synthetically generated long-context data in a post-training stage.
no code implementations • 7 Oct 2024 • Aniruddh Sriram, Fangyuan Xu, Eunsol Choi, Greg Durrett
By leveraging the AVeriTeC dataset, which annotates subquestions for claims with human written answers from evidence documents, we fine-tune Contriever with a contrastive objective based on multiple training signals, including distillation from GPT-4, evaluating subquestion answers, and gold labels in the dataset.
1 code implementation • 18 Sep 2024 • Zayne Sprague, Fangcong Yin, Juan Diego Rodriguez, Dongwei Jiang, Manya Wadhwa, Prasann Singhal, Xinyu Zhao, Xi Ye, Kyle Mahowald, Greg Durrett
Chain-of-thought (CoT) via prompting is the de facto method for eliciting reasoning capabilities from large language models (LLMs).
no code implementations • 8 Jul 2024 • Zeyu Leo Liu, Shrey Pandit, Xi Ye, Eunsol Choi, Greg Durrett
An instance in our benchmark consists of a synthetic API function update paired with a program synthesis example that uses the updated functionality; our goal is to update an LLM to be able to solve this program synthesis example without providing documentation of the update at inference time.
1 code implementation • 2 Jul 2024 • Manya Wadhwa, Xinyu Zhao, Junyi Jessy Li, Greg Durrett
Recent work has explored the capability of large language models (LLMs) to identify and correct errors in LLM-generated responses.
1 code implementation • 28 Jun 2024 • Anisha Gunjal, Greg Durrett
Automatic factuality verification of large language model (LLM) generations is becoming more and more widely used to combat hallucinations.
1 code implementation • 25 Jun 2024 • Thom Lake, Eunsol Choi, Greg Durrett
Alignment suppresses irrelevant and unhelpful content while shifting the output distribution toward longer responses that cover information spanning several responses from the base LLM, essentially presenting diverse information in a single response.
1 code implementation • 3 Jun 2024 • Fangcong Yin, Xi Ye, Greg Durrett
For truthfulness and reasoning tasks, we find that LoFiT's intervention vectors are more effective for LLM adaptation than vectors from representation intervention methods such as Inference-time Intervention.
1 code implementation • 16 May 2024 • Abhishek Divekar, Greg Durrett
It is often desirable to distill the capabilities of large language models (LLMs) into smaller student models due to compute and memory constraints.
1 code implementation • 2 May 2024 • Prasann Singhal, Nathan Lambert, Scott Niekum, Tanya Goyal, Greg Durrett
Varied approaches for aligning language models have been proposed, including supervised fine-tuning, RLHF, and direct optimization methods such as DPO.
2 code implementations • 16 Apr 2024 • Liyan Tang, Philippe Laban, Greg Durrett
We release LLM-AggreFact, code for data synthesis, and models.
1 code implementation • 16 Apr 2024 • Yating Wu, Ritika Mangla, Alexandros G. Dimakis, Greg Durrett, Junyi Jessy Li
QSALIENCE is instruction-tuned over our dataset of linguist-annotated salience scores of 1, 766 (context, question) pairs.
no code implementations • 7 Dec 2023 • Jarad Forristal, Niloofar Mireshghallah, Greg Durrett, Taylor Berg-Kirkpatrick
Recent work has shown that energy-based language modeling is an effective framework for controllable text generation because it enables flexible integration of arbitrary discriminators.
3 code implementations • 24 Oct 2023 • Zayne Sprague, Xi Ye, Kaj Bostrom, Swarat Chaudhuri, Greg Durrett
We evaluate a range of LLMs and prompting techniques on this dataset and characterize the gaps that remain for techniques like chain-of-thought to perform robust reasoning.
1 code implementation • 23 Oct 2023 • Yating Wu, Ritika Mangla, Greg Durrett, Junyi Jessy Li
Questions Under Discussion (QUD) is a versatile linguistic framework in which discourse progresses as continuously asking questions and answering them.
1 code implementation • 5 Oct 2023 • Prasann Singhal, Tanya Goyal, Jiacheng Xu, Greg Durrett
Great success has been reported using Reinforcement Learning from Human Feedback (RLHF) to align large language models, with open preference datasets enabling wider experimentation, particularly for "helpfulness" in tasks like dialogue and web question answering.
1 code implementation • 16 Sep 2023 • Juan Diego Rodriguez, Katrin Erk, Greg Durrett
Aligned paragraphs are sourced from Wikipedia pages in different languages, reflecting real information divergences observed in the wild.
1 code implementation • 5 Jul 2023 • Zayne Sprague, Kaj Bostrom, Swarat Chaudhuri, Greg Durrett
Specifically, we evaluate whether embedding spaces exhibit a property we call deductive additivity: the sum of premise statement embeddings should be close to embeddings of conclusions based on those premises.
1 code implementation • NeurIPS 2023 • Shankar Padmanabhan, Yasumasa Onoe, Michael J. Q. Zhang, Greg Durrett, Eunsol Choi
Then, we update the model parameters so that the distribution of the LM (the student) matches the distribution of the LM conditioned on the definition (the teacher) on the transfer set.
1 code implementation • 1 Jun 2023 • Prasann Singhal, Jiacheng Xu, Xi Ye, Greg Durrett
Standard decoding approaches for conditional text generation tasks typically search for an output hypothesis with high model probability, but this may not yield the best hypothesis according to human judgments of quality.
no code implementations • 30 May 2023 • Liyan Tang, Yifan Peng, Yanshan Wang, Ying Ding, Greg Durrett, Justin F. Rousseau
To tackle this problem, we propose a controlled text generation method that uses a novel contrastive learning strategy to encourage models to differentiate between generating likely and less likely outputs according to humans.
no code implementations • 29 May 2023 • Jiayi Wei, Greg Durrett, Isil Dillig
Developers often dedicate significant time to maintaining and refactoring existing code.
no code implementations • 24 May 2023 • Anisha Gunjal, Greg Durrett
Past work has studied event prediction and event language modeling, sometimes mediated through structured representations of knowledge in the form of event schemas.
1 code implementation • 24 May 2023 • Manya Wadhwa, Jifan Chen, Junyi Jessy Li, Greg Durrett
These scores should reflect the annotators' underlying assessments of the example.
1 code implementation • 19 May 2023 • Jifan Chen, Grace Kim, Aniruddh Sriram, Greg Durrett, Eunsol Choi
Evidence retrieval is a core part of automatic fact-checking.
1 code implementation • NeurIPS 2023 • Xi Ye, Qiaochu Chen, Isil Dillig, Greg Durrett
In this paper, we propose a new satisfiability-aided language modeling (SatLM) approach for improving the reasoning capabilities of LLMs.
1 code implementation • 2 May 2023 • Yasumasa Onoe, Michael J. Q. Zhang, Shankar Padmanabhan, Greg Durrett, Eunsol Choi
Pre-trained language models (LMs) are used for knowledge intensive tasks like question answering, but their knowledge gets continuously outdated as the world changes.
1 code implementation • 16 Mar 2023 • Jiayi Wei, Greg Durrett, Isil Dillig
There has been growing interest in automatically predicting missing type annotations in programs written in Python and JavaScript.
2 code implementations • 2 Mar 2023 • Ryo Kamoi, Tanya Goyal, Juan Diego Rodriguez, Greg Durrett
Textual entailment models are increasingly applied in settings like fact-checking, presupposition verification in question answering, or summary evaluation.
1 code implementation • 14 Feb 2023 • Mahnaz Koupaee, Greg Durrett, Nathanael Chambers, Niranjan Balasubramanian
Event scenarios are often complex and involve multiple event sequences connected through different entity participants.
1 code implementation • 9 Feb 2023 • Xi Ye, Greg Durrett
We first generate sets of candidate explanations for each example in the prompt using a leave-one-out scheme, then find an effective combination of these explanations with a two-stage framework.
1 code implementation • 29 Nov 2022 • Adithya Bhaskar, Alexander R. Fabbri, Greg Durrett
Large language models have shown impressive performance across a wide variety of tasks, including text summarization.
1 code implementation • 25 Nov 2022 • Xi Ye, Srinivasan Iyer, Asli Celikyilmaz, Ves Stoyanov, Greg Durrett, Ramakanth Pasunuru
Large language models (LLMs) have exhibited remarkable capabilities in learning from explanations in prompts, but there has been limited understanding of exactly how these explanations function or why they are effective.
2 code implementations • 1 Nov 2022 • Zayne Sprague, Kaj Bostrom, Swarat Chaudhuri, Greg Durrett
A growing body of work studies how to answer a question or verify a claim by generating a natural language "proof": a chain of deductive inferences yielding the answer based on a set of premises.
no code implementations • 13 Oct 2022 • Prasann Singhal, Jarad Forristal, Xi Ye, Greg Durrett
We address the task of predicting out-of-domain (OOD) performance in a few-shot fashion: given a few target-domain examples and a set of models with similar training performance, can we understand how these models will perform on OOD test data?
1 code implementation • 13 Oct 2022 • Ryo Kamoi, Tanya Goyal, Greg Durrett
Despite recent progress in abstractive summarization, models often generate summaries with factual errors.
1 code implementation • 12 Oct 2022 • Wei-Jen Ko, Yating Wu, Cutter Dalton, Dananjay Srinivas, Greg Durrett, Junyi Jessy Li
Human evaluation results show that QUD dependency parsing is possible for language models trained with this crowdsourced, generalizable annotation scheme.
1 code implementation • 26 Sep 2022 • Tanya Goyal, Junyi Jessy Li, Greg Durrett
Finally, we evaluate models on a setting beyond generic summarization, specifically keyword-based summarization, and show how dominant fine-tuning approaches compare to prompting.
1 code implementation • 25 May 2022 • Liyan Tang, Tanya Goyal, Alexander R. Fabbri, Philippe Laban, Jiacheng Xu, Semih Yavuz, Wojciech Kryściński, Justin F. Rousseau, Greg Durrett
We compare performance of state-of-the-art factuality metrics, including recent ChatGPT-based metrics, on this stratified benchmark and show that their performance varies significantly across different types of summarization models.
1 code implementation • 19 May 2022 • Tanya Goyal, Junyi Jessy Li, Greg Durrett
In this work, we introduce SNaC, a narrative coherence evaluation framework rooted in fine-grained annotations for long summaries.
no code implementations • 14 May 2022 • Jifan Chen, Aniruddh Sriram, Eunsol Choi, Greg Durrett
Verifying complex political claims is a challenging task, especially when politicians use various tactics to subtly misrepresent the facts.
1 code implementation • 6 May 2022 • Xi Ye, Greg Durrett
Does prompting a large language model (LLM) like GPT-3 with explanations improve in-context learning?
no code implementations • Findings (NAACL) 2022 • Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett
Given its wide coverage on entity knowledge and temporal indexing, our dataset can be used to evaluate LMs and techniques designed to modify or extend their knowledge.
no code implementations • 16 Jan 2022 • Kaj Bostrom, Zayne Sprague, Swarat Chaudhuri, Greg Durrett
In settings from fact-checking to question answering, we frequently want to know whether a collection of evidence (premises) entails a hypothesis.
1 code implementation • NAACL 2022 • Jiacheng Xu, Siddhartha Reddy Jonnalagadda, Greg Durrett
Conditional neural text generation models generate high-quality outputs, but often concentrate around a mode when what we really want is a diverse set of options.
1 code implementation • 1 Nov 2021 • Wei-Jen Ko, Cutter Dalton, Mark Simmons, Eliza Fisher, Greg Durrett, Junyi Jessy Li
While there has been substantial progress in text comprehension through simple factoid question answering, more holistic comprehension of a discourse still presents a major challenge (Dunietz et al., 2020).
no code implementations • Findings (ACL) 2022 • Tanya Goyal, Jiacheng Xu, Junyi Jessy Li, Greg Durrett
Across different datasets (CNN/DM, XSum, MediaSum) and summary properties, such as abstractiveness and hallucination, we study what the model learns at different stages of its fine-tuning process.
1 code implementation • ACL 2022 • Ojas Ahuja, Jiacheng Xu, Akshay Gupta, Kevin Horecka, Greg Durrett
Generic summaries try to cover an entire document and query-based summaries try to answer document-specific questions.
no code implementations • 15 Oct 2021 • Nila Selvaraj, Yasumasa Onoe, Greg Durrett
In this paper, we present a unified cross-lingual fine-grained entity typing model capable of handling over 100 languages and analyze this model's ability to generalize to languages and entities unseen during training.
no code implementations • 14 Oct 2021 • Liyan Tang, Dhruv Rajan, Suyash Mohan, Abhijeet Pradhan, R. Nick Bryan, Greg Durrett
We show that regularization with small amounts of evidence supervision during training can substantially improve the quality of extracted evidence.
2 code implementations • ACL 2022 • Xi Ye, Greg Durrett
Our approach first extracts a set of features combining human intuition about the task with model attributions generated by black box interpretation techniques, then uses a simple calibrator, in the form of a classifier, to predict whether the base model was correct or not.
2 code implementations • 3 Sep 2021 • Yasumasa Onoe, Michael J. Q. Zhang, Eunsol Choi, Greg Durrett
We introduce CREAK, a testbed for commonsense reasoning about entity knowledge, bridging fact-checking about entities (Harry Potter is a wizard and is skilled at riding a broomstick) with commonsense inferences (if you're good at a skill you can teach others how to do it).
no code implementations • ACL 2021 • Mahnaz Koupaee, Greg Durrett, Nathanael Chambers, Niranjan Balasubramanian
Event language models represent plausible sequences of events.
1 code implementation • ACL 2021 • Jiacheng Xu, Greg Durrett
Despite the prominence of neural abstractive summarization models, we know little about how they actually form summaries and how to understand where their decisions come from.
1 code implementation • 18 Apr 2021 • Jifan Chen, Eunsol Choi, Greg Durrett
To build robust question answering systems, we need the ability to verify whether answers to questions are truly correct, not just "good enough" in the context of imperfect QA datasets.
1 code implementation • EMNLP 2021 • Kaj Bostrom, Xinyu Zhao, Swarat Chaudhuri, Greg Durrett
Natural language is an attractive representation for this purpose -- it is both highly expressive and easy for humans to understand.
3 code implementations • NAACL 2021 • Tanya Goyal, Greg Durrett
Recent pre-trained abstractive summarization systems have started to achieve credible performance, but a major barrier to their use in practice is their propensity to output summaries that are not faithful to the input and that contain factual errors.
1 code implementation • NAACL 2021 • Elisa Ferracane, Greg Durrett, Junyi Jessy Li, Katrin Erk
Discourse signals are often implicit, leaving it up to the interpreter to draw the required inferences.
1 code implementation • EMNLP 2021 • Xi Ye, Rohan Nair, Greg Durrett
When a model attribution technique highlights a particular part of the input, a user might understand this highlight as making a statement about counterfactuals (Miller, 2019): if that part of the input were to change, the model's prediction might change as well.
no code implementations • EACL 2021 • Sagnik Majumder, Chinmoy Samant, Greg Durrett
While numerous methods have been proposed as defenses against adversarial examples in question answering (QA), these techniques are often model specific, require retraining of the model, and give only marginal improvements in performance over vanilla models.
1 code implementation • ACL 2021 • Yasumasa Onoe, Michael Boratko, Andrew McCallum, Greg Durrett
Neural entity typing models typically represent fine-grained entity types as vectors in a high-dimensional space, but such spaces are not well-suited to modeling these types' complex interdependencies.
Ranked #9 on
Entity Typing
on Open Entity
no code implementations • ACL 2021 • Shih-ting Lin, Nathanael Chambers, Greg Durrett
We propose a single model that addresses both temporal ordering, sorting given events into the order they occurred, and event infilling, predicting new events which fit into an existing temporally-ordered sequence.
2 code implementations • EACL (AdaptNLP) 2021 • Xinyu Zhao, Shih-ting Lin, Greg Durrett
A principal barrier to training temporal relation extraction models in new domains is the lack of varied, high quality examples and the challenge of collecting more.
1 code implementation • EMNLP 2020 • Jiacheng Xu, Shrey Desai, Greg Durrett
An advantage of seq2seq abstractive summarization models is that they generate text in a free-form manner, but this flexibility makes it difficult to interpret model behavior.
1 code implementation • EMNLP 2020 • Shrey Desai, Jiacheng Xu, Greg Durrett
Compressive summarization systems typically rely on a crafted set of syntactic rules to determine what spans of possible summary sentences can be deleted, then learn a model of what to actually delete by optimizing for content selection (ROUGE).
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Tanya Goyal, Greg Durrett
Experiments show that our dependency arc entailment model trained on this data can identify factual inconsistencies in paraphrasing and summarization better than sentence-level methods or those based on question generation, while additionally localizing the erroneous parts of the generation.
no code implementations • Findings (EMNLP) 2021 • Xi Ye, Qiaochu Chen, Isil Dillig, Greg Durrett
Multimodal program synthesis, which leverages different types of user input to synthesize a desired program, is an attractive way to scale program synthesis to challenging settings; however, it requires integrating noisy signals from the user, like natural language, with hard constraints on the program's behavior.
1 code implementation • EMNLP 2020 • Wei-Jen Ko, Te-Yuan Chen, Yiyan Huang, Greg Durrett, Junyi Jessy Li
Inquisitive probing questions come naturally to humans in a variety of settings, but is a challenging task for automatic systems.
no code implementations • 18 Sep 2020 • Shih-ting Lin, Greg Durrett
Current methods in open-domain question answering (QA) usually employ a pipeline of first retrieving relevant documents, then applying strong reading comprehension (RC) models to that retrieved text.
no code implementations • 17 Aug 2020 • Su Wang, Greg Durrett, Katrin Erk
We propose a method for controlled narrative/story generation where we are able to guide the model to produce coherent narratives with user-specified target endings by interpolation: for example, we are told that Jim went hiking and at the end Jim needed to be rescued, and we want the model to incrementally generate steps along the way.
2 code implementations • ACL 2020 • Tanya Goyal, Greg Durrett
Paraphrasing natural language sentences is a multifaceted process: it might involve replacing individual words or short phrases, local rearrangement of content, or high-level restructuring like topicalization or passivization.
no code implementations • ACL 2020 • Xi Ye, Qiaochu Chen, Isil Dillig, Greg Durrett
Existing datasets for regular expression (regex) generation from natural language are limited in complexity; compared to regex tasks that users post on StackOverflow, the regexes in these datasets are simple, and the language used to describe them is not diverse.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Yasumasa Onoe, Greg Durrett
On entity probing tasks involving recognizing entity identity, our embeddings used in parameter-free downstream models achieve competitive performance with ELMo- and BERT-based embeddings in trained models.
no code implementations • NAACL 2021 • Jifan Chen, Greg Durrett
Current textual question answering models achieve strong performance on in-domain test sets, but often do so by fitting surface-level patterns in the data, so they fail to generalize to out-of-distribution settings.
1 code implementation • ICLR 2020 • Jiayi Wei, Maruth Goyal, Greg Durrett, Isil Dillig
Given this program abstraction, we then use a graph neural network to propagate information between related type variables and eventually make type predictions.
1 code implementation • Findings of the Association for Computational Linguistics 2020 • Kaj Bostrom, Greg Durrett
We analyze differences between BPE and unigram LM tokenization, finding that the latter method recovers subword units that align more closely with morphology and avoids problems stemming from BPE's greedy construction procedure.
1 code implementation • EMNLP 2020 • Shrey Desai, Greg Durrett
Pre-trained Transformers are now ubiquitous in natural language processing, but despite their high end-task performance, little is known empirically about whether they are calibrated.
3 code implementations • 7 Oct 2019 • Jifan Chen, Shih-ting Lin, Greg Durrett
Our analysis shows the properties of chains that are crucial for high performance: in particular, modeling extraction sequentially is important, as is dealing with each candidate sentence in a context-aware way.
Ranked #4 on
Question Answering
on WikiHop
no code implementations • IJCNLP 2019 • Su Wang, Greg Durrett, Katrin Erk
The news coverage of events often contains not one but multiple incompatible accounts of what happened.
1 code implementation • 12 Sep 2019 • Yasumasa Onoe, Greg Durrett
For this problem, a domain is characterized not just by genre of text but even by factors as specific as the particular distribution of entities, as neural models tend to overfit by memorizing properties of frequent entities in a dataset.
1 code implementation • IJCNLP 2019 • Aditya Gupta, Greg Durrett
Tracking entities in procedural language requires understanding the transformations arising from actions on entities as well as those entities' interactions.
1 code implementation • 16 Aug 2019 • Xi Ye, Qiaochu Chen, Xinyu Wang, Isil Dillig, Greg Durrett
Our system achieves state-of-the-art performance on the prior datasets and solves 57% of the real-world dataset, which existing neural systems completely fail on.
3 code implementations • ACL 2019 • Tanya Goyal, Greg Durrett
Data-driven models have demonstrated state-of-the-art performance in inferring the temporal ordering of events in text.
1 code implementation • ACL 2019 • Elisa Ferracane, Greg Durrett, Junyi Jessy Li, Katrin Erk
Discourse structure is integral to understanding a text and is helpful in many NLP tasks.
1 code implementation • NAACL 2019 • Wei-Jen Ko, Greg Durrett, Junyi Jessy Li
Sequence-to-sequence models for open-domain dialogue generation tend to favor generic, uninformative responses.
1 code implementation • NAACL 2019 • Yasumasa Onoe, Greg Durrett
In this work, we propose a two-stage procedure for handling this type of data: denoise it with a learned model, then train our final model on clean and denoised distant data with standard supervised training.
Ranked #2 on
Entity Typing
on Ontonotes v5 (English)
no code implementations • NAACL 2019 • Jifan Chen, Greg Durrett
First, we explore sentence-factored models for these tasks; by design, these models cannot do multi-hop reasoning, but they are still able to solve a large number of examples in both datasets.
no code implementations • WS 2019 • Aditya Gupta, Greg Durrett
The global discrete state structure is explicitly modeled with a neural CRF over the changing hidden representation of the entity.
Ranked #2 on
Procedural Text Understanding
on ProPara
1 code implementation • IJCNLP 2019 • Jiacheng Xu, Greg Durrett
In this work, we present a neural model for single-document summarization based on joint extraction and syntactic compression.
2 code implementations • 13 Nov 2018 • Wei-Jen Ko, Greg Durrett, Junyi Jessy Li
Sentence specificity quantifies the level of detail in a sentence, characterizing the organization of information in discourse.
no code implementations • EMNLP 2018 • Su Wang, Eric Holgate, Greg Durrett, Katrin Erk
During natural disasters and conflicts, information about what happened is often confusing, messy, and distributed across many sources.
1 code implementation • EMNLP 2018 • David Mueller, Greg Durrett
To disambiguate between closely related concepts, entity linking systems need to effectively distill cues from their context, which may be quite noisy.
1 code implementation • EMNLP 2018 • Jiacheng Xu, Greg Durrett
A hallmark of variational autoencoders (VAEs) for text processing is their combination of powerful encoder-decoder models, such as LSTMs, with simple latent distributions, typically multivariate Gaussians.
Ranked #3 on
Topic Models
on AG News
1 code implementation • NAACL 2018 • Su Wang, Greg Durrett, Katrin Erk
Distributional data tells us that a man can swallow candy, but not that a man can swallow a paintball, since this is never attested.
1 code implementation • EMNLP 2017 • Greg Durrett, Jonathan K. Kummerfeld, Taylor Berg-Kirkpatrick, Rebecca S. Portnoff, Sadia Afroz, Damon McCoy, Kirill Levchenko, Vern Paxson
One weakness of machine-learned NLP models is that they typically perform poorly on out-of-domain data.
1 code implementation • NAACL 2016 • Matthew Francis-Landau, Greg Durrett, Dan Klein
A key challenge in entity linking is making effective use of contextual information to disambiguate mentions that might refer to different entities in different contexts.
no code implementations • ACL 2016 • Greg Durrett, Taylor Berg-Kirkpatrick, Dan Klein
We present a discriminative model for single-document summarization that integrally combines compression and anaphoricity constraints.
no code implementations • IJCNLP 2015 • Greg Durrett, Dan Klein
This paper describes a parsing model that combines the exact dynamic programming of CRF parsing with the rich nonlinear featurization of neural net approaches.
no code implementations • TACL 2014 • Greg Durrett, Dan Klein
We present a joint model of three core tasks in the entity analysis stack: coreference resolution (within-document clustering), named entity recognition (coarse semantic typing), and entity linking (matching to Wikipedia entities).
Ranked #28 on
Named Entity Recognition (NER)
on Ontonotes v5 (English)