no code implementations • EMNLP 2020 • Valentina Pyatkin, Ayal Klein, Reut Tsarfaty, Ido Dagan
Discourse relations describe how two propositions relate to one another, and identifying them automatically is an integral part of natural language understanding.
no code implementations • LREC 2022 • Merel Scholman, Valentina Pyatkin, Frances Yung, Ido Dagan, Reut Tsarfaty, Vera Demberg
The current contribution studies the effect of worker selection and training on the agreement on implicit relation labels between workers and gold labels, for both the DC and the QA method.
1 code implementation • NAACL 2022 • Ori Shapira, Ramakanth Pasunuru, Mohit Bansal, Ido Dagan, Yael Amsterdamer
Interactive summarization is a task that facilitates user-guided exploration of information within a document set.
no code implementations • EMNLP 2020 • Oren Barkan, Avi Caciularu, Ido Dagan
We propose the novel \textit{Within-Between} Relation model for recognizing lexical-semantic relations between words.
no code implementations • WASSA (ACL) 2022 • Ayal Klein, Oren Pereg, Daniel Korat, Vasudev Lal, Moshe Wasserblat, Ido Dagan
In this paper, we investigate and establish empirically a prior conjecture, which suggests that the linguistic relations connecting opinion terms to their aspects transfer well across domains and therefore can be leveraged for cross-domain aspect term extraction.
no code implementations • 8 Aug 2024 • Paul Roit, Aviv Slobodkin, Eran Hirsch, Arie Cattan, Ayal Klein, Valentina Pyatkin, Ido Dagan
Detecting semantic arguments of a predicate word has been conventionally modeled as a sentence-level task.
no code implementations • 29 Jun 2024 • Omer Goldman, Alon Jacovi, Aviv Slobodkin, Aviya Maimon, Ido Dagan, Reut Tsarfaty
By using a descriptive vocabulary and discussing the relevant properties of difficulty in long-context, we can implement more informed research in this area.
no code implementations • 20 Jun 2024 • Omri Berkovitch, Sapir Caduri, Noam Kahlon, Anatoly Efros, Avi Caciularu, Ido Dagan
By effectively comprehending user intentions through their actions and interactions with GUIs, agents will be better positioned to achieve these goals.
1 code implementation • 2 Jun 2024 • Ori Ernst, Ori Shapira, Aviv Slobodkin, Sharon Adar, Mohit Bansal, Jacob Goldberger, Ran Levy, Ido Dagan
Multi-document summarization (MDS) is a challenging task, often decomposed to subtasks of salience and redundancy detection, followed by text generation.
no code implementations • 31 May 2024 • Valentina Pyatkin, Bonnie Webber, Ido Dagan, Reut Tsarfaty
Semantically, superlatives perform a set comparison: something (or some things) has the min/max property out of a set.
no code implementations • 20 May 2024 • Chen Huang, Yang Deng, Wenqiang Lei, Jiancheng Lv, Ido Dagan
As such, informative or hard data is assigned to the expert for annotation, while easy data is handled by the model.
1 code implementation • 2 May 2024 • Lotem Golany, Filippo Galgani, Maya Mamo, Nimrod Parasol, Omer Vandsburger, Nadav Bar, Ido Dagan
Automating data generation with Large Language Models (LLMs) has become increasingly popular.
1 code implementation • 25 Mar 2024 • Aviv Slobodkin, Eran Hirsch, Arie Cattan, Tal Schuster, Ido Dagan
Recent efforts to address hallucinations in Large Language Models (LLMs) have focused on attributed text generation, which supplements generated texts with citations of supporting sources for post-generation fact-checking and corrections.
no code implementations • 22 Mar 2024 • Aviv Slobodkin, Ori Shapira, Ran Levy, Ido Dagan
This study lays the groundwork for further exploration of modular text generation in the multi-document setting, offering potential improvements in the quality and reliability of generated content.
1 code implementation • 7 Dec 2023 • Shmuel Amar, Liat Schiff, Ori Ernst, Asi Shefer, Ori Shapira, Ido Dagan
To advance research on more realistic scenarios, we introduce OpenAsp, a benchmark for multi-document \textit{open} aspect-based summarization.
1 code implementation • 19 Nov 2023 • Arie Cattan, Tom Hope, Doug Downey, Roy Bar-Haim, Lilach Eden, Yoav Kantor, Ido Dagan
Various NLP tasks require a complex hierarchical structure over nodes, where each node is a cluster of items.
coreference-resolution Cross Document Coreference Resolution
1 code implementation • 20 Oct 2023 • Moshe Berchansky, Peter Izsak, Avi Caciularu, Ido Dagan, Moshe Wasserblat
Fusion-in-Decoder (FiD) is an effective retrieval-augmented language model applied across a variety of open-domain tasks, such as question answering, fact checking, etc.
1 code implementation • 18 Oct 2023 • Aviv Slobodkin, Omer Goldman, Avi Caciularu, Ido Dagan, Shauli Ravfogel
In this paper, we explore the behavior of LLMs when presented with (un)answerable queries.
1 code implementation • 13 Oct 2023 • Aviv Slobodkin, Avi Caciularu, Eran Hirsch, Ido Dagan
Further, we substantially improve the silver training data quality via GPT-4 distillation.
no code implementations • 16 Aug 2023 • Aviv Slobodkin, Niv Nachum, Shmuel Amar, Ori Shapira, Ido Dagan
Current approaches for text summarization are predominantly automatic, with rather limited space for human intervention and control over the process.
1 code implementation • 24 May 2023 • Eran Hirsch, Valentina Pyatkin, Ruben Wolhandler, Avi Caciularu, Asi Shefer, Ido Dagan
In this paper, we suggest revisiting the sentence union generation task as an effective well-defined testbed for assessing text consolidation capabilities, decoupling the consolidation challenge from subjective content selection.
1 code implementation • 24 May 2023 • Avi Caciularu, Matthew E. Peters, Jacob Goldberger, Ido Dagan, Arman Cohan
The integration of multi-document pre-training objectives into language models has resulted in remarkable improvements in multi-document downstream tasks.
1 code implementation • 3 Apr 2023 • Valentina Pyatkin, Frances Yung, Merel C. J. Scholman, Reut Tsarfaty, Ido Dagan, Vera Demberg
Disagreement in natural language annotation has mostly been studied from a perspective of biases introduced by the annotators and the annotation frameworks.
2 code implementations • 24 Oct 2022 • Aviv Slobodkin, Paul Roit, Eran Hirsch, Ori Ernst, Ido Dagan
Producing a reduced version of a source text, as in generic or focused summarization, inherently involves two distinct subtasks: deciding on targeted content and generating a coherent text conveying it.
1 code implementation • 23 Oct 2022 • Alon Eirew, Avi Caciularu, Ido Dagan
The task of Cross-document Coreference Resolution has been traditionally formulated as requiring to identify all coreference links across a given set of documents.
Cross Document Coreference Resolution Open-Domain Question Answering +2
1 code implementation • 23 Oct 2022 • Ruben Wolhandler, Arie Cattan, Ori Ernst, Ido Dagan
To that end, we propose an automated measure for evaluating the degree to which a summary is ``disperse'', in the sense of the number of source documents needed to cover its content.
1 code implementation • 23 May 2022 • Ayal Klein, Eran Hirsch, Ron Eliav, Valentina Pyatkin, Avi Caciularu, Ido Dagan
Several recent works have suggested to represent semantic relations with questions and answers, decomposing textual information into separate interrogative natural language statements.
2 code implementations • NAACL 2022 • Ori Ernst, Avi Caciularu, Ori Shapira, Ramakanth Pasunuru, Mohit Bansal, Jacob Goldberger, Ido Dagan
Text clustering methods were traditionally incorporated into multi-document summarization (MDS) as a means for coping with considerable information repetition.
2 code implementations • NAACL 2022 • Avi Caciularu, Ido Dagan, Jacob Goldberger, Arman Cohan
Long-context question answering (QA) tasks require reasoning over a long document or multiple documents.
1 code implementation • NAACL 2022 • Daniela Brook Weiss, Paul Roit, Ori Ernst, Ido Dagan
NLP models that compare or consolidate information across multiple documents often struggle when challenged with recognizing substantial information redundancies across the texts.
1 code implementation • 3 Oct 2021 • Ori Shapira, Ramakanth Pasunuru, Ido Dagan, Yael Amsterdamer
Keyphrase extraction has been extensively researched within the single-document setting, with an abundance of methods, datasets and applications.
1 code implementation • EMNLP 2021 • Daniela Brook Weiss, Paul Roit, Ayal Klein, Ori Ernst, Ido Dagan
Multi-text applications, such as multi-document summarization, are typically required to model redundancies across related texts.
1 code implementation • EMNLP (ACL) 2021 • Eran Hirsch, Alon Eirew, Ori Shapira, Avi Caciularu, Arie Cattan, Ori Ernst, Ramakanth Pasunuru, Hadar Ronen, Mohit Bansal, Ido Dagan
We introduce iFacetSum, a web application for exploring topical document sets.
1 code implementation • EMNLP 2021 • Valentina Pyatkin, Paul Roit, Julian Michael, Reut Tsarfaty, Yoav Goldberg, Ido Dagan
We develop a two-stage model for this task, which first produces a context-independent question prototype for each role and then revises it to be contextually appropriate for the passage.
1 code implementation • Joint Conference on Lexical and Computational Semantics 2021 • Ohad Rozen, Shmuel Amar, Vered Shwartz, Ido Dagan
Our approach facilitates learning generic inference patterns requiring relational knowledge (e. g. inferences related to hypernymy) during training, while injecting on-demand the relevant relational facts (e. g. pangolin is an animal) at test time.
1 code implementation • Joint Conference on Lexical and Computational Semantics 2021 • Arie Cattan, Alon Eirew, Gabriel Stanovsky, Mandar Joshi, Ido Dagan
We point out that common evaluation practices for cross-document coreference resolution have been unrealistically permissive in their assumed settings, yielding inflated results.
coreference-resolution Cross Document Coreference Resolution
no code implementations • Joint Conference on Lexical and Computational Semantics 2021 • Avi Caciularu, Ido Dagan, Jacob Goldberger
We introduce a new approach for smoothing and improving the quality of word embeddings.
1 code implementation • Findings (ACL) 2021 • Arie Cattan, Alon Eirew, Gabriel Stanovsky, Mandar Joshi, Ido Dagan
Here, we introduce the first end-to-end model for CD coreference resolution from raw text, which extends the prominent model for within-document coreference to the CD setting.
coreference-resolution Cross Document Coreference Resolution
1 code implementation • NAACL 2021 • Ori Shapira, Ramakanth Pasunuru, Hadar Ronen, Mohit Bansal, Yael Amsterdamer, Ido Dagan
In this paper, we develop an end-to-end evaluation framework for interactive summarization, focusing on expansion-based interaction, which considers the accumulating information along a user session.
2 code implementations • AKBC 2021 • Arie Cattan, Sophie Johnson, Daniel Weld, Ido Dagan, Iz Beltagy, Doug Downey, Tom Hope
Determining coreference of concept mentions across multiple documents is a fundamental task in natural language understanding.
coreference-resolution Cross Document Coreference Resolution +1
no code implementations • 17 Apr 2021 • Ofer Sabo, Yanai Elazar, Yoav Goldberg, Ido Dagan
We explore Few-Shot Learning (FSL) for Relation Classification (RC).
2 code implementations • NAACL 2021 • Alon Eirew, Arie Cattan, Ido Dagan
To complement these resources and enhance future research, we present Wikipedia Event Coreference (WEC), an efficient methodology for gathering a large-scale dataset for cross-document event coreference from Wikipedia, where coreference links are not restricted within predefined topics.
no code implementations • EACL 2021 • James Ravenscroft, Amanda Clare, Arie Cattan, Ido Dagan, Maria Liakata
Cross-document co-reference resolution (CDCR) is the task of identifying and linking mentions to entities and concepts across many text documents.
1 code implementation • 29 Jan 2021 • James Ravenscroft, Arie Cattan, Amanda Clare, Ido Dagan, Maria Liakata
Cross-document co-reference resolution (CDCR) is the task of identifying and linking mentions to entities and concepts across many text documents.
2 code implementations • Findings (EMNLP) 2021 • Avi Caciularu, Arman Cohan, Iz Beltagy, Matthew E. Peters, Arie Cattan, Ido Dagan
We introduce a new pretraining approach geared for multi-document language modeling, incorporating two key ideas into the masked language modeling self-supervised objective.
Ranked #1 on Citation Recommendation on AAN test
1 code implementation • COLING 2020 • Ayal Klein, Jonathan Mamou, Valentina Pyatkin, Daniela Stepanov, Hangfeng He, Dan Roth, Luke Zettlemoyer, Ido Dagan
We propose a new semantic scheme for capturing predicate-argument relations for nominalizations, termed QANom.
1 code implementation • 6 Oct 2020 • Valentina Pyatkin, Ayal Klein, Reut Tsarfaty, Ido Dagan
Discourse relations describe how two propositions relate to one another, and identifying them automatically is an integral part of natural language understanding.
2 code implementations • EMNLP 2020 • Aaron Bornstein, Arie Cattan, Ido Dagan
Coreference annotation is an important, yet expensive and time consuming, task, which often involved expert annotators trained on complex decision guidelines.
2 code implementations • 23 Sep 2020 • Arie Cattan, Alon Eirew, Gabriel Stanovsky, Mandar Joshi, Ido Dagan
Recent evaluation protocols for Cross-document (CD) coreference resolution have often been inconsistent or lenient, leading to incomparable results across works and overestimation of performance.
coreference-resolution Cross Document Coreference Resolution +2
no code implementations • 17 Sep 2020 • Ori Shapira, Ramakanth Pasunuru, Hadar Ronen, Mohit Bansal, Yael Amsterdamer, Ido Dagan
Allowing users to interact with multi-document summarizers is a promising direction towards improving and customizing summary results.
1 code implementation • CoNLL (EMNLP) 2021 • Ori Ernst, Ori Shapira, Ramakanth Pasunuru, Michael Lepioshkin, Jacob Goldberger, Mohit Bansal, Ido Dagan
Aligning sentences in a reference summary with their counterparts in source documents was shown as a useful auxiliary summarization task, notably for generating training data for salience detection.
2 code implementations • Findings of the Association for Computational Linguistics 2020 • Yehudit Meged, Avi Caciularu, Vered Shwartz, Ido Dagan
We study the potential synergy between two different NLP tasks, both confronting predicate lexical variability: identifying predicate paraphrases, and event coreference resolution.
1 code implementation • ACL 2020 • Paul Roit, Ayal Klein, Daniela Stepanov, Jonathan Mamou, Julian Michael, Gabriel Stanovsky, Luke Zettlemoyer, Ido Dagan
Question-answer driven Semantic Role Labeling (QA-SRL) was proposed as an attractive open and natural flavour of SRL, potentially attainable from laymen.
1 code implementation • CONLL 2019 • Ohad Rozen, Vered Shwartz, Roee Aharoni, Ido Dagan
Phenomenon-specific "adversarial" datasets have been recently designed to perform targeted stress-tests for particular inference types.
no code implementations • WS 2019 • Yevgeniy Puzikov, Claire Gardent, Ido Dagan, Iryna Gurevych
End-to-end neural approaches have achieved state-of-the-art performance in many natural language processing (NLP) tasks.
1 code implementation • WS 2019 • Amit Moryossef, Ido Dagan, Yoav Goldberg
We follow the step-by-step approach to neural data-to-text generation we proposed in Moryossef et al (2019), in which the generation process is divided into a text-planning stage followed by a plan-realization stage.
1 code implementation • IJCNLP 2019 • Oren Pereg, Daniel Korat, Moshe Wasserblat, Jonathan Mamou, Ido Dagan
We present ABSApp, a portable system for weakly-supervised aspect-based sentiment extraction.
2 code implementations • IJCNLP 2019 • Florian Böhm, Yang Gao, Christian M. Meyer, Ori Shapira, Ido Dagan, Iryna Gurevych
Human evaluation experiments show that, compared to the state-of-the-art supervised-learning systems and ROUGE-as-rewards RL summarisation systems, the RL systems using our learned rewards during training generate summarieswith higher human ratings.
no code implementations • ACL 2019 • Tobias Falke, Leonardo F. R. Ribeiro, Prasetya Ajie Utama, Ido Dagan, Iryna Gurevych
While recent progress on abstractive summarization has led to remarkably fluent summaries, factual errors in generated summaries still severely limit their use in practice.
1 code implementation • ACL 2019 • Shany Barhom, Vered Shwartz, Alon Eirew, Michael Bugert, Nils Reimers, Ido Dagan
Our analysis confirms that all our representation elements, including the mention span itself, its context, and the relation to other mentions contribute to the model's success.
coreference-resolution Cross Document Coreference Resolution +4
no code implementations • WS 2019 • Simeng Sun, Ori Shapira, Ido Dagan, Ani Nenkova
We show that plain ROUGE F1 scores are not ideal for comparing current neural systems which on average produce different lengths.
1 code implementation • NAACL 2019 • Ori Shapira, David Gabay, Yang Gao, Hadar Ronen, Ramakanth Pasunuru, Mohit Bansal, Yael Amsterdamer, Ido Dagan
Conducting a manual evaluation is considered an essential part of summary evaluation methodology.
1 code implementation • NAACL 2019 • Amit Moryossef, Yoav Goldberg, Ido Dagan
We propose to split the generation process into a symbolic text-planning stage that is faithful to the input, followed by a neural generation stage that focuses only on realization.
Ranked #15 on Data-to-Text Generation on WebNLG
no code implementations • WS 2019 • Jonathan Mamou, Oren Pereg, Moshe Wasserblat, Ido Dagan
In this paper, we present a novel algorithm that combines multi-context term embeddings using a neural classifier and we test this approach on the use case of corpus-based term set expansion.
1 code implementation • TACL 2019 • Vered Shwartz, Ido Dagan
Building meaningful phrase representations is challenging because phrase meanings are not simply the sum of their constituent meanings.
no code implementations • EMNLP 2018 • Ori Shapira, David Gabay, Hadar Ronen, Judit Bar-Ilan, Yael Amsterdamer, Ani Nenkova, Ido Dagan
Practical summarization systems are expected to produce summaries of varying lengths, per user needs.
no code implementations • EMNLP 2018 • Gabriel Stanovsky, Ido Dagan
We propose a novel approach to semantic dependency parsing (SDP) by casting the task as an instance of multi-lingual machine translation, where each semantic representation is a different foreign dialect.
no code implementations • COLING 2018 • Jonathan Mamou, Oren Pereg, Moshe Wasserblat, Ido Dagan, Yoav Goldberg, Alon Eirew, Yael Green, Shira Guskin, Peter Izsak, Daniel Korat
We present SetExpander, a corpus-based system for expanding a seed set of terms into a more complete set of terms that belong to the same semantic class.
no code implementations • 26 Jul 2018 • Jonathan Mamou, Oren Pereg, Moshe Wasserblat, Ido Dagan, Yoav Goldberg, Alon Eirew, Yael Green, Shira Guskin, Peter Izsak, Daniel Korat
We present SetExpander, a corpus-based system for expanding a seed set of terms into a more complete set of terms that belong to the same semantic class.
no code implementations • NAACL 2018 • Gabriel Stanovsky, Julian Michael, Luke Zettlemoyer, Ido Dagan
We present data and methods that enable a supervised learning approach to Open Information Extraction (Open IE).
1 code implementation • ACL 2018 • Vered Shwartz, Ido Dagan
Revealing the implicit semantic relation between the constituents of a noun-compound is important for many NLP applications.
1 code implementation • NAACL 2018 • Julian Michael, Gabriel Stanovsky, Luheng He, Ido Dagan, Luke Zettlemoyer
We introduce Question-Answer Meaning Representations (QAMRs), which represent the predicate-argument structure of a sentence as a set of question-answer pairs.
no code implementations • EMNLP 2017 • Ori Shapira, Hadar Ronen, Meni Adler, Yael Amsterdamer, Judit Bar-Ilan, Ido Dagan
We present a novel interactive summarization system that is based on abstractive summarization, derived from a recent consolidated knowledge representation for multiple texts.
no code implementations • SEMEVAL 2017 • Vered Shwartz, Gabriel Stanovsky, Ido Dagan
We present a simple method for ever-growing extraction of predicate paraphrases from news headlines in Twitter.
no code implementations • EMNLP 2017 • Oren Melamud, Ido Dagan, Jacob Goldberger
Specifically, we show that with minor modifications to word2vec's algorithm, we get principled language models that are closely related to the well-established Noise Contrastive Estimation (NCE) based language models.
no code implementations • ACL 2017 • Gabriel Stanovsky, Judith Eckle-Kohler, Yevgeniy Puzikov, Ido Dagan, Iryna Gurevych
Previous models for the assessment of commitment towards a predicate in a sentence (also known as factuality prediction) were trained and tested against a specific annotated dataset, subsequently limiting the generality of their results.
1 code implementation • WS 2017 • Rachel Wities, Vered Shwartz, Gabriel Stanovsky, Meni Adler, Ori Shapira, Shyam Upadhyay, Dan Roth, Eugenio Martinez Camara, Iryna Gurevych, Ido Dagan
We propose to move from Open Information Extraction (OIE) ahead to Open Knowledge Representation (OKR), aiming to represent information conveyed jointly in a set of texts in an open text-based manner.
no code implementations • COLING 2016 • Omer Levy, Ido Dagan, Gabriel Stanovsky, Judith Eckle-Kohler, Iryna Gurevych
Sentence intersection captures the semantic overlap of two texts, generalizing over paradigms such as textual entailment and semantic text similarity.
Abstractive Text Summarization Natural Language Inference +2
1 code implementation • WS 2016 • Vered Shwartz, Ido Dagan
The reported results in the shared task bring this submission to the third place on subtask 1 (word relatedness), and the first place on subtask 2 (semantic relation classification), demonstrating the utility of integrating the complementary path-based and distributional information sources in recognizing concrete semantic relations.
no code implementations • 5 Sep 2016 • Oren Melamud, Ido Dagan, Jacob Goldberger
The obtained language modeling is closely related to NCE language models but is based on a simplified objective function.
1 code implementation • WS 2016 • Vered Shwartz, Ido Dagan
Recognizing various semantic relations between terms is beneficial for many NLP tasks.
no code implementations • LREC 2016 • Vasily Konovalov, Ron artstein, Oren Melamud, Ido Dagan
In this work, we introduce an annotated natural language human-agent dialogue corpus in the negotiation domain.
1 code implementation • ACL 2016 • Vered Shwartz, Yoav Goldberg, Ido Dagan
Detecting hypernymy relations is a key task in NLP, which is addressed in the literature using two complementary approaches.
no code implementations • 4 Mar 2016 • Gabriel Stanovsky, Jessica Ficler, Ido Dagan, Yoav Goldberg
Semantic NLP applications often rely on dependency trees to recognize major elements of the proposition structure of sentences.
Ranked #27 on Open Information Extraction on CaRB
no code implementations • TACL 2015 • Omer Levy, Yoav Goldberg, Ido Dagan
Recent trends suggest that neural-network-inspired word embedding models outperform traditional count-based distributional models on word similarity and analogy detection tasks.