no code implementations • EMNLP 2021 • Or Honovich, Leshem Choshen, Roee Aharoni, Ella Neeman, Idan Szpektor, Omri Abend
Neural knowledge-grounded generative models for dialogue often produce content that is factually inconsistent with the knowledge they rely on, making them unreliable and limiting their applicability.
Abstractive Text Summarization
Natural Language Inference
+3
1 code implementation • EMNLP 2020 • Liat Ein-Dor, Alon Halfon, Ariel Gera, Eyal Shnarch, Lena Dankin, Leshem Choshen, Marina Danilevsky, Ranit Aharonov, Yoav Katz, Noam Slonim
Here, we present a large-scale empirical study on active learning techniques for BERT-based classification, addressing a diverse set of AL strategies and datasets.
no code implementations • 24 May 2023 • Taelin Karidi, Leshem Choshen, Gal Patel, Omri Abend
For example, nouns and verbs are among the most frequent POS tags.
1 code implementation • 16 Mar 2023 • Alexander Yom Din, Taelin Karidi, Leshem Choshen, Mor Geva
Moreover, in the context of language modeling, our method allows "peeking" into early layer representations of GPT-2 and BERT, showing that often LMs already predict the final output in early layers.
no code implementations • 9 Feb 2023 • Almog Gueta, Elad Venezian, Colin Raffel, Noam Slonim, Yoav Katz, Leshem Choshen
Specifically, we show that starting from the center of the region is as good or better than the pre-trained model in 11 of 12 datasets and improves accuracy by 3. 06 on average.
no code implementations • 27 Jan 2023 • Alex Warstadt, Leshem Choshen, Aaron Mueller, Adina Williams, Ethan Wilcox, Chengxu Zhuang
In partnership with CoNLL and CMCL, we provide a platform for approaches to pretraining with a limited-size corpus sourced from data inspired by the input to children.
no code implementations • 2 Dec 2022 • Shachar Don-Yehiya, Elad Venezian, Colin Raffel, Noam Slonim, Yoav Katz, Leshem Choshen
We show that ColD Fusion yields comparable benefits to multitask pretraining by producing a model that (a) attains strong performance on all of the datasets it was multitask trained on and (b) is a better starting point for finetuning on unseen datasets.
1 code implementation • 10 Nov 2022 • Ella Neeman, Roee Aharoni, Or Honovich, Leshem Choshen, Idan Szpektor, Omri Abend
Question answering models commonly have access to two sources of "knowledge" during inference time: (1) parametric knowledge - the factual knowledge encoded in the model weights, and (2) contextual knowledge - external knowledge (e. g., a Wikipedia passage) given to the model to generate a grounded answer.
no code implementations • 31 Oct 2022 • Leshem Choshen, Elad Venezian, Shachar Don-Yehia, Noam Slonim, Yoav Katz
Such a model, finetuned on some source dataset, may provide a better starting point for a new finetuning process on a desired target dataset.
no code implementations • COLING 2022 • Asaf Yehudai, Leshem Choshen, Lior Fox, Omri Abend
Applying Reinforcement learning (RL) following maximum likelihood estimation (MLE) pre-training is a versatile method for enhancing neural machine translation (NMT) performance.
1 code implementation • 2 Aug 2022 • Eyal Shnarch, Alon Halfon, Ariel Gera, Marina Danilevsky, Yannis Katsis, Leshem Choshen, Martin Santillan Cooper, Dina Epelboim, Zheng Zhang, Dakuo Wang, Lucy Yip, Liat Ein-Dor, Lena Dankin, Ilya Shnayderman, Ranit Aharonov, Yunyao Li, Naftali Liberman, Philip Levin Slesarev, Gwilym Newton, Shila Ofek-Koifman, Noam Slonim, Yoav Katz
Text classification can be useful in many real-world scenarios, saving a lot of time for end users.
1 code implementation • 18 May 2022 • Shachar Don-Yehiya, Leshem Choshen, Omri Abend
We show that this augmentation method can improve the performance of the Quality-Estimation task as well.
1 code implementation • 11 May 2022 • Leshem Choshen, Ofir Shifman, Omri Abend
In Grammatical Error Correction, systems are evaluated by the number of errors they correct.
1 code implementation • 6 Apr 2022 • Leshem Choshen, Elad Venezian, Noam Slonim, Yoav Katz
We also show that fusing is often better than intertraining.
1 code implementation • ACL 2022 • Eyal Shnarch, Ariel Gera, Alon Halfon, Lena Dankin, Leshem Choshen, Ranit Aharonov, Noam Slonim
In real-world scenarios, a text classification task often begins with a cold start, when labeled data is scarce.
no code implementations • *SEM (NAACL) 2022 • Aviv Slobodkin, Leshem Choshen, Omri Abend
We further show an additional gain when using both semantic and syntactic structures in some language pairs.
1 code implementation • 6 Oct 2021 • Gal Patel, Leshem Choshen, Omri Abend
We present a methodology that explores how sentence structure is reflected in neural representations of machine translation systems.
1 code implementation • ACL 2022 • Leshem Choshen, Guy Hacohen, Daphna Weinshall, Omri Abend
These findings suggest that there is some mutual inductive bias that underlies these models' learning of linguistic phenomena.
1 code implementation • 23 Aug 2021 • Leshem Choshen, Idan Amit
We present ComSum, a data set of 7 million commit messages for text summarization.
no code implementations • 1 Jun 2021 • Ofek Rafaeli, Omri Abend, Leshem Choshen, Dmitry Nikolaev
In this research paper, I will elaborate on a method to evaluate machine translation models based on their performance on underlying syntactical phenomena between English and Arabic languages.
1 code implementation • 16 Apr 2021 • Or Honovich, Leshem Choshen, Roee Aharoni, Ella Neeman, Idan Szpektor, Omri Abend
Neural knowledge-grounded generative models for dialogue often produce content that is factually inconsistent with the knowledge they rely on, making them unreliable and limiting their applicability.
1 code implementation • NAACL 2021 • Aviv Slobodkin, Leshem Choshen, Omri Abend
Probing neural models for the ability to perform downstream tasks using their activation patterns is often used to localize what parts of the network specialize in performing what tasks.
1 code implementation • LREC 2022 • Piyawat Lertvittayakumjorn, Leshem Choshen, Eyal Shnarch, Francesca Toni
Data exploration is an important step of every data science and machine learning project, including those involving textual data.
1 code implementation • 6 Apr 2021 • Leshem Choshen, Matanel Oren, Dmitry Nikolaev, Omri Abend
SERRANT is a system and code for automatic classification of English grammatical errors that combines SErCl and ERRANT.
1 code implementation • 29 Jan 2021 • Leshem Choshen, Omri Abend
Notwithstanding recent advances, syntactic generalization remains a challenge for text decoders.
no code implementations • 1 Jan 2021 • Eyal Shnarch, Ariel Gera, Alon Halfon, Lena Dankin, Leshem Choshen, Ranit Aharonov, Noam Slonim
In such low resources scenarios, we suggest performing an unsupervised classification task prior to fine-tuning on the target task.
1 code implementation • CONLL 2020 • Leshem Choshen, Dmitry Nikolaev, Yevgeni Berzak, Omri Abend
We present a method for classifying syntactic errors in learner language, namely errors whose correction alters the morphosyntactic structure of a sentence.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Eyal Shnarch, Leshem Choshen, Guy Moshkowich, Noam Slonim, Ranit Aharonov
Approaching new data can be quite deterrent; you do not know how your categories of interest are realized in it, commonly, there is no labeled data at hand, and the performance of domain adaptation methods is unsatisfactory.
no code implementations • 25 Nov 2019 • Liat Ein-Dor, Eyal Shnarch, Lena Dankin, Alon Halfon, Benjamin Sznajder, Ariel Gera, Carlos Alzate, Martin Gleize, Leshem Choshen, Yufang Hou, Yonatan Bilu, Ranit Aharonov, Noam Slonim
One of the main tasks in argument mining is the retrieval of argumentative content pertaining to a given topic.
no code implementations • CONLL 2019 • Leshem Choshen, Omri Abend
We show that the state-of-the-art Transformer MT model is not biased towards monotonic reordering (unlike previous recurrent neural network models), but that nevertheless, long-distance dependencies remain a challenge for the model.
1 code implementation • 15 Sep 2019 • Leshem Choshen, Omri Abend
We show that the state of the art Transformer Machine Translation (MT) model is not biased towards monotonic reordering (unlike previous recurrent neural network models), but that nevertheless, long-distance dependencies remain a challenge for the model.
no code implementations • ACL 2019 • Martin Gleize, Eyal Shnarch, Leshem Choshen, Lena Dankin, Guy Moshkowich, Ranit Aharonov, Noam Slonim
With the advancement in argument detection, we suggest to pay more attention to the challenging task of identifying the more convincing arguments.
no code implementations • ICLR 2020 • Leshem Choshen, Lior Fox, Zohar Aizenbud, Omri Abend
Reinforcement learning (RL) is frequently used to increase performance in text generation tasks, including machine translation (MT), notably through the use of Minimum Risk Training (MRT) and Generative Adversarial Networks (GAN).
1 code implementation • WS 2019 • Yoav Kantor, Yoav Katz, Leshem Choshen, Edo Cohen-Karlik, Naftali Liberman, Assaf Toledo, Amir Menczel, Noam Slonim
We also present a spellchecker created for this task which outperforms standard spellcheckers tested on the task of spellchecking.
Ranked #4 on
Grammatical Error Correction
on BEA-2019 (test)
no code implementations • ICML 2020 • Guy Hacohen, Leshem Choshen, Daphna Weinshall
We further show that this pattern of results reflects the interplay between the way neural networks learn benchmark datasets.
1 code implementation • ACL 2019 • Leshem Choshen, Dan Eldad, Daniel Hershcovich, Elior Sulem, Omri Abend
The non-indexed parts of the Internet (the Darknet) have become a haven for both legal and illegal anonymous activity.
no code implementations • SEMEVAL 2019 • Daniel Hershcovich, Zohar Aizenbud, Leshem Choshen, Elior Sulem, Ari Rappoport, Omri Abend
We present the SemEval 2019 shared task on UCCA parsing in English, German and French, and discuss the participating systems and results.
1 code implementation • ACL 2018 • Leshem Choshen, Omri Abend
The prevalent use of too few references for evaluating text-to-text generation is known to bias estimates of their quality (henceforth, low coverage bias or LCB).
no code implementations • ACL 2018 • Eyal Shnarch, Carlos Alzate, Lena Dankin, Martin Gleize, Yufang Hou, Leshem Choshen, Ranit Aharonov, Noam Slonim
We propose a methodology to blend high quality but scarce strong labeled data with noisy but abundant weak labeled data during the training of neural networks.
no code implementations • 31 May 2018 • Daniel Hershcovich, Leshem Choshen, Elior Sulem, Zohar Aizenbud, Ari Rappoport, Omri Abend
Given the success of recent semantic parsing shared tasks (on SDP and AMR), we expect the task to have a significant contribution to the advancement of UCCA parsing in particular, and semantic parsing in general.
1 code implementation • ACL 2018 • Leshem Choshen, Omri Abend
Metric validation in Grammatical Error Correction (GEC) is currently done by observing the correlation between human and metric-induced rankings.
1 code implementation • 30 Apr 2018 • Leshem Choshen, Omri Abend
The prevalent use of too few references for evaluating text-to-text generation is known to bias estimates of their quality ({\it low coverage bias} or LCB).
1 code implementation • ICLR 2018 • Leshem Choshen, Lior Fox, Yonatan Loewenstein
We compare our approach to commonly used RL techniques, and show that using $E$-values improves learning and performance over traditional counters.
1 code implementation • NAACL 2018 • Leshem Choshen, Omri Abend
We propose USim, a semantic measure for Grammatical Error Correction (GEC) that measures the semantic faithfulness of the output to the source, thereby complementing existing reference-less measures (RLMs) for measuring the output's grammaticality.