Search Results for author: Leshem Choshen

Found 44 papers, 25 papers with code

Q^{2}: Evaluating Factual Consistency in Knowledge-Grounded Dialogues via Question Generation and Question Answering

no code implementations EMNLP 2021 Or Honovich, Leshem Choshen, Roee Aharoni, Ella Neeman, Idan Szpektor, Omri Abend

Neural knowledge-grounded generative models for dialogue often produce content that is factually inconsistent with the knowledge they rely on, making them unreliable and limiting their applicability.

Abstractive Text Summarization Natural Language Inference +3

Active Learning for BERT: An Empirical Study

1 code implementation EMNLP 2020 Liat Ein-Dor, Alon Halfon, Ariel Gera, Eyal Shnarch, Lena Dankin, Leshem Choshen, Marina Danilevsky, Ranit Aharonov, Yoav Katz, Noam Slonim

Here, we present a large-scale empirical study on active learning techniques for BERT-based classification, addressing a diverse set of AL strategies and datasets.

Active Learning Binary text classification +3

Jump to Conclusions: Short-Cutting Transformers With Linear Transformations

1 code implementation16 Mar 2023 Alexander Yom Din, Taelin Karidi, Leshem Choshen, Mor Geva

Moreover, in the context of language modeling, our method allows "peeking" into early layer representations of GPT-2 and BERT, showing that often LMs already predict the final output in early layers.

Decision Making Language Modelling

Knowledge is a Region in Weight Space for Fine-tuned Language Models

no code implementations9 Feb 2023 Almog Gueta, Elad Venezian, Colin Raffel, Noam Slonim, Yoav Katz, Leshem Choshen

Specifically, we show that starting from the center of the region is as good or better than the pre-trained model in 11 of 12 datasets and improves accuracy by 3. 06 on average.

Call for Papers -- The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus

no code implementations27 Jan 2023 Alex Warstadt, Leshem Choshen, Aaron Mueller, Adina Williams, Ethan Wilcox, Chengxu Zhuang

In partnership with CoNLL and CMCL, we provide a platform for approaches to pretraining with a limited-size corpus sourced from data inspired by the input to children.

Language Acquisition Language Modelling +1

ColD Fusion: Collaborative Descent for Distributed Multitask Finetuning

no code implementations2 Dec 2022 Shachar Don-Yehiya, Elad Venezian, Colin Raffel, Noam Slonim, Yoav Katz, Leshem Choshen

We show that ColD Fusion yields comparable benefits to multitask pretraining by producing a model that (a) attains strong performance on all of the datasets it was multitask trained on and (b) is a better starting point for finetuning on unseen datasets.

DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering

1 code implementation10 Nov 2022 Ella Neeman, Roee Aharoni, Or Honovich, Leshem Choshen, Idan Szpektor, Omri Abend

Question answering models commonly have access to two sources of "knowledge" during inference time: (1) parametric knowledge - the factual knowledge encoded in the model weights, and (2) contextual knowledge - external knowledge (e. g., a Wikipedia passage) given to the model to generate a grounded answer.

Data Augmentation Natural Questions +1

Where to start? Analyzing the potential value of intermediate models

no code implementations31 Oct 2022 Leshem Choshen, Elad Venezian, Shachar Don-Yehia, Noam Slonim, Yoav Katz

Such a model, finetuned on some source dataset, may provide a better starting point for a new finetuning process on a desired target dataset.

Reinforcement Learning with Large Action Spaces for Neural Machine Translation

no code implementations COLING 2022 Asaf Yehudai, Leshem Choshen, Lior Fox, Omri Abend

Applying Reinforcement learning (RL) following maximum likelihood estimation (MLE) pre-training is a versatile method for enhancing neural machine translation (NMT) performance.

Machine Translation NMT +5

PreQuEL: Quality Estimation of Machine Translation Outputs in Advance

1 code implementation18 May 2022 Shachar Don-Yehiya, Leshem Choshen, Omri Abend

We show that this augmentation method can improve the performance of the Quality-Estimation task as well.

Data Augmentation Machine Translation +1

Some Grammatical Errors are Frequent, Others are Important

1 code implementation11 May 2022 Leshem Choshen, Ofir Shifman, Omri Abend

In Grammatical Error Correction, systems are evaluated by the number of errors they correct.

Grammatical Error Correction

Fusing finetuned models for better pretraining

1 code implementation6 Apr 2022 Leshem Choshen, Elad Venezian, Noam Slonim, Yoav Katz

We also show that fusing is often better than intertraining.

On Neurons Invariant to Sentence Structural Changes in Neural Machine Translation

1 code implementation6 Oct 2021 Gal Patel, Leshem Choshen, Omri Abend

We present a methodology that explores how sentence structure is reflected in neural representations of machine translation systems.

Machine Translation Translation

The Grammar-Learning Trajectories of Neural Language Models

1 code implementation ACL 2022 Leshem Choshen, Guy Hacohen, Daphna Weinshall, Omri Abend

These findings suggest that there is some mutual inductive bias that underlies these models' learning of linguistic phenomena.

Inductive Bias

ComSum: Commit Messages Summarization and Meaning Preservation

1 code implementation23 Aug 2021 Leshem Choshen, Idan Amit

We present ComSum, a data set of 7 million commit messages for text summarization.

Text Summarization

Part of Speech and Universal Dependency effects on English Arabic Machine Translation

no code implementations1 Jun 2021 Ofek Rafaeli, Omri Abend, Leshem Choshen, Dmitry Nikolaev

In this research paper, I will elaborate on a method to evaluate machine translation models based on their performance on underlying syntactical phenomena between English and Arabic languages.

BIG-bench Machine Learning Machine Translation +1

$Q^{2}$: Evaluating Factual Consistency in Knowledge-Grounded Dialogues via Question Generation and Question Answering

1 code implementation16 Apr 2021 Or Honovich, Leshem Choshen, Roee Aharoni, Ella Neeman, Idan Szpektor, Omri Abend

Neural knowledge-grounded generative models for dialogue often produce content that is factually inconsistent with the knowledge they rely on, making them unreliable and limiting their applicability.

Abstractive Text Summarization Dialogue Evaluation +4

Mediators in Determining what Processing BERT Performs First

1 code implementation NAACL 2021 Aviv Slobodkin, Leshem Choshen, Omri Abend

Probing neural models for the ability to perform downstream tasks using their activation patterns is often used to localize what parts of the network specialize in performing what tasks.

SERRANT: a syntactic classifier for English Grammatical Error Types

1 code implementation6 Apr 2021 Leshem Choshen, Matanel Oren, Dmitry Nikolaev, Omri Abend

SERRANT is a system and code for automatic classification of English grammatical errors that combines SErCl and ERRANT.

General Classification

Enhancing the Transformer Decoder with Transition-based Syntax

1 code implementation29 Jan 2021 Leshem Choshen, Omri Abend

Notwithstanding recent advances, syntactic generalization remains a challenge for text decoders.

Machine Translation Text Generation +1

Classifying Syntactic Errors in Learner Language

1 code implementation CONLL 2020 Leshem Choshen, Dmitry Nikolaev, Yevgeni Berzak, Omri Abend

We present a method for classifying syntactic errors in learner language, namely errors whose correction alters the morphosyntactic structure of a sentence.

Classification General Classification +1

Unsupervised Expressive Rules Provide Explainability and Assist Human Experts Grasping New Domains

no code implementations Findings of the Association for Computational Linguistics 2020 Eyal Shnarch, Leshem Choshen, Guy Moshkowich, Noam Slonim, Ranit Aharonov

Approaching new data can be quite deterrent; you do not know how your categories of interest are realized in it, commonly, there is no labeled data at hand, and the performance of domain adaptation methods is unsatisfactory.

Domain Adaptation

Automatically Extracting Challenge Sets for Non-Local Phenomena in Neural Machine Translation

no code implementations CONLL 2019 Leshem Choshen, Omri Abend

We show that the state-of-the-art Transformer MT model is not biased towards monotonic reordering (unlike previous recurrent neural network models), but that nevertheless, long-distance dependencies remain a challenge for the model.

Machine Translation Translation

Automatically Extracting Challenge Sets for Non local Phenomena in Neural Machine Translation

1 code implementation15 Sep 2019 Leshem Choshen, Omri Abend

We show that the state of the art Transformer Machine Translation (MT) model is not biased towards monotonic reordering (unlike previous recurrent neural network models), but that nevertheless, long-distance dependencies remain a challenge for the model.

Machine Translation Translation

Are You Convinced? Choosing the More Convincing Evidence with a Siamese Network

no code implementations ACL 2019 Martin Gleize, Eyal Shnarch, Leshem Choshen, Lena Dankin, Guy Moshkowich, Ranit Aharonov, Noam Slonim

With the advancement in argument detection, we suggest to pay more attention to the challenging task of identifying the more convincing arguments.

On the Weaknesses of Reinforcement Learning for Neural Machine Translation

no code implementations ICLR 2020 Leshem Choshen, Lior Fox, Zohar Aizenbud, Omri Abend

Reinforcement learning (RL) is frequently used to increase performance in text generation tasks, including machine translation (MT), notably through the use of Minimum Risk Training (MRT) and Generative Adversarial Networks (GAN).

Machine Translation reinforcement-learning +3

The Language of Legal and Illegal Activity on the Darknet

1 code implementation ACL 2019 Leshem Choshen, Dan Eldad, Daniel Hershcovich, Elior Sulem, Omri Abend

The non-indexed parts of the Internet (the Darknet) have become a haven for both legal and illegal anonymous activity.

POS

SemEval-2019 Task 1: Cross-lingual Semantic Parsing with UCCA

no code implementations SEMEVAL 2019 Daniel Hershcovich, Zohar Aizenbud, Leshem Choshen, Elior Sulem, Ari Rappoport, Omri Abend

We present the SemEval 2019 shared task on UCCA parsing in English, German and French, and discuss the participating systems and results.

UCCA Parsing

Inherent Biases in Reference-based Evaluation for Grammatical Error Correction

1 code implementation ACL 2018 Leshem Choshen, Omri Abend

The prevalent use of too few references for evaluating text-to-text generation is known to bias estimates of their quality (henceforth, low coverage bias or LCB).

Grammatical Error Correction Text Generation +1

SemEval 2019 Shared Task: Cross-lingual Semantic Parsing with UCCA - Call for Participation

no code implementations31 May 2018 Daniel Hershcovich, Leshem Choshen, Elior Sulem, Zohar Aizenbud, Ari Rappoport, Omri Abend

Given the success of recent semantic parsing shared tasks (on SDP and AMR), we expect the task to have a significant contribution to the advancement of UCCA parsing in particular, and semantic parsing in general.

UCCA Parsing

Automatic Metric Validation for Grammatical Error Correction

1 code implementation ACL 2018 Leshem Choshen, Omri Abend

Metric validation in Grammatical Error Correction (GEC) is currently done by observing the correlation between human and metric-induced rankings.

Grammatical Error Correction

Inherent Biases in Reference based Evaluation for Grammatical Error Correction and Text Simplification

1 code implementation30 Apr 2018 Leshem Choshen, Omri Abend

The prevalent use of too few references for evaluating text-to-text generation is known to bias estimates of their quality ({\it low coverage bias} or LCB).

Grammatical Error Correction Text Generation +1

DORA The Explorer: Directed Outreaching Reinforcement Action-Selection

1 code implementation ICLR 2018 Leshem Choshen, Lior Fox, Yonatan Loewenstein

We compare our approach to commonly used RL techniques, and show that using $E$-values improves learning and performance over traditional counters.

Reinforcement Learning (RL)

Reference-less Measure of Faithfulness for Grammatical Error Correction

1 code implementation NAACL 2018 Leshem Choshen, Omri Abend

We propose USim, a semantic measure for Grammatical Error Correction (GEC) that measures the semantic faithfulness of the output to the source, thereby complementing existing reference-less measures (RLMs) for measuring the output's grammaticality.

Grammatical Error Correction

Cannot find the paper you are looking for? You can Submit a new open access paper.