Search Results for author: Roee Aharoni

Found 33 papers, 12 papers with code

Morphological Inflection Generation with Hard Monotonic Attention

1 code implementation ACL 2017 Roee Aharoni, Yoav Goldberg

We present a neural model for morphological inflection generation which employs a hard attention mechanism, inspired by the nearly-monotonic alignment commonly found between the characters in a word and the characters in its inflection.

Hard Attention Morphological Inflection

Towards String-to-Tree Neural Machine Translation

no code implementations ACL 2017 Roee Aharoni, Yoav Goldberg

We present a simple method to incorporate syntactic information about the target language in a neural machine translation system by translating into linearized, lexicalized constituency trees.

Machine Translation NMT +1

Split and Rephrase: Better Evaluation and a Stronger Baseline

2 code implementations2 May 2018 Roee Aharoni, Yoav Goldberg

To aid this, we present a new train-development-test data split and neural models augmented with a copy-mechanism, outperforming the best reported baseline by 8. 68 BLEU and fostering further progress on the task.

Memorization Sentence +1

Split and Rephrase: Better Evaluation and Stronger Baselines

1 code implementation ACL 2018 Roee Aharoni, Yoav Goldberg

To aid this, we present a new train-development-test data split and neural models augmented with a copy-mechanism, outperforming the best reported baseline by 8. 68 BLEU and fostering further progress on the task.

Machine Translation Memorization +2

Massively Multilingual Neural Machine Translation

no code implementations NAACL 2019 Roee Aharoni, Melvin Johnson, Orhan Firat

Our experiments on a large-scale dataset with 102 languages to and from English and up to one million examples per direction also show promising results, surpassing strong bilingual baselines and encouraging future work on massively multilingual NMT.

Machine Translation NMT +1

Filling Gender & Number Gaps in Neural Machine Translation with Black-box Context Injection

no code implementations8 Mar 2019 Amit Moryossef, Roee Aharoni, Yoav Goldberg

When translating from a language that does not morphologically mark information such as gender and number into a language that does, translation systems must "guess" this missing information, often leading to incorrect translations in the given context.

Machine Translation Translation

The Missing Ingredient in Zero-Shot Neural Machine Translation

no code implementations17 Mar 2019 Naveen Arivazhagan, Ankur Bapna, Orhan Firat, Roee Aharoni, Melvin Johnson, Wolfgang Macherey

Multilingual Neural Machine Translation (NMT) models are capable of translating between multiple source and target languages.

Machine Translation NMT +1

Filling Gender \& Number Gaps in Neural Machine Translation with Black-box Context Injection

no code implementations WS 2019 Amit Moryossef, Roee Aharoni, Yoav Goldberg

When translating from a language that does not morphologically mark information such as gender and number into a language that does, translation systems must {``}guess{''} this missing information, often leading to incorrect translations in the given context.

Machine Translation Translation

Diversify Your Datasets: Analyzing Generalization via Controlled Variance in Adversarial Datasets

1 code implementation CONLL 2019 Ohad Rozen, Vered Shwartz, Roee Aharoni, Ido Dagan

Phenomenon-specific "adversarial" datasets have been recently designed to perform targeted stress-tests for particular inference types.

Unsupervised Domain Clusters in Pretrained Language Models

1 code implementation ACL 2020 Roee Aharoni, Yoav Goldberg

The notion of "in-domain data" in NLP is often over-simplistic and vague, as textual data varies in many nuanced linguistic aspects such as topic, style or level of formality.

Machine Translation Sentence +1

Real-Time Sign Language Detection using Human Pose Estimation

no code implementations11 Aug 2020 Amit Moryossef, Ioannis Tsochantaridis, Roee Aharoni, Sarah Ebling, Srini Narayanan

We propose a lightweight real-time sign language detection model, as we identify the need for such a case in videoconferencing.

Optical Flow Estimation Pose Estimation

KoBE: Knowledge-Based Machine Translation Evaluation

1 code implementation Findings of the Association for Computational Linguistics 2020 Zorik Gekhman, Roee Aharoni, Genady Beryozkin, Markus Freitag, Wolfgang Macherey

Our approach achieves the highest correlation with human judgements on 9 out of the 18 language pairs from the WMT19 benchmark for evaluation without references, which is the largest number of wins for a single evaluation method on this task.

Machine Translation Sentence +1

$Q^{2}$: Evaluating Factual Consistency in Knowledge-Grounded Dialogues via Question Generation and Question Answering

1 code implementation16 Apr 2021 Or Honovich, Leshem Choshen, Roee Aharoni, Ella Neeman, Idan Szpektor, Omri Abend

Neural knowledge-grounded generative models for dialogue often produce content that is factually inconsistent with the knowledge they rely on, making them unreliable and limiting their applicability.

Abstractive Text Summarization Dialogue Evaluation +4

DisentQA: Disentangling Parametric and Contextual Knowledge with Counterfactual Question Answering

1 code implementation10 Nov 2022 Ella Neeman, Roee Aharoni, Or Honovich, Leshem Choshen, Idan Szpektor, Omri Abend

Question answering models commonly have access to two sources of "knowledge" during inference time: (1) parametric knowledge - the factual knowledge encoded in the model weights, and (2) contextual knowledge - external knowledge (e. g., a Wikipedia passage) given to the model to generate a grounded answer.

counterfactual Data Augmentation +2

Multilingual Sequence-to-Sequence Models for Hebrew NLP

no code implementations19 Dec 2022 Matan Eyal, Hila Noga, Roee Aharoni, Idan Szpektor, Reut Tsarfaty

We demonstrate that by casting tasks in the Hebrew NLP pipeline as text-to-text tasks, we can leverage powerful multilingual, pretrained sequence-to-sequence models as mT5, eliminating the need for a specialized, morpheme-based, separately fine-tuned decoder.

named-entity-recognition Named Entity Recognition +1

mFACE: Multilingual Summarization with Factual Consistency Evaluation

no code implementations20 Dec 2022 Roee Aharoni, Shashi Narayan, Joshua Maynez, Jonathan Herzig, Elizabeth Clark, Mirella Lapata

Abstractive summarization has enjoyed renewed interest in recent years, thanks to pre-trained language models and the availability of large-scale datasets.

Abstractive Text Summarization

q2d: Turning Questions into Dialogs to Teach Models How to Search

no code implementations27 Apr 2023 Yonatan Bitton, Shlomi Cohen-Ganor, Ido Hakimi, Yoad Lewenberg, Roee Aharoni, Enav Weinreb

One of the exciting capabilities of recent language models for dialog is their ability to independently search for relevant information to ground a given dialog response.

Language Modelling Large Language Model +1

Surfacing Biases in Large Language Models using Contrastive Input Decoding

no code implementations12 May 2023 Gal Yona, Or Honovich, Itay Laish, Roee Aharoni

We use CID to highlight context-specific biases that are hard to detect with standard decoding strategies and quantify the effect of different input perturbations.

Text Generation

What You See is What You Read? Improving Text-Image Alignment Evaluation

1 code implementation NeurIPS 2023 Michal Yarom, Yonatan Bitton, Soravit Changpinyo, Roee Aharoni, Jonathan Herzig, Oran Lang, Eran Ofek, Idan Szpektor

Automatically determining whether a text and a corresponding image are semantically aligned is a significant challenge for vision-language models, with applications in generative text-to-image and image-to-text tasks.

Question Answering Question Generation +5

TrueTeacher: Learning Factual Consistency Evaluation with Large Language Models

1 code implementation18 May 2023 Zorik Gekhman, Jonathan Herzig, Roee Aharoni, Chen Elkind, Idan Szpektor

Factual consistency evaluation is often conducted using Natural Language Inference (NLI) models, yet these models exhibit limited success in evaluating summaries.

Natural Language Inference Synthetic Data Generation

A Comprehensive Evaluation of Tool-Assisted Generation Strategies

no code implementations16 Oct 2023 Alon Jacovi, Avi Caciularu, Jonathan Herzig, Roee Aharoni, Bernd Bohnet, Mor Geva

A growing area of research investigates augmenting language models with tools (e. g., search engines, calculators) to overcome their shortcomings (e. g., missing or incorrect knowledge, incorrect logical inferences).

Retrieval

Multilingual Instruction Tuning With Just a Pinch of Multilinguality

no code implementations3 Jan 2024 Uri Shaham, Jonathan Herzig, Roee Aharoni, Idan Szpektor, Reut Tsarfaty, Matan Eyal

As instruction-tuned large language models (LLMs) gain global adoption, their ability to follow instructions in multiple languages becomes increasingly crucial.

Cross-Lingual Transfer Instruction Following

Narrowing the Knowledge Evaluation Gap: Open-Domain Question Answering with Multi-Granularity Answers

no code implementations9 Jan 2024 Gal Yona, Roee Aharoni, Mor Geva

In this work, we propose GRANOLA QA, a novel evaluation setting where a predicted answer is evaluated in terms of accuracy and informativeness against a set of multi-granularity answers.

Informativeness Open-Domain Question Answering

A Chain-of-Thought Is as Strong as Its Weakest Link: A Benchmark for Verifiers of Reasoning Chains

no code implementations1 Feb 2024 Alon Jacovi, Yonatan Bitton, Bernd Bohnet, Jonathan Herzig, Or Honovich, Michael Tseng, Michael Collins, Roee Aharoni, Mor Geva

REVEAL includes comprehensive labels for the relevance, attribution to evidence passages, and logical correctness of each reasoning step in a language model's answer, across a variety of datasets and state-of-the-art language models.

Open-Domain Question Answering

MiMiC: Minimally Modified Counterfactuals in the Representation Space

no code implementations15 Feb 2024 Shashwat Singh, Shauli Ravfogel, Jonathan Herzig, Roee Aharoni, Ryan Cotterell, Ponnurangam Kumaraguru

We demonstrate the effectiveness of the proposed approaches in mitigating bias in multiclass classification and in reducing the generation of toxic language, outperforming strong baselines.

Q^{2}: Evaluating Factual Consistency in Knowledge-Grounded Dialogues via Question Generation and Question Answering

no code implementations EMNLP 2021 Or Honovich, Leshem Choshen, Roee Aharoni, Ella Neeman, Idan Szpektor, Omri Abend

Neural knowledge-grounded generative models for dialogue often produce content that is factually inconsistent with the knowledge they rely on, making them unreliable and limiting their applicability.

Abstractive Text Summarization Natural Language Inference +3

Cannot find the paper you are looking for? You can Submit a new open access paper.