Search Results for author: Ariel Gera

Found 19 papers, 9 papers with code

Active Learning for BERT: An Empirical Study

1 code implementation EMNLP 2020 Liat Ein-Dor, Alon Halfon, Ariel Gera, Eyal Shnarch, Lena Dankin, Leshem Choshen, Marina Danilevsky, Ranit Aharonov, Yoav Katz, Noam Slonim

Here, we present a large-scale empirical study on active learning techniques for BERT-based classification, addressing a diverse set of AL strategies and datasets.

Active Learning Binary text classification +3

WildIFEval: Instruction Following in the Wild

1 code implementation9 Mar 2025 Gili Lior, Asaf Yehudai, Ariel Gera, Liat Ein-Dor

In this work, we introduce WildIFEval - a large-scale dataset of 12K real user instructions with diverse, multi-constraint conditions.

Instruction Following

The Mighty ToRR: A Benchmark for Table Reasoning and Robustness

1 code implementation26 Feb 2025 Shir Ashury-Tahan, Yifan Mai, Rajmohan C, Ariel Gera, Yotam Perlitz, Asaf Yehudai, Elron Bandel, Leshem Choshen, Eyal Shnarch, Percy Liang, Michal Shmueli-Scheuer

ToRR goes beyond model performance rankings, and is designed to reflect whether models can handle tabular data consistently and robustly, across a variety of common table representation formats.

JuStRank: Benchmarking LLM Judges for System Ranking

no code implementations12 Dec 2024 Ariel Gera, Odellia Boni, Yotam Perlitz, Roy Bar-Haim, Lilach Eden, Asaf Yehudai

To address this gap, we conduct the first large-scale study of LLM judges as system rankers.

Benchmarking

Do These LLM Benchmarks Agree? Fixing Benchmark Evaluation with BenchBench

1 code implementation18 Jul 2024 Yotam Perlitz, Ariel Gera, Ofir Arviv, Asaf Yehudai, Elron Bandel, Eyal Shnarch, Michal Shmueli-Scheuer, Leshem Choshen

Despite the crucial role of BAT for benchmark builders and consumers, there are no standardized procedures for such agreement testing.

Language Modelling

Label-Efficient Model Selection for Text Generation

no code implementations12 Feb 2024 Shir Ashury-Tahan, Ariel Gera, Benjamin Sznajder, Leshem Choshen, Liat Ein-Dor, Eyal Shnarch

Model selection for a given target task can be costly, as it may entail extensive annotation of the quality of outputs of different models.

model Model Selection +1

Unitxt: Flexible, Shareable and Reusable Data Preparation and Evaluation for Generative AI

1 code implementation25 Jan 2024 Elron Bandel, Yotam Perlitz, Elad Venezian, Roni Friedman-Melamed, Ofir Arviv, Matan Orbach, Shachar Don-Yehyia, Dafna Sheinwald, Ariel Gera, Leshem Choshen, Michal Shmueli-Scheuer, Yoav Katz

In the dynamic landscape of generative NLP, traditional text processing pipelines limit research flexibility and reproducibility, as they are tailored to specific dataset, task, and model combinations.

Efficient Benchmarking of Language Models

no code implementations22 Aug 2023 Yotam Perlitz, Elron Bandel, Ariel Gera, Ofir Arviv, Liat Ein-Dor, Eyal Shnarch, Noam Slonim, Michal Shmueli-Scheuer, Leshem Choshen

The increasing versatility of language models (LMs) has given rise to a new class of benchmarks that comprehensively assess a broad range of capabilities.

Benchmarking

Active Learning for Natural Language Generation

no code implementations24 May 2023 Yotam Perlitz, Ariel Gera, Michal Shmueli-Scheuer, Dafna Sheinwald, Noam Slonim, Liat Ein-Dor

In this paper, we present a first systematic study of active learning for NLG, considering a diverse set of tasks and multiple leading selection strategies, and harnessing a strong instruction-tuned model.

Active Learning text-classification +2

The Benefits of Bad Advice: Autocontrastive Decoding across Model Layers

1 code implementation2 May 2023 Ariel Gera, Roni Friedman, Ofir Arviv, Chulaka Gunasekara, Benjamin Sznajder, Noam Slonim, Eyal Shnarch

Applying language models to natural language processing tasks typically relies on the representations in the final model layer, as intermediate hidden layer representations are presumed to be less informative.

Language Modeling Language Modelling +1

Zero-Shot Text Classification with Self-Training

1 code implementation31 Oct 2022 Ariel Gera, Alon Halfon, Eyal Shnarch, Yotam Perlitz, Liat Ein-Dor, Noam Slonim

Recent advances in large pretrained language models have increased attention to zero-shot text classification.

Natural Language Inference text-classification +2

Cluster & Tune: Enhance BERT Performance in Low Resource Text Classification

no code implementations1 Jan 2021 Eyal Shnarch, Ariel Gera, Alon Halfon, Lena Dankin, Leshem Choshen, Ranit Aharonov, Noam Slonim

In such low resources scenarios, we suggest performing an unsupervised classification task prior to fine-tuning on the target task.

Clustering General Classification +2

Financial Event Extraction Using Wikipedia-Based Weak Supervision

no code implementations WS 2019 Liat Ein-Dor, Ariel Gera, Orith Toledo-Ronen, Alon Halfon, Benjamin Sznajder, Lena Dankin, Yonatan Bilu, Yoav Katz, Noam Slonim

Extraction of financial and economic events from text has previously been done mostly using rule-based methods, with more recent works employing machine learning techniques.

BIG-bench Machine Learning Event Extraction

Argument Invention from First Principles

no code implementations ACL 2019 Yonatan Bilu, Ariel Gera, Daniel Hershcovich, Benjamin Sznajder, Dan Lahav, Guy Moshkowich, Anael Malet, Assaf Gavron, Noam Slonim

In this work we aim to explicitly define a taxonomy of such principled recurring arguments, and, given a controversial topic, to automatically identify which of these arguments are relevant to the topic.

Controversy in Context

no code implementations20 Aug 2019 Benjamin Sznajder, Ariel Gera, Yonatan Bilu, Dafna Sheinwald, Ella Rabinovich, Ranit Aharonov, David Konopnicki, Noam Slonim

With the growing interest in social applications of Natural Language Processing and Computational Argumentation, a natural question is how controversial a given concept is.

Cannot find the paper you are looking for? You can Submit a new open access paper.