Search Results for author: Alon Albalak

Found 23 papers, 15 papers with code

Big-Math: A Large-Scale, High-Quality Math Dataset for Reinforcement Learning in Language Models

1 code implementation24 Feb 2025 Alon Albalak, Duy Phung, Nathan Lile, Rafael Rafailov, Kanishk Gandhi, Louis Castricato, Anikait Singh, Chase Blagden, Violet Xiang, Dakota Mahan, Nick Haber

However, existing open math datasets either contain a small collection of high-quality, human-written problems or a large corpus of machine-generated problems of uncertain quality, forcing researchers to choose between quality and quantity.

GSM8K Math +2

Towards System 2 Reasoning in LLMs: Learning How to Think With Meta Chain-of-Though

no code implementations8 Jan 2025 Violet Xiang, Charlie Snell, Kanishk Gandhi, Alon Albalak, Anikait Singh, Chase Blagden, Duy Phung, Rafael Rafailov, Nathan Lile, Dakota Mahan, Louis Castricato, Jan-Philipp Franken, Nick Haber, Chelsea Finn

We propose a novel framework, Meta Chain-of-Thought (Meta-CoT), which extends traditional Chain-of-Thought (CoT) by explicitly modeling the underlying reasoning required to arrive at a particular CoT.

Synthetic Data Generation

Generative Reward Models

no code implementations2 Oct 2024 Dakota Mahan, Duy Van Phung, Rafael Rafailov, Chase Blagden, Nathan Lile, Louis Castricato, Jan-Philipp Fränken, Chelsea Finn, Alon Albalak

We introduce GenRM, an iterative algorithm that trains an LLM on self-generated reasoning traces, leading to synthetic preference labels matching human preference judgments.

reinforcement-learning Reinforcement Learning

Generalization v.s. Memorization: Tracing Language Models' Capabilities Back to Pretraining Data

no code implementations20 Jul 2024 Xinyi Wang, Antonis Antoniades, Yanai Elazar, Alfonso Amayuelas, Alon Albalak, Kexun Zhang, William Yang Wang

Furthermore, while model performance improves across all tasks as LLM size increases, only factual question answering shows an increase in memorization, whereas machine translation and reasoning tasks exhibit greater generalization, producing more novel outputs.

Language Modelling Machine Translation +6

Efficient Online Data Mixing For Language Model Pre-Training

1 code implementation5 Dec 2023 Alon Albalak, Liangming Pan, Colin Raffel, William Yang Wang

The data used to pretrain large language models has a decisive impact on a model's downstream performance, which has led to a large body of work on data selection methods that aim to automatically determine the most suitable data to use for pretraining.

Language Modeling Language Modelling +1

Logic-LM: Empowering Large Language Models with Symbolic Solvers for Faithful Logical Reasoning

1 code implementation20 May 2023 Liangming Pan, Alon Albalak, Xinyi Wang, William Yang Wang

We also introduce a self-refinement module, which utilizes the symbolic solver's error messages to revise symbolic formalizations.

Logical Reasoning

Improving Few-Shot Generalization by Exploring and Exploiting Auxiliary Data

1 code implementation NeurIPS 2023 Alon Albalak, Colin Raffel, William Yang Wang

In this work, we focus on Few-shot Learning with Auxiliary Data (FLAD), a training paradigm that assumes access to auxiliary data during few-shot learning in hopes of improving generalization.

Few-Shot Learning

CausalDialogue: Modeling Utterance-level Causality in Conversations

1 code implementation20 Dec 2022 Yi-Lin Tuan, Alon Albalak, Wenda Xu, Michael Saxon, Connor Pryor, Lise Getoor, William Yang Wang

Despite their widespread adoption, neural conversation models have yet to exhibit natural chat capabilities with humans.

Dialogue Generation Diversity

An Exploration of Data Efficiency in Intra-Dataset Task Transfer for Dialog Understanding

no code implementations21 Oct 2022 Josiah Ross, Luke Yoffe, Alon Albalak, William Yang Wang

Transfer learning is an exciting area of Natural Language Processing that has the potential to both improve model performance and increase data efficiency.

Transfer Learning

Data-Efficiency with a Single GPU: An Exploration of Transfer Methods for Small Language Models

no code implementations8 Oct 2022 Alon Albalak, Akshat Shrivastava, Chinnadhurai Sankar, Adithya Sagar, Mike Ross

Multi-task learning (MTL), instruction tuning, and prompting have recently been shown to improve the generalizability of large language models to new tasks.

Multi-Task Learning

Emotion Recognition in Conversation using Probabilistic Soft Logic

no code implementations14 Jul 2022 Eriq Augustine, Pegah Jandaghi, Alon Albalak, Connor Pryor, Charles Dickens, William Wang, Lise Getoor

Creating agents that can both appropriately respond to conversations and understand complex human linguistic tendencies and social cues has been a long standing challenge in the NLP community.

Emotion Recognition in Conversation Logical Reasoning +2

NeuPSL: Neural Probabilistic Soft Logic

1 code implementation27 May 2022 Connor Pryor, Charles Dickens, Eriq Augustine, Alon Albalak, William Wang, Lise Getoor

In this paper, we introduce Neural Probabilistic Soft Logic (NeuPSL), a novel neuro-symbolic (NeSy) framework that unites state-of-the-art symbolic reasoning with the low-level perception of deep neural networks.

FETA: A Benchmark for Few-Sample Task Transfer in Open-Domain Dialogue

1 code implementation12 May 2022 Alon Albalak, Yi-Lin Tuan, Pegah Jandaghi, Connor Pryor, Luke Yoffe, Deepak Ramachandran, Lise Getoor, Jay Pujara, William Yang Wang

Task transfer, transferring knowledge contained in related tasks, holds the promise of reducing the quantity of labeled data required to fine-tune language models.

Dialogue Understanding Domain Adaptation +1

Addressing Issues of Cross-Linguality in Open-Retrieval Question Answering Systems For Emergent Domains

1 code implementation26 Jan 2022 Alon Albalak, Sharon Levy, William Yang Wang

Open-retrieval question answering systems are generally trained and tested on large datasets in well-established domains.

Question Answering Retrieval +1

D-REX: Dialogue Relation Extraction with Explanations

1 code implementation NLP4ConvAI (ACL) 2022 Alon Albalak, Varun Embar, Yi-Lin Tuan, Lise Getoor, William Yang Wang

Existing research studies on cross-sentence relation extraction in long-form multi-party conversations aim to improve relation extraction without considering the explainability of such methods.

Dialog Relation Extraction Relation +3

Modeling Disclosive Transparency in NLP Application Descriptions

1 code implementation EMNLP 2021 Michael Saxon, Sharon Levy, Xinyi Wang, Alon Albalak, William Yang Wang

Broader disclosive transparency$-$truth and clarity in communication regarding the function of AI systems$-$is widely considered desirable.

Fairness Language Modeling +2

Cannot find the paper you are looking for? You can Submit a new open access paper.