Search Results for author: Alon Albalak

Found 15 papers, 10 papers with code

Efficient Online Data Mixing For Language Model Pre-Training

no code implementations5 Dec 2023 Alon Albalak, Liangming Pan, Colin Raffel, William Yang Wang

The data used to pretrain large language models has a decisive impact on a model's downstream performance, which has led to a large body of work on data selection methods that aim to automatically determine the most suitable data to use for pretraining.

Language Modelling

Logic-LM: Empowering Large Language Models with Symbolic Solvers for Faithful Logical Reasoning

1 code implementation20 May 2023 Liangming Pan, Alon Albalak, Xinyi Wang, William Yang Wang

We also introduce a self-refinement module, which utilizes the symbolic solver's error messages to revise symbolic formalizations.

Logical Reasoning

Improving Few-Shot Generalization by Exploring and Exploiting Auxiliary Data

1 code implementation NeurIPS 2023 Alon Albalak, Colin Raffel, William Yang Wang

In this work, we focus on Few-shot Learning with Auxiliary Data (FLAD), a training paradigm that assumes access to auxiliary data during few-shot learning in hopes of improving generalization.

Few-Shot Learning

CausalDialogue: Modeling Utterance-level Causality in Conversations

1 code implementation20 Dec 2022 Yi-Lin Tuan, Alon Albalak, Wenda Xu, Michael Saxon, Connor Pryor, Lise Getoor, William Yang Wang

Despite their widespread adoption, neural conversation models have yet to exhibit natural chat capabilities with humans.

Dialogue Generation

An Exploration of Data Efficiency in Intra-Dataset Task Transfer for Dialog Understanding

no code implementations21 Oct 2022 Josiah Ross, Luke Yoffe, Alon Albalak, William Yang Wang

Transfer learning is an exciting area of Natural Language Processing that has the potential to both improve model performance and increase data efficiency.

Transfer Learning

Data-Efficiency with a Single GPU: An Exploration of Transfer Methods for Small Language Models

no code implementations8 Oct 2022 Alon Albalak, Akshat Shrivastava, Chinnadhurai Sankar, Adithya Sagar, Mike Ross

Multi-task learning (MTL), instruction tuning, and prompting have recently been shown to improve the generalizability of large language models to new tasks.

Multi-Task Learning

Emotion Recognition in Conversation using Probabilistic Soft Logic

no code implementations14 Jul 2022 Eriq Augustine, Pegah Jandaghi, Alon Albalak, Connor Pryor, Charles Dickens, William Wang, Lise Getoor

Creating agents that can both appropriately respond to conversations and understand complex human linguistic tendencies and social cues has been a long standing challenge in the NLP community.

Emotion Recognition in Conversation Logical Reasoning +2

NeuPSL: Neural Probabilistic Soft Logic

no code implementations27 May 2022 Connor Pryor, Charles Dickens, Eriq Augustine, Alon Albalak, William Wang, Lise Getoor

In this paper, we introduce Neural Probabilistic Soft Logic (NeuPSL), a novel neuro-symbolic (NeSy) framework that unites state-of-the-art symbolic reasoning with the low-level perception of deep neural networks.

FETA: A Benchmark for Few-Sample Task Transfer in Open-Domain Dialogue

1 code implementation12 May 2022 Alon Albalak, Yi-Lin Tuan, Pegah Jandaghi, Connor Pryor, Luke Yoffe, Deepak Ramachandran, Lise Getoor, Jay Pujara, William Yang Wang

Task transfer, transferring knowledge contained in related tasks, holds the promise of reducing the quantity of labeled data required to fine-tune language models.

Dialogue Understanding Domain Adaptation +1

Addressing Issues of Cross-Linguality in Open-Retrieval Question Answering Systems For Emergent Domains

1 code implementation26 Jan 2022 Alon Albalak, Sharon Levy, William Yang Wang

Open-retrieval question answering systems are generally trained and tested on large datasets in well-established domains.

Question Answering Retrieval +1

D-REX: Dialogue Relation Extraction with Explanations

1 code implementation NLP4ConvAI (ACL) 2022 Alon Albalak, Varun Embar, Yi-Lin Tuan, Lise Getoor, William Yang Wang

Existing research studies on cross-sentence relation extraction in long-form multi-party conversations aim to improve relation extraction without considering the explainability of such methods.

Dialog Relation Extraction Relation +3

Modeling Disclosive Transparency in NLP Application Descriptions

1 code implementation EMNLP 2021 Michael Saxon, Sharon Levy, Xinyi Wang, Alon Albalak, William Yang Wang

Broader disclosive transparency$-$truth and clarity in communication regarding the function of AI systems$-$is widely considered desirable.

Fairness Language Modelling +1

Cannot find the paper you are looking for? You can Submit a new open access paper.