Search Results for author: Archiki Prasad

Found 15 papers, 12 papers with code

Learning to Generate Unit Tests for Automated Debugging

1 code implementation3 Feb 2025 Archiki Prasad, Elias Stengel-Eskin, Justin Chih-Yao Chen, Zaid Khan, Mohit Bansal

However, we uncover a trade-off between generating unit test inputs that reveal errors when given a faulty code and correctly predicting the unit test output without access to the gold solution.

HumanEval Large Language Model

Self-Consistency Preference Optimization

no code implementations6 Nov 2024 Archiki Prasad, Weizhe Yuan, Richard Yuanzhe Pang, Jing Xu, Maryam Fazel-Zarandi, Mohit Bansal, Sainbayar Sukhbaatar, Jason Weston, Jane Yu

Self-alignment, whereby models learn to improve themselves without human annotation, is a rapidly growing research area.

GSM8K Math

LASeR: Learning to Adaptively Select Reward Models with Multi-Armed Bandits

1 code implementation2 Oct 2024 Duy Nguyen, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal

Our results on commonsense and math reasoning tasks demonstrate that LASeR can boost iterative LLM optimization by optimizing for multiple RMs, improving the absolute average accuracy of Llama-3-8B over three datasets by 2. 67% over training with ensemble RM scores while also showing superior training efficiency (e. g., a 2x speedup).

Instruction Following Math +1

MAgICoRe: Multi-Agent, Iterative, Coarse-to-Fine Refinement for Reasoning

1 code implementation18 Sep 2024 Justin Chih-Yao Chen, Archiki Prasad, Swarnadeep Saha, Elias Stengel-Eskin, Mohit Bansal

Moreover, to ensure effective refinement, we employ a multi-agent loop with three agents: Solver, Reviewer (which generates targeted feedback based on step-wise RM scores), and the Refiner (which incorporates feedback).

Math

AdaCAD: Adaptively Decoding to Balance Conflicts between Contextual and Parametric Knowledge

1 code implementation11 Sep 2024 Han Wang, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal

Knowledge conflict arises from discrepancies between information in the context of a large language model (LLM) and the knowledge stored in its parameters.

Language Modelling Large Language Model +1

System-1.x: Learning to Balance Fast and Slow Planning with Language Models

1 code implementation19 Jul 2024 Swarnadeep Saha, Archiki Prasad, Justin Chih-Yao Chen, Peter Hase, Elias Stengel-Eskin, Mohit Bansal

To this end, we propose the System-1. x Planner, a controllable planning framework with LLMs that is capable of generating hybrid plans and balancing between the two planning modes based on the difficulty of the problem at hand.

Soft Self-Consistency Improves Language Model Agents

1 code implementation20 Feb 2024 Han Wang, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal

Current "sample and select" methods such as self-consistency (SC) rely on majority voting to score answers.

Language Modeling Language Modelling +2

ReGAL: Refactoring Programs to Discover Generalizable Abstractions

1 code implementation29 Jan 2024 Elias Stengel-Eskin, Archiki Prasad, Mohit Bansal

While large language models (LLMs) are increasingly being used for program synthesis, they lack the global view needed to develop useful abstractions; they generally predict programs one at a time, often repeating the same functionality.

Date Understanding Math +2

ADaPT: As-Needed Decomposition and Planning with Language Models

1 code implementation8 Nov 2023 Archiki Prasad, Alexander Koller, Mareike Hartmann, Peter Clark, Ashish Sabharwal, Mohit Bansal, Tushar Khot

Large Language Models (LLMs) are increasingly being used for interactive decision-making tasks requiring planning and adapting to the environment.

Decision Making

Rephrase, Augment, Reason: Visual Grounding of Questions for Vision-Language Models

1 code implementation9 Oct 2023 Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal

An increasing number of vision-language tasks can be handled with little to no training, i. e., in a zero and few-shot manner, by marrying large language models (LLMs) to vision encoders, resulting in large vision-language models (LVLMs).

Language Modelling Question Answering +2

ReCEval: Evaluating Reasoning Chains via Correctness and Informativeness

1 code implementation21 Apr 2023 Archiki Prasad, Swarnadeep Saha, Xiang Zhou, Mohit Bansal

Multi-step reasoning ability is fundamental to many natural language tasks, yet it is unclear what constitutes a good reasoning chain and how to evaluate them.

Informativeness Natural Language Inference +1

GrIPS: Gradient-free, Edit-based Instruction Search for Prompting Large Language Models

2 code implementations14 Mar 2022 Archiki Prasad, Peter Hase, Xiang Zhou, Mohit Bansal

Providing natural language instructions in prompts is a useful new paradigm for improving task performance of large language models in a zero-shot setting.

The Effectiveness of Intermediate-Task Training for Code-Switched Natural Language Understanding

no code implementations EMNLP (MRL) 2021 Archiki Prasad, Mohammad Ali Rehan, Shreya Pathak, Preethi Jyothi

In this work, we propose the use of bilingual intermediate pretraining as a reliable technique to derive large and consistent performance gains on three different NLP tasks using code-switched text.

Language Modelling Natural Language Inference +4

How Accents Confound: Probing for Accent Information in End-to-End Speech Recognition Systems

no code implementations ACL 2020 Archiki Prasad, Preethi Jyothi

We use a state-of-the-art end-to-end ASR system, comprising convolutional and recurrent layers, that is trained on a large amount of US-accented English speech and evaluate the model on speech samples from seven different English accents.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

Cannot find the paper you are looking for? You can Submit a new open access paper.