Search Results for author: Justin Chih-Yao Chen

Found 9 papers, 6 papers with code

MAMM-Refine: A Recipe for Improving Faithfulness in Generation with Multi-Agent Collaboration

no code implementations19 Mar 2025 David Wan, Justin Chih-Yao Chen, Elias Stengel-Eskin, Mohit Bansal

We investigate how iterative collaboration among multiple instances and types of large language models (LLMs) enhances subtasks in the refinement process, such as error detection, critiquing unfaithful sentences, and making corrections based on critiques.

Long Form Question Answering Reranking

Symbolic Mixture-of-Experts: Adaptive Skill-based Routing for Heterogeneous Reasoning

no code implementations7 Mar 2025 Justin Chih-Yao Chen, Sukwon Yun, Elias Stengel-Eskin, Tianlong Chen, Mohit Bansal

We propose a skill-based recruiting strategy that dynamically selects the most relevant set of expert LLMs for diverse reasoning tasks based on their strengths.

Math Mixture-of-Experts +1

Learning to Generate Unit Tests for Automated Debugging

1 code implementation3 Feb 2025 Archiki Prasad, Elias Stengel-Eskin, Justin Chih-Yao Chen, Zaid Khan, Mohit Bansal

However, we uncover a trade-off between generating unit test inputs that reveal errors when given a faulty code and correctly predicting the unit test output without access to the gold solution.

HumanEval Large Language Model +1

MAgICoRe: Multi-Agent, Iterative, Coarse-to-Fine Refinement for Reasoning

1 code implementation18 Sep 2024 Justin Chih-Yao Chen, Archiki Prasad, Swarnadeep Saha, Elias Stengel-Eskin, Mohit Bansal

Moreover, to ensure effective refinement, we employ a multi-agent loop with three agents: Solver, Reviewer (which generates targeted feedback based on step-wise RM scores), and the Refiner (which incorporates feedback).

Math

System-1.x: Learning to Balance Fast and Slow Planning with Language Models

1 code implementation19 Jul 2024 Swarnadeep Saha, Archiki Prasad, Justin Chih-Yao Chen, Peter Hase, Elias Stengel-Eskin, Mohit Bansal

To this end, we propose the System-1. x Planner, a controllable planning framework with LLMs that is capable of generating hybrid plans and balancing between the two planning modes based on the difficulty of the problem at hand.

MAGDi: Structured Distillation of Multi-Agent Interaction Graphs Improves Reasoning in Smaller Language Models

1 code implementation2 Feb 2024 Justin Chih-Yao Chen, Swarnadeep Saha, Elias Stengel-Eskin, Mohit Bansal

Experiments on seven widely used commonsense and math reasoning benchmarks show that MAGDi improves the reasoning capabilities of smaller models, outperforming several methods that distill from a single teacher and multiple teachers.

Language Modelling Large Language Model +1

Location-Aware Visual Question Generation with Lightweight Models

1 code implementation23 Oct 2023 Nicholas Collin Suwono, Justin Chih-Yao Chen, Tun Min Hung, Ting-Hao Kenneth Huang, I-Bin Liao, Yung-Hui Li, Lun-Wei Ku, Shao-Hua Sun

This work introduces a novel task, location-aware visual question generation (LocaVQG), which aims to generate engaging questions from data relevant to a particular geographical location.

Dataset Generation Question Generation +1

ReConcile: Round-Table Conference Improves Reasoning via Consensus among Diverse LLMs

1 code implementation22 Sep 2023 Justin Chih-Yao Chen, Swarnadeep Saha, Mohit Bansal

In each round, ReConcile initiates discussion between agents via a 'discussion prompt' that consists of (a) grouped answers and explanations generated by each agent in the previous round, (b) their confidence scores, and (c) demonstrations of answer-rectifying human explanations, used for convincing other agents.

Math

Cannot find the paper you are looking for? You can Submit a new open access paper.