Search Results for author: Jiuhai Chen

Found 17 papers, 7 papers with code

Automated Data Curation for Robust Language Model Fine-Tuning

no code implementations19 Mar 2024 Jiuhai Chen, Jonas Mueller

We introduce an automated data curation pipeline CLEAR (Confidence-based LLM Evaluation And Rectification) for instruction tuning datasets, that can be used with any LLM and fine-tuning procedure.

Language Modelling Text Generation

Can LLMs Speak For Diverse People? Tuning LLMs via Debate to Generate Controllable Controversial Statements

1 code implementation16 Feb 2024 Ming Li, Jiuhai Chen, Lichang Chen, Tianyi Zhou

To examine DEBATunE, we curate the largest dataset of debate topics so far, which covers 710 controversial topics and corresponding arguments for each topic.

Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning

2 code implementations15 Feb 2024 Ming Li, Lichang Chen, Jiuhai Chen, Shwai He, Jiuxiang Gu, Tianyi Zhou

Instruction tuning is critical to large language models (LLMs) for achieving better instruction following and task adaptation capabilities but its success heavily relies on the training data quality.

Data Augmentation Instruction Following

ODIN: Disentangled Reward Mitigates Hacking in RLHF

no code implementations11 Feb 2024 Lichang Chen, Chen Zhu, Davit Soselia, Jiuhai Chen, Tianyi Zhou, Tom Goldstein, Heng Huang, Mohammad Shoeybi, Bryan Catanzaro

In this work, we study the issue of reward hacking on the response length, a challenge emerging in Reinforcement Learning from Human Feedback (RLHF) on LLMs.

Reflection-Tuning: Data Recycling Improves LLM Instruction-Tuning

2 code implementations18 Oct 2023 Ming Li, Lichang Chen, Jiuhai Chen, Shwai He, Heng Huang, Jiuxiang Gu, Tianyi Zhou

Recent advancements in Large Language Models (LLMs) have expanded the horizons of natural language understanding and generation.

Natural Language Understanding

Quantifying Uncertainty in Answers from any Language Model and Enhancing their Trustworthiness

no code implementations30 Aug 2023 Jiuhai Chen, Jonas Mueller

We introduce BSDetector, a method for detecting bad and speculative answers from a pretrained Large Language Model by estimating a numeric confidence score for any output it generated.

Language Modelling Large Language Model +1

InstructZero: Efficient Instruction Optimization for Black-Box Large Language Models

1 code implementation5 Jun 2023 Lichang Chen, Jiuhai Chen, Tom Goldstein, Heng Huang, Tianyi Zhou

Large language models~(LLMs) are instruction followers, but it can be challenging to find the best instruction for different situations, especially for black-box LLMs on which backpropagation is forbidden.

Bayesian Optimization

When do you need Chain-of-Thought Prompting for ChatGPT?

no code implementations6 Apr 2023 Jiuhai Chen, Lichang Chen, Heng Huang, Tianyi Zhou

However, it is not clear whether CoT is still effective on more recent instruction finetuned (IFT) LLMs such as ChatGPT.

Arithmetic Reasoning Memorization

How Many Demonstrations Do You Need for In-context Learning?

no code implementations14 Mar 2023 Jiuhai Chen, Lichang Chen, Chen Zhu, Tianyi Zhou

Moreover, ICL (with and w/o CoT) using only one correct demo significantly outperforms all-demo ICL adopted by most previous works, indicating the weakness of LLMs in finding correct demo(s) for input queries, which is difficult to evaluate on the biased datasets.

In-Context Learning

A Robust Stacking Framework for Training Deep Graph Models with Multifaceted Node Features

no code implementations16 Jun 2022 Jiuhai Chen, Jonas Mueller, Vassilis N. Ioannidis, Tom Goldstein, David Wipf

Graph Neural Networks (GNNs) with numerical node features and graph structure as inputs have demonstrated superior performance on various supervised learning tasks with graph data.

Does your graph need a confidence boost? Convergent boosted smoothing on graphs with tabular node features

1 code implementation26 Oct 2021 Jiuhai Chen, Jonas Mueller, Vassilis N. Ioannidis, Soji Adeshina, Yangkun Wang, Tom Goldstein, David Wipf

For supervised learning with tabular data, decision tree ensembles produced via boosting techniques generally dominate real-world applications involving iid training/test sets.

Why Propagate Alone? Parallel Use of Labels and Features on Graphs

no code implementations ICLR 2022 Yangkun Wang, Jiarui Jin, Weinan Zhang, Yongyi Yang, Jiuhai Chen, Quan Gan, Yong Yu, Zheng Zhang, Zengfeng Huang, David Wipf

In this regard, it has recently been proposed to use a randomly-selected portion of the training labels as GNN inputs, concatenated with the original node features for making predictions on the remaining labels.

Node Property Prediction Property Prediction

A Closer Look at Distribution Shifts and Out-of-Distribution Generalization on Graphs

no code implementations29 Sep 2021 Mucong Ding, Kezhi Kong, Jiuhai Chen, John Kirchenbauer, Micah Goldblum, David Wipf, Furong Huang, Tom Goldstein

We observe that in most cases, we need both a suitable domain generalization algorithm and a strong GNN backbone model to optimize out-of-distribution test performance.

Domain Generalization Graph Classification +1

Convergent Boosted Smoothing for Modeling GraphData with Tabular Node Features

no code implementations ICLR 2022 Jiuhai Chen, Jonas Mueller, Vassilis N. Ioannidis, Soji Adeshina, Yangkun Wang, Tom Goldstein, David Wipf

Many practical modeling tasks require making predictions using tabular data composed of heterogeneous feature types (e. g., text-based, categorical, continuous, etc.).

Particle-based Energetic Variational Inference

1 code implementation14 Apr 2020 Yiwei Wang, Jiuhai Chen, Chun Liu, Lulu Kang

Using the EVI framework, we can derive many existing Particle-based Variational Inference (ParVI) methods, including the popular Stein Variational Gradient Descent (SVGD) approach.

Variational Inference

Cannot find the paper you are looking for? You can Submit a new open access paper.