Search Results for author: Alexander Bukharin

Found 10 papers, 4 papers with code

HelpSteer2-Preference: Complementing Ratings with Preferences

no code implementations2 Oct 2024 Zhilin Wang, Alexander Bukharin, Olivier Delalleau, Daniel Egert, Gerald Shen, Jiaqi Zeng, Oleksii Kuchaiev, Yi Dong

Reward models are critical for aligning models to follow instructions, and are typically trained following one of two popular paradigms: Bradley-Terry style or Regression style.

regression

RNR: Teaching Large Language Models to Follow Roles and Rules

no code implementations10 Sep 2024 Kuan Wang, Alexander Bukharin, Haoming Jiang, Qingyu Yin, Zhengyang Wang, Tuo Zhao, Jingbo Shang, Chao Zhang, Bing Yin, Xian Li, Jianshu Chen, Shiyang Li

However, existing models trained on open-source IFT datasets only have the ability to follow instructions from users, and often fail to follow complex role and rules specified by developers, a. k. a.

Instruction Following

Robust Reinforcement Learning from Corrupted Human Feedback

no code implementations21 Jun 2024 Alexander Bukharin, Ilgee Hong, Haoming Jiang, Zichong Li, Qingru Zhang, Zixuan Zhang, Tuo Zhao

To tackle this challenge, we propose a robust RLHF approach -- $R^3M$, which models the potentially corrupted preference label as sparse outliers.

reinforcement-learning Reinforcement Learning +1

Adaptive Preference Scaling for Reinforcement Learning with Human Feedback

no code implementations4 Jun 2024 Ilgee Hong, Zichong Li, Alexander Bukharin, Yixiao Li, Haoming Jiang, Tianbao Yang, Tuo Zhao

By incorporating an adaptive scaling parameter into the loss for each pair, our method increases the flexibility of the reward function.

reinforcement-learning Reinforcement Learning +1

Data Diversity Matters for Robust Instruction Tuning

no code implementations21 Nov 2023 Alexander Bukharin, Shiyang Li, Zhengyang Wang, Jingfeng Yang, Bing Yin, Xian Li, Chao Zhang, Tuo Zhao, Haoming Jiang

QDIT provides a simple method to simultaneously control dataset diversity and quality, allowing us to conduct an in-depth study on the effect of diversity and quality on instruction tuning performance.

Diversity Instruction Following

Deep Reinforcement Learning from Hierarchical Preference Design

1 code implementation6 Sep 2023 Alexander Bukharin, Yixiao Li, Pengcheng He, Tuo Zhao

Researchers typically utilize feedback signals from the environment to handcraft a reward function, but this process is not always effective due to the varying scale and intricate dependencies of the feedback signals.

Deep Reinforcement Learning reinforcement-learning +1

Machine Learning Force Fields with Data Cost Aware Training

1 code implementation5 Jun 2023 Alexander Bukharin, Tianyi Liu, Shengjie Wang, Simiao Zuo, Weihao Gao, Wen Yan, Tuo Zhao

To address this issue, we propose a multi-stage computational framework -- ASTEROID, which lowers the data cost of MLFFs by leveraging a combination of cheap inaccurate data and expensive accurate data.

AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning

2 code implementations18 Mar 2023 Qingru Zhang, Minshuo Chen, Alexander Bukharin, Nikos Karampatziakis, Pengcheng He, Yu Cheng, Weizhu Chen, Tuo Zhao

Therefore, many fine-tuning methods are proposed to learn incremental updates of pre-trained weights in a parameter efficient way, e. g., low-rank increments.

parameter-efficient fine-tuning Question Answering +1

Early Detection of COVID-19 Hotspots Using Spatio-Temporal Data

no code implementations31 May 2021 Shixiang Zhu, Alexander Bukharin, Liyan Xie, Khurram Yamin, Shihao Yang, Pinar Keskinocak, Yao Xie

Recently, the Centers for Disease Control and Prevention (CDC) has worked with other federal agencies to identify counties with increasing coronavirus disease 2019 (COVID-19) incidence (hotspots) and offers support to local health departments to limit the spread of the disease.

Cannot find the paper you are looking for? You can Submit a new open access paper.