Search Results for author: Peixin Cao

Found 1 papers, 1 papers with code

Adversarial Preference Optimization

1 code implementation14 Nov 2023 Pengyu Cheng, Yifan Yang, Jian Li, Yong Dai, Tianhao Hu, Peixin Cao, Nan Du

Human preference alignment is essential to improve the interaction quality of large language models (LLMs).

Cannot find the paper you are looking for? You can Submit a new open access paper.