1 code implementation • 16 Oct 2024 • Yaxi Lu, Shenzhi Yang, Cheng Qian, Guirong Chen, Qinyu Luo, Yesai Wu, Huadong Wang, Xin Cong, Zhong Zhang, Yankai Lin, Weiwen Liu, Yasheng Wang, Zhiyuan Liu, Fangming Liu, Maosong Sun
The labeled data is used to train a reward model that simulates human judgment and serves as an automatic evaluator of the proactiveness of LLM agents.
1 code implementation • 6 Jun 2024 • Xinhao Yao, Xiaolin Hu, Shenzhi Yang, Yong liu
Pre-trained large language models (LLMs) based on Transformer have demonstrated striking in-context learning (ICL) abilities.