1 code implementation • 20 Feb 2024 • Che Zhang, Zhenyang Xiao, Chengcheng Han, Yixin Lian, Yuejian Fang
After integrating the original CoT data and checking-correction data for training, we observe that models could improve their self-checking capabilities, thereby enhancing their self-correction capacity and eliminating the need for external feedback or ground truth labels to ascertain the endpoint of correction.
1 code implementation • 8 Oct 2023 • Chengcheng Han, Xiaowei Du, Che Zhang, Yixin Lian, Xiang Li, Ming Gao, Baoyuan Wang
Chain-of-Thought (CoT) prompting has proven to be effective in enhancing the reasoning capabilities of Large Language Models (LLMs) with at least 100 billion parameters.
1 code implementation • 11 Jul 2023 • Che Zhang, Ping'an Liu, Zhenyang Xiao, Haojun Fei
The study of human values is essential in both practical and theoretical domains.