1 code implementation • 30 Apr 2025 • Haotian Luo, Haiying He, Yibo Wang, Jinluan Yang, Rui Liu, Naiqiang Tan, Xiaochun Cao, DaCheng Tao, Li Shen
To address this, we propose a novel two-stage framework for adaptive and efficient reasoning.
1 code implementation • 30 Jan 2025 • Yibo Wang, Tiansheng Huang, Li Shen, Huanjin Yao, Haotian Luo, Rui Liu, Naiqiang Tan, Jiaxing Huang, DaCheng Tao
Mainstream defenses aim to vaccinate the model such that the later harmful fine-tuning attack is less effective.
1 code implementation • 22 Jan 2025 • Haotian Luo, Li Shen, Haiying He, Yibo Wang, Shiwei Liu, Wei Li, Naiqiang Tan, Xiaochun Cao, DaCheng Tao
Experiments on various mathematical reasoning benchmarks show that O1-Pruner not only significantly reduces inference overhead but also achieves higher accuracy, providing a novel and promising solution to this challenge.
no code implementations • 9 Apr 2024 • Haotian Luo
Consequently, we utilize the distance between sample pairs in the latent space to guide DPO-based alignment training.
1 code implementation • 8 Apr 2024 • Shibo Hao, Yi Gu, Haotian Luo, Tianyang Liu, Xiyan Shao, Xinyuan Wang, Shuhua Xie, Haodi Ma, Adithya Samavedhi, Qiyue Gao, Zhen Wang, Zhiting Hu
(2) We develop LLM Reasoners, a library for standardized modular implementation of existing and new reasoning algorithms, under a unified formulation of the search, reward, and world model components.
no code implementations • 25 Nov 2023 • Haotian Luo, Yixin Liu, Peidong Liu, Xianggen Liu
Therefore, we present vector-quantized prompts as the cues to control the generation of pre-trained models.
no code implementations • 3 Nov 2023 • Haotian Luo, Kunming Wu, Cheng Dai, Sixian Ding, Xinhao Chen
RNN-like language models are getting renewed attention from NLP researchers in recent years and several models have made significant progress, which demonstrates performance comparable to traditional transformers.
2 code implementations • 25 Oct 2023 • Xinyuan Wang, Chenxi Li, Zhen Wang, Fan Bai, Haotian Luo, Jiayou Zhang, Nebojsa Jojic, Eric P. Xing, Zhiting Hu
Highly effective, task-specific prompts are often heavily engineered by experts to integrate detailed instructions and domain insights based on a deep understanding of both instincts of large language models (LLMs) and the intricacies of the target task.