Search Results for author: Yuxuan Yao

Found 14 papers, 6 papers with code

Activation-Guided Consensus Merging for Large Language Models

no code implementations20 May 2025 Yuxuan Yao, Shuqi Liu, Zehua Liu, Qintong Li, Mingyang Liu, Xiongwei Han, Zhijiang Guo, Han Wu, Linqi Song

While existing training-based and prompt-based approaches face significant challenges in terms of efficiency and stability, model merging emerges as a promising strategy to integrate the diverse capabilities of different Large Language Models (LLMs) into a unified model.

Unlocking Efficient Long-to-Short LLM Reasoning with Model Merging

1 code implementation26 Mar 2025 Han Wu, Yuxuan Yao, Shuqi Liu, Zehua Liu, Xiaojin Fu, Xiongwei Han, Xing Li, Hui-Ling Zhen, Tao Zhong, Mingxuan Yuan

Model merging, on the other hand, offers a cost-effective and robust alternative by integrating the quick-thinking capabilities of System 1 models with the methodical reasoning of System 2 models.

Prompt Engineering Reinforcement Learning (RL)

From System 1 to System 2: A Survey of Reasoning Large Language Models

1 code implementation24 Feb 2025 Zhong-Zhi Li, Duzhen Zhang, Ming-Liang Zhang, Jiaxin Zhang, Zengyan Liu, Yuxuan Yao, Haotian Xu, Junhao Zheng, Pei-Jie Wang, Xiuyi Chen, Yingying Zhang, Fei Yin, Jiahua Dong, Zhijiang Guo, Le Song, Cheng-Lin Liu

Achieving human-level intelligence requires refining the transition from the fast, intuitive System 1 to the slower, more deliberate System 2 reasoning.

Logical Reasoning

LoRE-Merging: Exploring Low-Rank Estimation For Large Language Model Merging

no code implementations15 Feb 2025 Zehua Liu, Han Wu, Yuxuan Yao, Ruifeng She, Xiongwei Han, Tao Zhong, Mingxuan Yuan

While most current approaches rely on further training techniques, such as fine-tuning or reinforcement learning, to enhance model capacities, model merging stands out for its ability of improving models without requiring any additional training.

Language Modeling Language Modelling +1

Reflective Gaussian Splatting

no code implementations26 Dec 2024 Yuxuan Yao, Zixuan Zeng, Chun Gu, Xiatian Zhu, Li Zhang

Novel view synthesis has experienced significant advancements owing to increasingly capable NeRF- and 3DGS-based methods.

3DGS NeRF +2

IRGS: Inter-Reflective Gaussian Splatting with 2D Gaussian Ray Tracing

no code implementations CVPR 2025 Chun Gu, Xiaofei Wei, Zixuan Zeng, Yuxuan Yao, Li Zhang

In inverse rendering, accurately modeling visibility and indirect radiance for incident light is essential for capturing secondary effects.

3DGS Inverse Rendering

Determine-Then-Ensemble: Necessity of Top-k Union for Large Language Model Ensembling

no code implementations3 Oct 2024 Yuxuan Yao, Han Wu, Mingyang Liu, Sichun Luo, Xiongwei Han, Jie Liu, Zhijiang Guo, Linqi Song

Large language models (LLMs) exhibit varying strengths and weaknesses across different tasks, prompting recent studies to explore the benefits of ensembling models to leverage their complementary advantages.

Language Modeling Language Modelling +2

MR-Ben: A Meta-Reasoning Benchmark for Evaluating System-2 Thinking in LLMs

no code implementations20 Jun 2024 Zhongshen Zeng, Yinhong Liu, Yingjia Wan, Jingyao Li, Pengguang Chen, Jianbo Dai, Yuxuan Yao, Rongwu Xu, Zehan Qi, Wanru Zhao, Linling Shen, Jianqiao Lu, Haochen Tan, Yukang Chen, Hao Zhang, Zhan Shi, Bailin Wang, Zhijiang Guo, Jiaya Jia

Large language models (LLMs) have shown increasing capability in problem-solving and decision-making, largely based on the step-by-step chain-of-thought reasoning processes.

Decision Making

Privacy in LLM-based Recommendation: Recent Advances and Future Directions

no code implementations3 Jun 2024 Sichun Luo, Wei Shao, Yuxuan Yao, Jian Xu, Mingyang Liu, Qintong Li, Bowei He, Maolin Wang, Guanzhi Deng, Hanxu Hou, Xinyi Zhang, Linqi Song

Nowadays, large language models (LLMs) have been integrated with conventional recommendation models to improve recommendation performance.

Learning From Correctness Without Prompting Makes LLM Efficient Reasoner

1 code implementation28 Mar 2024 Yuxuan Yao, Han Wu, Zhijiang Guo, Biyan Zhou, Jiahui Gao, Sichun Luo, Hanxu Hou, Xiaojin Fu, Linqi Song

Large language models (LLMs) have demonstrated outstanding performance across various tasks, yet they still exhibit limitations such as hallucination, unfaithful reasoning, and toxic content.

Hallucination

Can LLM Substitute Human Labeling? A Case Study of Fine-grained Chinese Address Entity Recognition Dataset for UAV Delivery

1 code implementation10 Mar 2024 Yuxuan Yao, Sichun Luo, Haohan Zhao, Guanzhi Deng, Linqi Song

We present CNER-UAV, a fine-grained \textbf{C}hinese \textbf{N}ame \textbf{E}ntity \textbf{R}ecognition dataset specifically designed for the task of address resolution in \textbf{U}nmanned \textbf{A}erial \textbf{V}ehicle delivery systems.

NER

Integrating Large Language Models into Recommendation via Mutual Augmentation and Adaptive Aggregation

no code implementations25 Jan 2024 Sichun Luo, Yuxuan Yao, Bowei He, Yinya Huang, Aojun Zhou, Xinyi Zhang, Yuanzhang Xiao, Mingjie Zhan, Linqi Song

Conventional recommendation methods have achieved notable advancements by harnessing collaborative or sequential information from user behavior.

Data Augmentation

RecRanker: Instruction Tuning Large Language Model as Ranker for Top-k Recommendation

1 code implementation26 Dec 2023 Sichun Luo, Bowei He, Haohan Zhao, Wei Shao, Yanlin Qi, Yinya Huang, Aojun Zhou, Yuxuan Yao, Zongpeng Li, Yuanzhang Xiao, Mingjie Zhan, Linqi Song

Large Language Models (LLMs) have demonstrated remarkable capabilities and have been extensively deployed across various domains, including recommender systems.

In-Context Learning Language Modeling +4

Fine-grained Conversational Decoding via Isotropic and Proximal Search

1 code implementation12 Oct 2023 Yuxuan Yao, Han Wu, Qiling Xu, Linqi Song

General-purpose text decoding approaches are usually adopted for dialogue response generation.

Informativeness Response Generation

Cannot find the paper you are looking for? You can Submit a new open access paper.