no code implementations • 28 May 2025 • Siqi Fan, Peng Han, Shuo Shang, Yequan Wang, Aixin Sun
With reference to the instruct model, we formally define reasoning efficiency and observe a potential reasoning efficiency scaling law in LLMs.
no code implementations • 30 Mar 2025 • Siqi Fan, Xiusheng Huang, Yiqun Yao, Xuezhi Fang, Kang Liu, Peng Han, Shuo Shang, Aixin Sun, Yequan Wang
However, during multi-turn, multi-agent interactions, LLMs begin to exhibit consistent, character-like behaviors, hinting at a form of emergent lifelong learning.
no code implementations • 11 Mar 2025 • Siqi Fan, Xuezhi Fang, Xingrun Xing, Peng Han, Shuo Shang, Yequan Wang
Experiments on large language models (\ie the Llama) with $7 \sim 70$ billion parameters show that $D^3$ can achieve an average 1. 5x speedup compared with the full-inference pipeline while maintaining comparable performance with nearly no performance drop ($<1\%$) on the GSM8K and BBH benchmarks.
1 code implementation • 22 Nov 2024 • Silin Zhou, Shuo Shang, Lisi Chen, Peng Han, Christian S. Jensen
Trajectory representation learning (TRL) maps trajectories to vectors that can be used for many downstream tasks.
1 code implementation • 24 Oct 2024 • Duc Kieu, Tung Kieu, Peng Han, Bin Yang, Christian S. Jensen, Bac Le
However, existing methods assume the input is fixed-topology road networks and static traffic time series.
no code implementations • 5 Sep 2024 • Xin Jiang, Xiang Li, Wenjia Ma, Xuezhi Fang, Yiqun Yao, Naitong Yu, Xuying Meng, Peng Han, Jing Li, Aixin Sun, Yequan Wang
Sketch comprises the following components: (1) a suite of task description schemas and prompt templates encompassing various NLP tasks; (2) a user-friendly, interactive process for building structured output LLM services tailored to various NLP tasks; (3) an open-source dataset for output format control, along with tools for dataset construction; and (4) an open-source model based on LLaMA3-8B-Instruct that adeptly comprehends and adheres to output formatting instructions.
1 code implementation • 8 Aug 2024 • Yiqun Yao, Wenjia Ma, Xuezhi Fang, Xin Jiang, Xiang Li, Xuying Meng, Peng Han, Jing Li, Aixin Sun, Yequan Wang
Controlling the format of outputs generated by large language models (LLMs) is a critical functionality in various applications.
no code implementations • 8 Jul 2024 • Chang Gong, Di Yao, Jin Wang, Wenbin Li, Lanting Fang, Yongtao Xie, Kaiyu Feng, Peng Han, Jingping Bi
In this paper, we unveil the issues of unobserved confounders and heterogeneity in partial observation and come up with a new problem of root cause analysis with partially observed data.
no code implementations • 3 Jul 2024 • Chengrui Huang, Zhengliang Shi, Yuntao Wen, Xiuying Chen, Peng Han, Shen Gao, Shuo Shang
Tool learning methods have enhanced the ability of large language models (LLMs) to interact with real-world applications.
1 code implementation • 18 May 2024 • Xinli Guo, Weidong Zhang, Ruonan Liu, Peng Han, Hongtian Chen
A novel 3DGS-based SLAM approach with a fusion of deep visual feature, dual keyframe selection and 3DGS is presented in this paper.
no code implementations • 9 Apr 2024 • Shen Gao, Yifan Wang, Jiabao Fang, Lisi Chen, Peng Han, Shuo Shang
Recommendation systems play a crucial role in various domains, suggesting items based on user behavior. However, the lack of transparency in presenting recommendations can lead to user confusion.
no code implementations • 4 Mar 2024 • Siqi Fan, Xin Jiang, Xiang Li, Xuying Meng, Peng Han, Shuo Shang, Aixin Sun, Yequan Wang, Zhongyuan Wang
That is, not all layers of LLMs are necessary during inference.
no code implementations • 7 Feb 2024 • Yuanfang Zhang, Junxuan Li, Kaiqing Luo, Yiying Yang, Jiayi Han, Nian Liu, Denghui Qin, Peng Han, Chengpei Xu
Extensive experiments demonstrate that by leveraging V2V communication, the SSC performance can be increased by 8. 3% on geometric metric IoU and 6. 0% mIOU.
no code implementations • 27 Dec 2023 • Minbo Ma, Jilin Hu, Christian S. Jensen, Fei Teng, Peng Han, Zhiqiang Xu, Tianrui Li
Spatio-temporal forecasting of future values of spatially correlated time series is important across many cyber-physical systems (CPS).
no code implementations • 7 Sep 2023 • Xiang Li, Yiqun Yao, Xin Jiang, Xuezhi Fang, Xuying Meng, Siqi Fan, Peng Han, Jing Li, Li Du, Bowen Qin, Zheng Zhang, Aixin Sun, Yequan Wang
Large language models (LLMs) are considered important approaches towards foundational machine intelligence, achieving remarkable success in Natural Language Processing and multimodal tasks, among others.
1 code implementation • 14 Apr 2023 • Yiqun Yao, Siqi Fan, Xiusheng Huang, Xuezhi Fang, Xiang Li, Ziyi Ni, Xin Jiang, Xuying Meng, Peng Han, Shuo Shang, Kang Liu, Aixin Sun, Yequan Wang
With around 14% of the one-time pre-training cost, we can accurately forecast the loss for models up to 52B.
1 code implementation • 23 Jun 2022 • Shufang Xie, Rui Yan, Peng Han, Yingce Xia, Lijun Wu, Chenjuan Guo, Bin Yang, Tao Qin
We observe that the same intermediate molecules are visited many times in the searching process, and they are usually independently treated in previous tree-based methods (e. g., AND-OR tree search, Monte Carlo tree search).
Ranked #2 on
Multi-step retrosynthesis
on USPTO-190
no code implementations • 8 Oct 2021 • Peng Han, Yanning Guo, Chuanjiang Li, Hui Zhi, Yueyong Lv
This paper proposed a novel large neighborhood search-adaptive genetic algorithm (LNS-AGA) for many-to-many on-orbit repairing mission planning of geosynchronous orbit (GEO) satellites with mission deadline constraint.
no code implementations • 29 Oct 2019 • Rui Fan, Yu-An Wang, Lei Qiao, Ruiwen Yao, Peng Han, Weidong Zhang, Ioannis Pitas, Ming Liu
This linear model is then utilized to reduce the redundant information in the left and right road images.
no code implementations • 5 Feb 2015 • Peng Han
The experimental results demonstrate the promising performance of the proposed method for segmentation with selectively propagated constraints.