no code implementations • 11 Mar 2025 • Siqi Fan, Xuezhi Fang, Xingrun Xing, Peng Han, Shuo Shang, Yequan Wang
Experiments on large language models (\ie the Llama) with $7 \sim 70$ billion parameters show that $D^3$ can achieve an average 1. 5x speedup compared with the full-inference pipeline while maintaining comparable performance with nearly no performance drop ($<1\%$) on the GSM8K and BBH benchmarks.
no code implementations • 7 Jan 2025 • Xiaoqing Zhang, Ang Lv, YuHan Liu, Flood Sung, Wei Liu, Shuo Shang, Xiuying Chen, Rui Yan
Recognizing the lack of multi-task datasets with diverse many-shot distributions, we develop the Many-Shot ICL Benchmark (ICL-50)-a large-scale benchmark of 50 tasks that cover shot numbers from 1 to 350 within sequences of up to 8, 000 tokens-for fine-tuning purposes.
no code implementations • 22 Nov 2024 • Silin Zhou, Shuo Shang, Lisi Chen, Christian S. Jensen, Panos Kalnis
Trajectory representation learning (TRL) maps trajectories to vectors that can then be used for various downstream tasks, including trajectory similarity computation, trajectory classification, and travel-time estimation.
1 code implementation • 22 Nov 2024 • Silin Zhou, Shuo Shang, Lisi Chen, Peng Han, Christian S. Jensen
Trajectory representation learning (TRL) maps trajectories to vectors that can be used for many downstream tasks.
no code implementations • 10 Nov 2024 • Shuqi Li, Yuebo Sun, Yuxin Lin, Xin Gao, Shuo Shang, Rui Yan
On the other hand, there is substantial noise existing in the news data leading to extracting effective information with difficulty.
1 code implementation • 23 Sep 2024 • Qinzhuo Wu, Weikai Xu, Wei Liu, Tao Tan, Jianfeng Liu, Ang Li, Jian Luan, Bin Wang, Shuo Shang
These fine-tuned VLMs may still ignore the relationships between UI pages, neglect the roles of elements in page transitions and lack inter-UI understanding.
no code implementations • 3 Jul 2024 • Chengrui Huang, Zhengliang Shi, Yuntao Wen, Xiuying Chen, Peng Han, Shen Gao, Shuo Shang
Tool learning methods have enhanced the ability of large language models (LLMs) to interact with real-world applications.
1 code implementation • 1 Jul 2024 • Shihan Deng, Weikai Xu, Hongda Sun, Wei Liu, Tao Tan, Jianfeng Liu, Ang Li, Jian Luan, Bin Wang, Rui Yan, Shuo Shang
With the remarkable advancements of large language models (LLMs), LLM-based agents have become a research hotspot in human-computer interaction.
no code implementations • 28 Jun 2024 • Shen Gao, Yuntao Wen, Minghang Zhu, Jianing Wei, YuHan Cheng, Qunzi Zhang, Shuo Shang
In this paper, we propose \textbf{A}gent-based \textbf{S}imulated \textbf{F}inancial \textbf{M}arket (ASFM), which first constructs a simulated stock market with a real order matching system.
no code implementations • 28 May 2024 • Hongda Sun, Hongzhan Lin, Haiyu Yan, Chen Zhu, Yang song, Xin Gao, Shuo Shang, Rui Yan
To this end, we propose MockLLM, a novel applicable framework that divides the person-job matching process into two modules: mock interview generation and two-sided evaluation in handshake protocol, jointly enhancing their performance through collaborative behaviors between interviewers and candidates.
no code implementations • 9 Apr 2024 • Shen Gao, Yifan Wang, Jiabao Fang, Lisi Chen, Peng Han, Shuo Shang
Recommendation systems play a crucial role in various domains, suggesting items based on user behavior. However, the lack of transparency in presenting recommendations can lead to user confusion.
1 code implementation • 8 Apr 2024 • Shen Gao, Hao Li, Chengrui Huang, Quan Tu, Zhiliang Tian, Minlie Huang, Shuo Shang
The framework employs a novel 360$^\circ$ performance assessment method for multi-perspective performance evaluation with fine-grained assessment.
1 code implementation • 8 Mar 2024 • Hongda Sun, Yuxuan Liu, ChengWei Wu, Haiyu Yan, Cheng Tai, Xin Gao, Shuo Shang, Rui Yan
Open-domain question answering (ODQA) has emerged as a pivotal research spotlight in information systems.
no code implementations • 5 Mar 2024 • Chuanqi Cheng, Quan Tu, Shuo Shang, Cunli Mao, Zhengtao Yu, Wei Wu, Rui Yan
Personalized dialogue systems have gained significant attention in recent years for their ability to generate responses in alignment with different personas.
no code implementations • 4 Mar 2024 • Siqi Fan, Xin Jiang, Xiang Li, Xuying Meng, Peng Han, Shuo Shang, Aixin Sun, Yequan Wang, Zhongyuan Wang
That is, not all layers of LLMs are necessary during inference.
no code implementations • 28 Feb 2024 • Yang Cao, Shuo Shang, Jun Wang, Wei zhang
This paper explores providing explainability for session-based recommendation (SR) by path reasoning.
1 code implementation • 28 Oct 2023 • Hongda Sun, Weikai Xu, Wei Liu, Jian Luan, Bin Wang, Shuo Shang, Ji-Rong Wen, Rui Yan
Recent advances in large language models (LLMs) have revolutionized the landscape of reasoning tasks.
no code implementations • 24 Oct 2023 • Yuxiang Wang, Xiao Yan, Chuang Hu, Fangcheng Fu, Wentao Zhang, Hao Wang, Shuo Shang, Jiawei Jiang
For graph self-supervised learning (GSSL), masked autoencoder (MAE) follows the generative paradigm and learns to reconstruct masked graph edges or node features.
1 code implementation • 31 Aug 2023 • Qiang Huang, Jiawei Jiang, Xi Susie Rao, Ce Zhang, Zhichao Han, Zitao Zhang, Xin Wang, Yongjun He, Quanqing Xu, Yang Zhao, Chuang Hu, Shuo Shang, Bo Du
To handle graphs in which features or connectivities are evolving over time, a series of temporal graph neural networks (TGNNs) have been proposed.
1 code implementation • 20 Aug 2023 • Quan Tu, Chuanqi Chen, Jinpeng Li, Yanran Li, Shuo Shang, Dongyan Zhao, Ran Wang, Rui Yan
In our modern, fast-paced, and interconnected world, the importance of mental well-being has grown into a matter of great urgency.
1 code implementation • 14 Apr 2023 • Yiqun Yao, Siqi Fan, Xiusheng Huang, Xuezhi Fang, Xiang Li, Ziyi Ni, Xin Jiang, Xuying Meng, Peng Han, Shuo Shang, Kang Liu, Aixin Sun, Yequan Wang
With around 14% of the one-time pre-training cost, we can accurately forecast the loss for models up to 52B.