1 code implementation • 8 May 2025 • Shiqi Chen, Jinghan Zhang, Tongyao Zhu, Wei Liu, Siyang Gao, Miao Xiong, Manling Li, Junxian He
Vision-Language Models (VLMs) combine visual perception with the general capabilities, such as reasoning, of Large Language Models (LLMs).
no code implementations • 20 Mar 2025 • Jinghan Zhang, Xiting Wang, Fengran Mo, Yeyang Zhou, Wanfu Gao, Kunpeng Liu
In large language model (LLM) reasoning, multi-step processes have proven effective for solving complex tasks.
1 code implementation • 3 Mar 2025 • Shiqi Chen, Tongyao Zhu, Ruochen Zhou, Jinghan Zhang, Siyang Gao, Juan Carlos Niebles, Mor Geva, Junxian He, Jiajun Wu, Manling Li
By tracing attention distribution over the image through out intermediate layers, we observe that successful spatial reasoning correlates strongly with the model's ability to align its attention distribution with actual object locations, particularly differing between familiar and unfamiliar spatial relationships.
no code implementations • 17 Feb 2025 • Zaitian Wang, Jinghan Zhang, Xinhao Zhang, Kunpeng Liu, Pengfei Wang, Yuanchun Zhou
Data augmentation is an essential technique in natural language processing (NLP) for enriching training datasets by generating diverse samples.
no code implementations • 29 Jan 2025 • Xinhao Zhang, Jinghan Zhang, Fengran Mo, Dongjie Wang, Yanjie Fu, Kunpeng Liu
Therefore, we design a knowledge augmentation method LEKA for knowledge transfer that actively searches for suitable knowledge sources that can enrich the target domain's knowledge.
1 code implementation • 28 Dec 2024 • Henry J. Xie, Jinghan Zhang, Xinhao Zhang, Kunpeng Liu
We develop a novel and comprehensive framework for investigating how effective LLMs are at measuring and scoring empathy of responses in dialogues, and what methods can be employed to deepen our understanding of LLM scoring.
no code implementations • 4 Nov 2024 • Xinhao Zhang, Jinghan Zhang, Wujun Si, Kunpeng Liu
Deep Reinforcement Learning has shown excellent performance in generating efficient solutions for complex tasks.
no code implementations • 4 Nov 2024 • Jinghan Zhang, Henry Xie, Xinhao Zhang, Kunpeng Liu
In the financial field, precise risk assessment tools are essential for decision-making.
no code implementations • 31 Oct 2024 • Jinghan Zhang, Fengran Mo, Xiting Wang, Kunpeng Liu
Recent advances in large language models (LLMs) have demonstrated their potential in handling complex reasoning tasks, which are usually achieved by constructing a thought chain to guide the model to solve the problem with multi-step thinking.
no code implementations • 17 Jun 2024 • Xinhao Zhang, Jinghan Zhang, Fengran Mo, Yuzhong Chen, Kunpeng Liu
Feature generation can significantly enhance learning outcomes, particularly for tasks with limited data.
no code implementations • 6 Jun 2024 • Jinghan Zhang, Xiting Wang, Yiqiao Jin, Changyu Chen, Xinhao Zhang, Kunpeng Liu
The reward model for Reinforcement Learning from Human Feedback (RLHF) has proven effective in fine-tuning Large Language Models (LLMs).
1 code implementation • 4 Jun 2024 • Jinghan Zhang, Xiting Wang, Weijieying Ren, Lu Jiang, Dongjie Wang, Kunpeng Liu
To address these limitations, we introduce the Retrieval Augmented Thought Tree (RATT), a novel thought structure that considers both overall logical soundness and factual correctness at each step of the thinking process.
no code implementations • 4 Jun 2024 • Xinhao Zhang, Jinghan Zhang, Banafsheh Rekabdar, Yuanchun Zhou, Pengfei Wang, Kunpeng Liu
The representation of feature space is a crucial environment where data points get vectorized and embedded for upcoming modeling.
1 code implementation • 15 Apr 2024 • Yuzhen Huang, Jinghan Zhang, Zifei Shan, Junxian He
We open-source our compression datasets as well as our data collection pipelines to facilitate future researchers to assess compression properly.
no code implementations • 18 Oct 2023 • Le Ma, Ran Zhang, Yikun Han, Shirui Yu, Zaitian Wang, Zhiyuan Ning, Jinghan Zhang, Ping Xu, Pengjiang Li, Wei Ju, Chong Chen, Dongjie Wang, Kunpeng Liu, Pengyang Wang, Pengfei Wang, Yanjie Fu, Chunjiang Liu, Yuanchun Zhou, Chang-Tien Lu
In this work, we present a comprehensive review of the relevant algorithms to provide a general understanding of this booming research area.
1 code implementation • NeurIPS 2023 • Shiqi Chen, Yiran Zhao, Jinghan Zhang, I-Chun Chern, Siyang Gao, PengFei Liu, Junxian He
In this benchmark, we collect responses generated from LLMs and annotate factuality labels in a fine-grained manner.
2 code implementations • 26 Jun 2023 • Jinghan Zhang, Shiqi Chen, Junteng Liu, Junxian He
In this paper, we propose to compose these parameter-efficient modules through linear arithmetic operations in the weight space, thereby integrating different module capabilities.
1 code implementation • NeurIPS 2023 • Yuzhen Huang, Yuzhuo Bai, Zhihao Zhu, Junlei Zhang, Jinghan Zhang, Tangjun Su, Junteng Liu, Chuancheng Lv, Yikai Zhang, Jiayi Lei, Yao Fu, Maosong Sun, Junxian He
We present C-Eval, the first comprehensive Chinese evaluation suite designed to assess advanced knowledge and reasoning abilities of foundation models in a Chinese context.