Search Results for author: Huanxuan Liao

Found 11 papers, 10 papers with code

Dynamic Parametric Retrieval Augmented Generation for Test-time Knowledge Enhancement

1 code implementation31 Mar 2025 Yuqiao Tan, Shizhu He, Huanxuan Liao, Jun Zhao, Kang Liu

Retrieval-augmented generation (RAG) enhances large language models (LLMs) by retrieving relevant documents from external sources and incorporating them into the context.

Hallucination RAG +1

Evaluating Personalized Tool-Augmented LLMs from the Perspectives of Personalization and Proactivity

1 code implementation2 Mar 2025 Yupu Hao, Pengfei Cao, Zhuoran Jin, Huanxuan Liao, Yubo Chen, Kang Liu, Jun Zhao

Personalized tool utilization is essential for aligning large language models (LLMs) with user preference in interaction scenarios with various tools.

Text Generation

DATA: Decomposed Attention-based Task Adaptation for Rehearsal-Free Continual Learning

1 code implementation17 Feb 2025 Huanxuan Liao, Shizhu He, Yupu Hao, Jun Zhao, Kang Liu

For new tasks, DATA dynamically adjusts the weights of adapters of different ranks based on their relevance and distinction from previous tasks, allowing the model to acquire new task-specific skills while effectively retaining previously learned knowledge.

Continual Learning

CITI: Enhancing Tool Utilizing Ability in Large Language Models without Sacrificing General Performance

1 code implementation20 Sep 2024 Yupu Hao, Pengfei Cao, Zhuoran Jin, Huanxuan Liao, Yubo Chen, Kang Liu, Jun Zhao

However, previous works predominantly focus on improving model's tool-utilizing accuracy and the ability to generalize to new, unseen tools, excessively forcing LLMs to adjust specific tool-invoking pattern without considering the harm to model's general performance.

Neural-Symbolic Collaborative Distillation: Advancing Small Language Models for Complex Reasoning Tasks

1 code implementation20 Sep 2024 Huanxuan Liao, Shizhu He, Yao Xu, Yuanzhe Zhang, Kang Liu, Jun Zhao

By decoupling general and specialized capabilities, the proposed NesyCD can achieve superior performance cost-effectively, utilizing smaller models and blending parameterized neural networks with symbolic KB.

ARC GSM8K +1

$\textit{SKIntern}$: Internalizing Symbolic Knowledge for Distilling Better CoT Capabilities into Small Language Models

1 code implementation20 Sep 2024 Huanxuan Liao, Shizhu He, Yupu Hao, Xiang Li, Yuanzhe Zhang, Jun Zhao, Kang Liu

By efficiently internalizing knowledge, $\textit{SKIntern}$ reduces computational overhead and speeds up the reasoning process by focusing solely on the question during inference.

From Instance Training to Instruction Learning: Task Adapters Generation from Instructions

2 code implementations18 Jun 2024 Huanxuan Liao, Shizhu He, Yao Xu, Yuanzhe Zhang, Yanchao Hao, Shengping Liu, Kang Liu, Jun Zhao

Within this context, we introduce Task Adapters Generation from Instructions (TAGI), which automatically constructs the task-specific model in a parameter generation manner based on the given task instructions without retraining for unseen tasks.

Knowledge Distillation

Awakening Augmented Generation: Learning to Awaken Internal Knowledge of Large Language Models for Question Answering

1 code implementation22 Mar 2024 Huanxuan Liao, Shizhu He, Yao Xu, Yuanzhe Zhang, Kang Liu, Shengping Liu, Jun Zhao

Retrieval-Augmented-Generation and Generation-Augmented-Generation have been proposed to enhance the knowledge required for question answering with Large Language Models (LLMs) by leveraging richer context.

Open-Domain Question Answering Out-of-Distribution Generalization

LMTuner: An user-friendly and highly-integrable Training Framework for fine-tuning Large Language Models

1 code implementation20 Aug 2023 Yixuan Weng, Zhiqi Wang, Huanxuan Liao, Shizhu He, Shengping Liu, Kang Liu, Jun Zhao

With the burgeoning development in the realm of large language models (LLMs), the demand for efficient incremental training tailored to specific industries and domains continues to increase.

Cannot find the paper you are looking for? You can Submit a new open access paper.