Search Results for author: Chak Tou Leong

Found 15 papers, 8 papers with code

STeCa: Step-level Trajectory Calibration for LLM Agent Learning

1 code implementation20 Feb 2025 Hanlin Wang, Jian Wang, Chak Tou Leong, Wenjie Li

To address this, we highlight the importance of timely calibration and the need to automatically construct calibration trajectories for training agents.

Decision Making Language Modeling +2

Why Safeguarded Ships Run Aground? Aligned Large Language Models' Safety Mechanisms Tend to Be Anchored in The Template Region

no code implementations19 Feb 2025 Chak Tou Leong, Qingyu Yin, Jian Wang, Wenjie Li

The safety alignment of large language models (LLMs) remains vulnerable, as their initial behavior can be easily jailbroken by even relatively simple attacks.

Decision Making Safety Alignment

TokenSkip: Controllable Chain-of-Thought Compression in LLMs

1 code implementation17 Feb 2025 Heming Xia, Yongqi Li, Chak Tou Leong, Wenjie Wang, Wenjie Li

Chain-of-Thought (CoT) has been proven effective in enhancing the reasoning capabilities of large language models (LLMs).

GSM8K

Subtle Errors Matter: Preference Learning via Error-injected Self-editing

no code implementations9 Oct 2024 Kaishuai Xu, Tiezheng Yu, Wenjun Hou, Yi Cheng, Chak Tou Leong, Liangyou Li, Xin Jiang, Lifeng Shang, Qun Liu, Wenjie Li

In this work, we propose a novel preference learning framework called eRror-Injected Self-Editing (RISE), which injects predefined subtle errors into partial tokens of correct solutions to construct hard pairs for error mitigation.

GSM8K Math +1

Deeper Insights Without Updates: The Power of In-Context Learning Over Fine-Tuning

1 code implementation7 Oct 2024 Qingyu Yin, Xuzheng He, Luoao Deng, Chak Tou Leong, Fan Wang, Yanzhao Yan, Xiaoyu Shen, Qiang Zhang

Fine-tuning and in-context learning (ICL) are two prevalent methods in imbuing large language models with task-specific knowledge.

In-Context Learning

E2CL: Exploration-based Error Correction Learning for Embodied Agents

no code implementations5 Sep 2024 Hanlin Wang, Chak Tou Leong, Jian Wang, Wenjie Li

Language models are exhibiting increasing capability in knowledge utilization and reasoning.

AutoPal: Autonomous Adaptation to Users for Personal AI Companionship

no code implementations20 Jun 2024 Yi Cheng, Wenge Liu, Kaishuai Xu, Wenjun Hou, Yi Ouyang, Chak Tou Leong, Xian Wu, Yefeng Zheng

However, imbuing agents with autonomous adaptability presents unique challenges, including identifying optimal adaptations to meet users' expectations and ensuring a smooth transition during the adaptation process.

No Two Devils Alike: Unveiling Distinct Mechanisms of Fine-tuning Attacks

no code implementations25 May 2024 Chak Tou Leong, Yi Cheng, Kaishuai Xu, Jian Wang, Hanlin Wang, Wenjie Li

In particular, we analyze the two most representative types of attack approaches: Explicit Harmful Attack (EHA) and Identity-Shifting Attack (ISA).

Safety Alignment

Instruct Once, Chat Consistently in Multiple Rounds: An Efficient Tuning Framework for Dialogue

1 code implementation10 Feb 2024 Jian Wang, Chak Tou Leong, Jiashuo Wang, Dongding Lin, Wenjie Li, Xiao-Yong Wei

Tuning language models for dialogue generation has been a prevalent paradigm for building capable dialogue agents.

Dialogue Generation

Mitigating Unhelpfulness in Emotional Support Conversations with Multifaceted AI Feedback

1 code implementation11 Jan 2024 Jiashuo Wang, Chunpu Xu, Chak Tou Leong, Wenjie Li, Jing Li

An emotional support conversation system aims to alleviate users' emotional distress and assist them in addressing their challenges.

Contrastive Learning

COOPER: Coordinating Specialized Agents towards a Complex Dialogue Goal

1 code implementation19 Dec 2023 Yi Cheng, Wenge Liu, Jian Wang, Chak Tou Leong, Yi Ouyang, Wenjie Li, Xian Wu, Yefeng Zheng

In recent years, there has been a growing interest in exploring dialogues with more complex goals, such as negotiation, persuasion, and emotional support, which go beyond traditional service-focused dialogue systems.

Self-Detoxifying Language Models via Toxification Reversal

2 code implementations14 Oct 2023 Chak Tou Leong, Yi Cheng, Jiashuo Wang, Jian Wang, Wenjie Li

Drawing on this idea, we devise a method to identify the toxification direction from the normal generation process to the one prompted with the negative prefix, and then steer the generation to the reversed direction by manipulating the information movement within the attention layers.

Language Modeling Language Modelling

Target-oriented Proactive Dialogue Systems with Personalization: Problem Formulation and Dataset Curation

1 code implementation11 Oct 2023 Jian Wang, Yi Cheng, Dongding Lin, Chak Tou Leong, Wenjie Li

Target-oriented dialogue systems, designed to proactively steer conversations toward predefined targets or accomplish specific system-side goals, are an exciting area in conversational AI.

Cannot find the paper you are looking for? You can Submit a new open access paper.