no code implementations • 28 Oct 2024 • Weizhe Chen, Zhicheng Zhang, Guanlin Liu, Renjie Zheng, Wenlei Shi, Chen Dun, Zheng Wu, Xing Jin, Lin Yan
Since the release of ChatGPT, large language models (LLMs) have demonstrated remarkable capabilities across various domains.
no code implementations • 23 Oct 2024 • Ning Dai, Zheng Wu, Renjie Zheng, Ziyun Wei, Wenlei Shi, Xing Jin, Guanlin Liu, Chen Dun, Liang Huang, Lin Yan
Reinforcement Learning (RL) with unit test feedback has enhanced large language models (LLMs) code generation, but relies on sparse rewards provided only after complete code evaluation, limiting learning efficiency and incremental improvements.
no code implementations • 11 Oct 2024 • Guanlin Liu, Kaixuan Ji, Renjie Zheng, Zheng Wu, Chen Dun, Quanquan Gu, Lin Yan
Reinforcement Learning (RL) plays a crucial role in aligning large language models (LLMs) with human preferences and improving their ability to perform complex tasks.
no code implementations • 1 Nov 2023 • Ziqing Lu, Guanlin Liu, Lifeng Cai, Weiyu Xu
Finding optimal adversarial attack strategies is an important topic in reinforcement learning and the Markov decision process.
no code implementations • 15 Jul 2023 • Guanlin Liu, Zhihan Zhou, Han Liu, Lifeng Lai
Robust reinforcement learning (RL) aims to find a policy that optimizes the worst-case performance in the face of uncertainties.
no code implementations • 10 Dec 2021 • Guanlin Liu, Lifeng Lai
We show that, in both white-box and black-box settings, the proposed attack schemes can force the LinUCB agent to pull a target arm very frequently by spending only logarithm cost.
no code implementations • NeurIPS 2021 • Guanlin Liu, Lifeng Lai
In this paper, we introduce a new class of attacks named action poisoning attacks, where an adversary can change the action signal selected by the agent.
no code implementations • 19 Feb 2020 • Guanlin Liu, Lifeng Lai
To defend against this class of attacks, we introduce a novel algorithm that is robust to action-manipulation attacks when an upper bound for the total attack cost is given.