Search Results for author: Zhehua Zhou

Found 9 papers, 2 papers with code

Multilingual Blending: LLM Safety Alignment Evaluation with Language Mixture

no code implementations10 Jul 2024 Jiayang Song, Yuheng Huang, Zhehua Zhou, Lei Ma

As safety remains a crucial concern throughout the development lifecycle of Large Language Models (LLMs), researchers and industrial practitioners have increasingly focused on safeguarding and aligning LLM behaviors with human preferences and ethical standards.

Safety Alignment

GenSafe: A Generalizable Safety Enhancer for Safe Reinforcement Learning Algorithms Based on Reduced Order Markov Decision Process Model

no code implementations6 Jun 2024 Zhehua Zhou, Xuan Xie, Jiayang Song, Zhan Shu, Lei Ma

To address this issue, we introduce in this work a novel Generalizable Safety enhancer (GenSafe) that is able to overcome the challenge of data insufficiency and enhance the performance of SRL approaches.

Autonomous Vehicles Deep Reinforcement Learning +2

Online Safety Analysis for LLMs: a Benchmark, an Assessment, and a Path Forward

no code implementations12 Apr 2024 Xuan Xie, Jiayang Song, Zhehua Zhou, Yuheng Huang, Da Song, Lei Ma

To bridge this gap, we conduct in this work a comprehensive evaluation of the effectiveness of existing online safety analysis methods on LLMs.

Fairness

ISR-LLM: Iterative Self-Refined Large Language Model for Long-Horizon Sequential Task Planning

2 code implementations26 Aug 2023 Zhehua Zhou, Jiayang Song, Kunpeng Yao, Zhan Shu, Lei Ma

Motivated by the substantial achievements observed in Large Language Models (LLMs) in the field of natural language processing, recent research has commenced investigations into the application of LLMs for complex, long-horizon sequential task planning challenges in robotics.

Language Modeling Language Modelling +2

Data Generation Method for Learning a Low-dimensional Safe Region in Safe Reinforcement Learning

no code implementations10 Sep 2021 Zhehua Zhou, Ozgur S. Oguz, Yi Ren, Marion Leibold, Martin Buss

Safe reinforcement learning aims to learn a control policy while ensuring that neither the system nor the environment gets damaged during the learning process.

reinforcement-learning Reinforcement Learning +2

Off-Policy Risk-Sensitive Reinforcement Learning Based Constrained Robust Optimal Control

no code implementations10 Jun 2020 Cong Li, Qingchen Liu, Zhehua Zhou, Martin Buss, Fangzhou Liu

By introducing pseudo controls and risk-sensitive input and state penalty terms, the constrained robust stabilization problem of the original system is converted into an equivalent optimal control problem of an auxiliary system.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.