no code implementations • 12 Nov 2023 • Zeyang Li, Chuxiong Hu, WeiYe Zhao, Changliu Liu
This paper presents a theoretical framework that bridges the advantages of both RMPC and RL to synthesize safety filters for nonlinear systems with state- and action-dependent uncertainty.
1 code implementation • 20 Oct 2023 • WeiYe Zhao, Feihan Li, Yifan Sun, Rui Chen, Tianhao Wei, Changliu Liu
In recent years, trust region on-policy reinforcement learning has achieved impressive results in addressing complex control tasks and gaming scenarios.
no code implementations • 21 Sep 2023 • Rui Chen, WeiYe Zhao, Changliu Liu
This paper introduces an approach for synthesizing feasible safety indices to derive safe control laws under state-dependent control spaces.
1 code implementation • 21 Jun 2023 • WeiYe Zhao, Rui Chen, Yifan Sun, Tianhao Wei, Changliu Liu
In particular, we introduce the framework of Maximum Markov Decision Process, and prove that the worst-case safety violation is bounded under SCPO.
1 code implementation • 23 May 2023 • WeiYe Zhao, Rui Chen, Yifan Sun, Ruixuan Liu, Tianhao Wei, Changliu Liu
Due to the diversity of algorithms and tasks, it remains difficult to compare existing safe RL algorithms.
no code implementations • 6 Feb 2023 • WeiYe Zhao, Tairan He, Rui Chen, Tianhao Wei, Changliu Liu
Despite the tremendous success of Reinforcement Learning (RL) algorithms in simulation environments, applying RL to real-world applications still faces many challenges.
no code implementations • 24 Jan 2023 • Tairan He, WeiYe Zhao, Changliu Liu
Results show that the converged policies with intrinsic costs in all environments achieve zero constraint violation and comparable performance with baselines.
1 code implementation • 18 Apr 2021 • WeiYe Zhao, Suqin He, Changliu Liu
Tolerance estimation problems are prevailing in engineering applications.