Search Results for author: Nai-Chieh Huang

Found 3 papers, 0 papers with code

PPO-Clip Attains Global Optimality: Towards Deeper Understandings of Clipping

no code implementations19 Dec 2023 Nai-Chieh Huang, Ping-Chun Hsieh, Kuo-Hao Ho, I-Chen Wu

Our findings highlight the $O(1/\sqrt{T})$ min-iterate convergence rate specifically in the context of neural function approximation.

Accelerated Policy Gradient: On the Convergence Rates of the Nesterov Momentum for Reinforcement Learning

no code implementations18 Oct 2023 Yen-ju Chen, Nai-Chieh Huang, Ping-Chun Hsieh

In response to this gap, we adapt the celebrated Nesterov's accelerated gradient (NAG) method to policy optimization in RL, termed \textit{Accelerated Policy Gradient} (APG).

Policy Gradient Methods reinforcement-learning +1

Neural PPO-Clip Attains Global Optimality: A Hinge Loss Perspective

no code implementations26 Oct 2021 Nai-Chieh Huang, Ping-Chun Hsieh, Kuo-Hao Ho, Hsuan-Yu Yao, Kai-Chun Hu, Liang-Chun Ouyang, I-Chen Wu

Policy optimization is a fundamental principle for designing reinforcement learning algorithms, and one example is the proximal policy optimization algorithm with a clipped surrogate objective (PPO-Clip), which has been popularly used in deep reinforcement learning due to its simplicity and effectiveness.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.