no code implementations • 7 Mar 2024 • Tairan He, Zhengyi Luo, Wenli Xiao, Chong Zhang, Kris Kitani, Changliu Liu, Guanya Shi
We present Human to Humanoid (H2O), a reinforcement learning (RL) based framework that enables real-time whole-body teleoperation of a full-sized humanoid robot with only an RGB camera.
no code implementations • 31 Jan 2024 • Tairan He, Chong Zhang, Wenli Xiao, Guanqi He, Changliu Liu, Guanya Shi
Legged robots navigating cluttered environments must be jointly agile for efficient task execution and safe to avoid collisions with obstacles or humans.
1 code implementation • 8 Oct 2023 • Wenli Xiao, Tairan He, John Dolan, Guanya Shi
In contrast, policy adaptation based on reinforcement learning (RL) offers versatility and generalizability but presents safety and robustness challenges.
no code implementations • 6 Feb 2023 • WeiYe Zhao, Tairan He, Rui Chen, Tianhao Wei, Changliu Liu
Despite the tremendous success of Reinforcement Learning (RL) algorithms in simulation environments, applying RL to real-world applications still faces many challenges.
1 code implementation • 2 Feb 2023 • Minghuan Liu, Tairan He, Weinan Zhang, Shuicheng Yan, Zhongwen Xu
Specifically, we present Adversarial Imitation Learning with Patch Rewards (PatchAIL), which employs a patch-based discriminator to measure the expertise of different local parts from given images and provide patch rewards.
no code implementations • 24 Jan 2023 • Tairan He, WeiYe Zhao, Changliu Liu
Results show that the converged policies with intrinsic costs in all environments achieve zero constraint violation and comparable performance with baselines.
no code implementations • 12 Oct 2022 • Tairan He, Yuge Zhang, Kan Ren, Minghuan Liu, Che Wang, Weinan Zhang, Yuqing Yang, Dongsheng Li
A good state representation is crucial to solving complicated reinforcement learning (RL) challenges.
no code implementations • 29 Sep 2021 • Tairan He, Yuge Zhang, Kan Ren, Che Wang, Weinan Zhang, Dongsheng Li, Yuqing Yang
A good state representation is crucial to reinforcement learning (RL) while an ideal representation is hard to learn only with signals from the RL objective.
1 code implementation • 20 Apr 2020 • Minghuan Liu, Tairan He, Minkai Xu, Wei-Nan Zhang
We tackle a common scenario in imitation learning (IL), where agents try to recover the optimal policy from expert demonstrations without further access to the expert or environment reward signals.