Search Results for author: Grace Zhang

Found 6 papers, 3 papers with code

Efficient Multi-Task Reinforcement Learning via Selective Behavior Sharing

no code implementations1 Feb 2023 Grace Zhang, Ayush Jain, Injune Hwang, Shao-Hua Sun, Joseph J. Lim

The ability to leverage shared behaviors between tasks is critical for sample-efficient multi-task reinforcement learning (MTRL).

reinforcement-learning Reinforcement Learning (RL)

CoMPS: Continual Meta Policy Search

no code implementations ICLR 2022 Glen Berseth, Zhiwei Zhang, Grace Zhang, Chelsea Finn, Sergey Levine

Beyond simply transferring past experience to new tasks, our goal is to devise continual reinforcement learning algorithms that learn to learn, using their experience on previous tasks to learn new tasks more quickly.

Continual Learning Continuous Control +5

Policy Transfer across Visual and Dynamics Domain Gaps via Iterative Grounding

1 code implementation1 Jul 2021 Grace Zhang, Linghan Zhong, Youngwoon Lee, Joseph J. Lim

In this paper, we propose a novel policy transfer method with iterative "environment grounding", IDAPT, that alternates between (1) directly minimizing both visual and dynamics domain gaps by grounding the source environment in the target environment domains, and (2) training a policy on the grounded source environment.

Advantage-Weighted Regression: Simple and Scalable Off-Policy Reinforcement Learning

5 code implementations1 Oct 2019 Xue Bin Peng, Aviral Kumar, Grace Zhang, Sergey Levine

In this paper, we aim to develop a simple and scalable reinforcement learning algorithm that uses standard supervised learning methods as subroutines.

Continuous Control OpenAI Gym +3

Advantage Weighted Regression: Simple and Scalable Off-Policy Reinforcement Learning

no code implementations25 Sep 2019 Xue Bin Peng, Aviral Kumar, Grace Zhang, Sergey Levine

In this paper, we aim to develop a simple and scalable reinforcement learning algorithm that uses standard supervised learning methods as subroutines.

Continuous Control OpenAI Gym +3

MCP: Learning Composable Hierarchical Control with Multiplicative Compositional Policies

1 code implementation NeurIPS 2019 Xue Bin Peng, Michael Chang, Grace Zhang, Pieter Abbeel, Sergey Levine

In this work, we propose multiplicative compositional policies (MCP), a method for learning reusable motor skills that can be composed to produce a range of complex behaviors.

Continuous Control

Cannot find the paper you are looking for? You can Submit a new open access paper.