Search Results for author: Cong Lu

Found 12 papers, 8 papers with code

Policy-Guided Diffusion

1 code implementation9 Apr 2024 Matthew Thomas Jackson, Michael Tryfan Matthews, Cong Lu, Benjamin Ellis, Shimon Whiteson, Jakob Foerster

Our approach provides an effective alternative to autoregressive offline world models, opening the door to the controllable generation of synthetic training data.

The Edge-of-Reach Problem in Offline Model-Based Reinforcement Learning

2 code implementations19 Feb 2024 Anya Sims, Cong Lu, Yee Whye Teh

The prevailing theoretical understanding is that this can then be viewed as online reinforcement learning in an approximate dynamics model, and any remaining gap is therefore assumed to be due to the imperfect dynamics model.

Model-based Reinforcement Learning reinforcement-learning

Synthetic Experience Replay

1 code implementation NeurIPS 2023 Cong Lu, Philip J. Ball, Yee Whye Teh, Jack Parker-Holder

We believe that synthetic training data could open the door to realizing the full potential of deep learning for replay-based RL algorithms from limited data.

Reinforcement Learning (RL) Self-Supervised Learning

On Pathologies in KL-Regularized Reinforcement Learning from Expert Demonstrations

1 code implementation NeurIPS 2021 Tim G. J. Rudner, Cong Lu, Michael A. Osborne, Yarin Gal, Yee Whye Teh

KL-regularized reinforcement learning from expert demonstrations has proved successful in improving the sample efficiency of deep reinforcement learning algorithms, allowing them to be applied to challenging physical real-world tasks.

reinforcement-learning Reinforcement Learning (RL)

Go-Explore Complex 3D Game Environments for Automated Reachability Testing

no code implementations1 Sep 2022 Cong Lu, Raluca Georgescu, Johan Verwey

Modern AAA video games feature huge game levels and maps which are increasingly hard for level testers to cover exhaustively.

Bayesian Generational Population-Based Training

2 code implementations19 Jul 2022 Xingchen Wan, Cong Lu, Jack Parker-Holder, Philip J. Ball, Vu Nguyen, Binxin Ru, Michael A. Osborne

Leveraging the new highly parallelizable Brax physics engine, we show that these innovations lead to large performance gains, significantly outperforming the tuned baseline while learning entire configurations on the fly.

Bayesian Optimization Reinforcement Learning (RL)

Challenges and Opportunities in Offline Reinforcement Learning from Visual Observations

2 code implementations9 Jun 2022 Cong Lu, Philip J. Ball, Tim G. J. Rudner, Jack Parker-Holder, Michael A. Osborne, Yee Whye Teh

Using this suite of benchmarking tasks, we show that simple modifications to two popular vision-based online reinforcement learning algorithms, DreamerV2 and DrQ-v2, suffice to outperform existing offline RL methods and establish competitive baselines for continuous control in the visual domain.

Benchmarking Continuous Control +3

Revisiting Design Choices in Offline Model-Based Reinforcement Learning

no code implementations8 Oct 2021 Cong Lu, Philip J. Ball, Jack Parker-Holder, Michael A. Osborne, Stephen J. Roberts

Significant progress has been made recently in offline model-based reinforcement learning, approaches which leverage a learned dynamics model.

Bayesian Optimization Model-based Reinforcement Learning +2

Revisiting Design Choices in Offline Model Based Reinforcement Learning

no code implementations NeurIPS 2021 Cong Lu, Philip Ball, Jack Parker-Holder, Michael Osborne, S Roberts

Offline reinforcement learning enables agents to make use of large pre-collected datasets of environment transitions and learn control policies without the need for potentially expensive or unsafe online data collection.

Bayesian Optimization Model-based Reinforcement Learning +3

Augmented World Models Facilitate Zero-Shot Dynamics Generalization From a Single Offline Environment

no code implementations ICLR Workshop SSL-RL 2021 Philip J. Ball, Cong Lu, Jack Parker-Holder, Stephen Roberts

Reinforcement learning from large-scale offline datasets provides us with the ability to learn policies without potentially unsafe or impractical exploration.

Zero-shot Generalization

Exploration in Approximate Hyper-State Space for Meta Reinforcement Learning

1 code implementation2 Oct 2020 Luisa Zintgraf, Leo Feng, Cong Lu, Maximilian Igl, Kristian Hartikainen, Katja Hofmann, Shimon Whiteson

To rapidly learn a new task, it is often essential for agents to explore efficiently -- especially when performance matters from the first timestep.

Meta-Learning Meta Reinforcement Learning +2

Cannot find the paper you are looking for? You can Submit a new open access paper.