Search Results for author: Trevor McInroe

Found 6 papers, 4 papers with code

LLM-Personalize: Aligning LLM Planners with Human Preferences via Reinforced Self-Training for Housekeeping Robots

no code implementations22 Apr 2024 Dongge Han, Trevor McInroe, Adam Jelley, Stefano V. Albrecht, Peter Bell, Amos Storkey

We introduce LLM-Personalize, a novel framework with an optimization pipeline designed to personalize LLM planners for household robotics.

Planning to Go Out-of-Distribution in Offline-to-Online Reinforcement Learning

no code implementations9 Oct 2023 Trevor McInroe, Adam Jelley, Stefano V. Albrecht, Amos Storkey

Offline pretraining with a static dataset followed by online fine-tuning (offline-to-online, or OtO) is a paradigm well matched to a real-world RL deployment process.

Continuous Control Offline RL

Temporal Disentanglement of Representations for Improved Generalisation in Reinforcement Learning

1 code implementation12 Jul 2022 Mhairi Dunion, Trevor McInroe, Kevin Sebastian Luck, Josiah P. Hanna, Stefano V. Albrecht

Reinforcement Learning (RL) agents are often unable to generalise well to environment variations in the state space that were not observed during training.

Disentanglement reinforcement-learning +1

Multi-Horizon Representations with Hierarchical Forward Models for Reinforcement Learning

1 code implementation22 Jun 2022 Trevor McInroe, Lukas Schäfer, Stefano V. Albrecht

Learning control from pixels is difficult for reinforcement learning (RL) agents because representation learning and policy learning are intertwined.

reinforcement-learning Reinforcement Learning (RL) +1

Learning Temporally-Consistent Representations for Data-Efficient Reinforcement Learning

2 code implementations11 Oct 2021 Trevor McInroe, Lukas Schäfer, Stefano V. Albrecht

Deep reinforcement learning (RL) agents that exist in high-dimensional state spaces, such as those composed of images, have interconnected learning burdens.

reinforcement-learning Reinforcement Learning (RL) +1

Cannot find the paper you are looking for? You can Submit a new open access paper.