Search Results for author: Ian Davies

Found 8 papers, 2 papers with code

One Step at a Time: Pros and Cons of Multi-Step Meta-Gradient Reinforcement Learning

no code implementations30 Oct 2021 Clément Bonnet, Paul Caron, Thomas Barrett, Ian Davies, Alexandre Laterre

Self-tuning algorithms that adapt the learning process online encourage more effective and robust learning.

Kernel Identification Through Transformers

1 code implementation NeurIPS 2021 Fergus Simpson, Ian Davies, Vidhi Lalchand, Alessandro Vullo, Nicolas Durrande, Carl Rasmussen

Kernel selection plays a central role in determining the performance of Gaussian Process (GP) models, as the chosen kernel determines both the inductive biases and prior support of functions under the GP prior.

regression

Learning to Safely Exploit a Non-Stationary Opponent

no code implementations NeurIPS 2021 Zheng Tian, Hang Ren, Yaodong Yang, Yuchen Sun, Ziqi Han, Ian Davies, Jun Wang

On the other hand, overfitting to an opponent (i. e., exploiting only one specific type of opponent) makes the learning player easily exploitable by others.

Learning to Model Opponent Learning

1 code implementation6 Jun 2020 Ian Davies, Zheng Tian, Jun Wang

In this work, we develop a novel approach to modelling an opponent's learning dynamics which we term Learning to Model Opponent Learning (LeMOL).

Decision Making Multi-agent Reinforcement Learning

Joint Perception and Control as Inference with an Object-based Implementation

no code implementations4 Mar 2019 Minne Li, Zheng Tian, Pranav Nashikkar, Ian Davies, Ying Wen, Jun Wang

Existing model-based reinforcement learning methods often study perception modeling and decision making separately.

Bayesian Inference Decision Making +2

Learning to Communicate Implicitly By Actions

no code implementations10 Oct 2018 Zheng Tian, Shihao Zou, Ian Davies, Tim Warr, Lisheng Wu, Haitham Bou Ammar, Jun Wang

The auxiliary reward for communication is integrated into the learning of the policy module.

Cannot find the paper you are looking for? You can Submit a new open access paper.