no code implementations • 4 Sep 2022 • Felix Leibfried
And second, in model-based RL where agents aim to learn about the environment they are operating in, the model-learning part can be naturally phrased as an inference problem over the process that governs environment dynamics.
1 code implementation • 12 Apr 2021 • Vincent Dutordoir, Hugh Salimbeni, Eric Hambro, John McLeod, Felix Leibfried, Artem Artemev, Mark van der Wilk, James Hensman, Marc P. Deisenroth, ST John
GPflux is compatible with and built on top of the Keras deep learning eco-system.
2 code implementations • 26 Mar 2021 • John McLeod, Hrvoje Stojic, Vincent Adam, Dongho Kim, Jordi Grau-Moya, Peter Vrancx, Felix Leibfried
This paves the way for new research directions, e. g. investigating uncertainty-aware environment models that are not necessarily neural-network-based, or developing algorithms to solve industrially-motivated benchmarks that share characteristics with real-world problems.
Model-based Reinforcement Learning
reinforcement-learning
+2
no code implementations • 27 Dec 2020 • Felix Leibfried, Vincent Dutordoir, ST John, Nicolas Durrande
In this context, a convenient choice for approximate inference is variational inference (VI), where the problem of Bayesian inference is cast as an optimization problem -- namely, to maximize a lower bound of the log marginal likelihood.
no code implementations • 11 Sep 2019 • Felix Leibfried, Jordi Grau-Moya
While this has been initially proposed for Markov Decision Processes (MDPs) in tabular settings, it was recently shown that a similar principle leads to significant improvements over vanilla SQL in RL for high-dimensional domains with discrete actions and function approximators.
no code implementations • NeurIPS 2019 • Felix Leibfried, Sergio Pascual-Diaz, Jordi Grau-Moya
In this paper, we investigate the use of empowerment in the presence of an extrinsic reward signal.
no code implementations • ICLR 2019 • Jordi Grau-Moya, Felix Leibfried, Peter Vrancx
We show that the prior optimization introduces a mutual-information regularizer in the RL objective.
2 code implementations • 12 Oct 2018 • Tim Pearce, Felix Leibfried, Alexandra Brintrup, Mohamed Zaki, Andy Neely
Ensembling NNs provides an easily implementable, scalable method for uncertainty quantification, however, it has been criticised for not being Bayesian.
no code implementations • 6 Sep 2018 • Felix Leibfried, Peter Vrancx
This paper proposes a new optimization objective for value-based deep reinforcement learning.
no code implementations • 16 Apr 2018 • Zhen Peng, Tim Genewein, Felix Leibfried, Daniel A. Braun
Here we consider perception and action as two serial information channels with limited information-processing capacity.
no code implementations • 9 Feb 2018 • Jordi Grau-Moya, Felix Leibfried, Haitham Bou-Ammar
Within the context of video games the notion of perfectly rational agents can be undesirable as it leads to uninteresting situations, where humans face tough adversarial decision makers.
no code implementations • 6 Aug 2017 • Felix Leibfried, Jordi Grau-Moya, Haitham Bou-Ammar
Different learning outcomes can be demonstrated by tuning a Lagrange multiplier accordingly.
no code implementations • 21 Nov 2016 • Felix Leibfried, Nate Kushman, Katja Hofmann
Reinforcement learning is concerned with identifying reward-maximizing behaviour policies in environments that are initially unknown.
no code implementations • 7 Apr 2016 • Jordi Grau-Moya, Felix Leibfried, Tim Genewein, Daniel A. Braun
As limit cases, this generalized scheme includes standard value iteration with a known model, Bayesian MDP planning, and robust planning.
no code implementations • 26 Feb 2016 • Felix Leibfried, Daniel Alexander Braun
Bounded rational decision-makers transform sensory input into motor output under limited computational resources.