no code implementations • 11 Sep 2023 • Marin Vlastelica, Sebastian Blaes, Cristina Pineri, Georg Martius
We introduce a simple but effective method for managing risk in model-based reinforcement learning with trajectory sampling that involves probabilistic safety constraints and balancing of optimism in the face of epistemic uncertainty and pessimism in the face of aleatoric uncertainty of an ensemble of stochastic neural networks. Various experiments indicate that the separation of uncertainties is essential to performing well with data-driven MPC approaches in uncertain and safety-critical control environments.
Model-based Reinforcement Learning reinforcement-learning +1
no code implementations • 15 Aug 2023 • Nico Gürtler, Felix Widmaier, Cansu Sancaktar, Sebastian Blaes, Pavel Kolev, Stefan Bauer, Manuel Wüthrich, Markus Wulfmeier, Martin Riedmiller, Arthur Allshire, Qiang Wang, Robert McCarthy, Hangyeol Kim, Jongchan Baek, Wookyong Kwon, Shanliang Qian, Yasunori Toshimitsu, Mike Yan Michelis, Amirhossein Kazemipour, Arman Raayatsanati, Hehui Zheng, Barnabas Gavin Cangan, Bernhard Schölkopf, Georg Martius
For this reason, a large part of the reinforcement learning (RL) community uses simulators to develop and benchmark algorithms.
2 code implementations • 28 Jul 2023 • Nico Gürtler, Sebastian Blaes, Pavel Kolev, Felix Widmaier, Manuel Wüthrich, Stefan Bauer, Bernhard Schölkopf, Georg Martius
To coordinate the efforts of the research community toward tackling this problem, we propose a benchmark including: i) a large collection of data for offline learning from a dexterous manipulation platform on two tasks, obtained with capable RL agents trained in simulation; ii) the option to execute learned policies on a real-world robotic system and a simulation for efficient debugging.
no code implementations • 16 Sep 2022 • Chenhao Li, Sebastian Blaes, Pavel Kolev, Marin Vlastelica, Jonas Frey, Georg Martius
Learning diverse skills is one of the main challenges in robotics.
no code implementations • 23 Jun 2022 • Chenhao Li, Marin Vlastelica, Sebastian Blaes, Jonas Frey, Felix Grimminger, Georg Martius
Learning agile skills is one of the main challenges in robotics.
no code implementations • 22 Jun 2022 • Cansu Sancaktar, Sebastian Blaes, Georg Martius
It has been a long-standing dream to design artificial agents that explore their environment efficiently via intrinsic motivation, similar to how children perform curious free play.
no code implementations • ICLR 2021 • Cristina Pinneri, Shambhuraj Sawant, Sebastian Blaes, Georg Martius
Solving high-dimensional, continuous robotic tasks is a challenging optimization problem.
1 code implementation • 14 Aug 2020 • Cristina Pinneri, Shambhuraj Sawant, Sebastian Blaes, Jan Achterhold, Joerg Stueckler, Michal Rolinek, Georg Martius
However, their sampling inefficiency prevents them from being used for real-time planning and control.
Model-based Reinforcement Learning reinforcement-learning +1
1 code implementation • NeurIPS 2019 • Sebastian Blaes, Marin Vlastelica Pogančić, Jia-Jie Zhu, Georg Martius
We present a novel intrinsically motivated agent that learns how to control the environment in the fastest possible manner by optimizing learning progress.