1 code implementation • 7 Mar 2023 • Daniel Palenicek, Michael Lutter, Joao Carvalho, Jan Peters
Therefore, we conclude that the limitation of model-based value expansion methods is not the model accuracy of the learned models.
no code implementations • 28 Mar 2022 • Daniel Palenicek, Michael Lutter, Jan Peters
Model-based value expansion methods promise to improve the quality of value function targets and, thereby, the effectiveness of value function learning.
1 code implementation • 5 Oct 2021 • Michael Lutter, Boris Belousov, Shie Mannor, Dieter Fox, Animesh Garg, Jan Peters
Especially for continuous control, solving this differential equation and its extension the Hamilton-Jacobi-Isaacs equation, is important as it yields the optimal policy that achieves the maximum reward on a give task.
1 code implementation • 5 Oct 2021 • Michael Lutter, Jan Peters
Especially for learning dynamics models, these black-box models are not desirable as the underlying principles are well understood and the standard deep networks can learn dynamics that violate these principles.
no code implementations • 29 Sep 2021 • Michael Lutter, Leonard Hasenclever, Arunkumar Byravan, Gabriel Dulac-Arnold, Piotr Trochim, Nicolas Heess, Josh Merel, Yuval Tassa
This paper sets out to disambiguate the role of different design choices for learning dynamics models, by comparing their performance to planning with a ground-truth model -- the simulator.
1 code implementation • 25 May 2021 • Michael Lutter, Shie Mannor, Jan Peters, Dieter Fox, Animesh Garg
The adversarial perturbations encourage a optimal policy that is robust to changes in the dynamics.
1 code implementation • 10 May 2021 • Michael Lutter, Shie Mannor, Jan Peters, Dieter Fox, Animesh Garg
This algorithm enables dynamic programming for continuous states and actions with a known dynamics model.
no code implementations • 3 Nov 2020 • Michael Lutter, Johannes Silberbauer, Joe Watson, Jan Peters
A limitation of model-based reinforcement learning (MBRL) is the exploitation of errors in the learned models.
Model-based Reinforcement Learning reinforcement-learning +1
no code implementations • 26 Oct 2020 • Kai Ploeger, Michael Lutter, Jan Peters
Robots that can learn in the physical world will be important to en-able robots to escape their stiff and pre-programmed movements.
no code implementations • 19 Oct 2020 • Michael Lutter, Johannes Silberbauer, Joe Watson, Jan Peters
In this work, we examine a spectrum of hybrid model for the domain of multi-body robot dynamics.
no code implementations • ICLR Workshop DeepDiffEq 2019 • Michael Lutter, Jan Peters
Therefore, differential equations are a promising approach to incorporate prior knowledge in machine learning models to obtain robust and interpretable models.
no code implementations • 13 Sep 2019 • Michael Lutter, Boris Belousov, Kim Listmann, Debora Clever, Jan Peters
The corresponding optimal value function is learned end-to-end by embedding a deep differential network in the Hamilton-Jacobi-Bellmann differential equation and minimizing the error of this equality while simultaneously decreasing the discounting from short- to far-sighted to enable the learning.
1 code implementation • 10 Jul 2019 • Michael Lutter, Kim Listmann, Jan Peters
Applying Deep Learning to control has a lot of potential for enabling the intelligent design of robot control laws.
3 code implementations • ICLR 2019 • Michael Lutter, Christian Ritter, Jan Peters
DeLaN can learn the equations of motion of a mechanical system (i. e., system dynamics) with a deep network efficiently while ensuring physical plausibility.