no code implementations • 2 May 2023 • Geoffrey Cideron, Baruch Tabanpour, Sebastian Curi, Sertan Girgin, Leonard Hussenot, Gabriel Dulac-Arnold, Matthieu Geist, Olivier Pietquin, Robert Dadashi
We consider the Imitation Learning (IL) setup where expert data are not collected on the actual deployment environment but on a different version.
no code implementations • 4 Jul 2022 • Sebastian Curi, Armin Lederer, Sandra Hirche, Andreas Krause
Ensuring safety is a crucial challenge when deploying reinforcement learning (RL) to real-world systems.
no code implementations • 9 Apr 2022 • Bhavya Sukhija, Nathanael Köhler, Miguel Zamora, Simon Zimmermann, Sebastian Curi, Andreas Krause, Stelian Coros
In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car, and gives good performance in combination with trajectory optimization methods.
1 code implementation • ICLR 2022 • Yarden As, Ilnura Usmanova, Sebastian Curi, Andreas Krause
Improving sample-efficiency and safety are crucial challenges when deploying reinforcement learning in high-stakes real world applications.
no code implementations • 18 Mar 2021 • Sebastian Curi, Ilija Bogunovic, Andreas Krause
In real-world tasks, reinforcement learning (RL) agents frequently encounter situations that are not present during training time.
Deep Reinforcement Learning Model-based Reinforcement Learning +2
1 code implementation • ICLR 2021 • Núria Armengol Urpí, Sebastian Curi, Andreas Krause
We demonstrate empirically that in the presence of natural distribution-shifts, O-RAAC learns policies with good average performance.
no code implementations • 21 Oct 2020 • Joan Bas-Serrano, Sebastian Curi, Andreas Krause, Gergely Neu
We propose a new reinforcement learning algorithm derived from a regularized linear-programming formulation of optimal control in MDPs.
no code implementations • 19 Jun 2020 • Lenart Treven, Sebastian Curi, Mojmir Mutny, Andreas Krause
The principal task to control dynamical systems is to ensure their stability.
1 code implementation • NeurIPS 2020 • Sebastian Curi, Felix Berkenkamp, Andreas Krause
Based on this theoretical foundation, we show how optimistic exploration can be easily combined with state-of-the-art reinforcement learning algorithms and different probabilistic models.
Model-based Reinforcement Learning reinforcement-learning +2
no code implementations • L4DC 2020 • Sebastian Curi, Silvan Melchior, Felix Berkenkamp, Andreas Krause
Finally, we show experimentally that our learning algorithm performs well in stable and unstable real systems with hidden states.
1 code implementation • NeurIPS 2020 • Sebastian Curi, Kfir. Y. Levy, Stefanie Jegelka, Andreas Krause
In high-stakes machine learning applications, it is crucial to not only perform well on average, but also when restricted to difficult examples.
1 code implementation • 16 Jul 2019 • Silvan Melchior, Sebastian Curi, Felix Berkenkamp, Andreas Krause
Finally, we show experimentally that our learning algorithm performs well in stable and unstable real systems with hidden states.
no code implementations • 28 Jun 2019 • Marcello Fiducioso, Sebastian Curi, Benedikt Schumacher, Markus Gwerder, Andreas Krause
Furthermore, this successful attempt paves the way for further use at different levels of HVAC systems, with promising energy, operational, and commissioning costs savings, and it is a practical demonstration of the positive effects that Artificial Intelligence can have on environmental sustainability.
1 code implementation • 29 Mar 2019 • Zalán Borsos, Sebastian Curi, Kfir. Y. Levy, Andreas Krause
Adaptive importance sampling for stochastic optimization is a promising approach that offers improved convergence through variance reduction.
no code implementations • 19 Jun 2018 • Sebastian Curi, Kfir. Y. Levy, Andreas Krause
To this end, we introduce a novel estimation algorithm that explicitly trades off bias and variance to optimally reduce the overall estimation error.