Search Results for author: Sebastian Curi

Found 15 papers, 6 papers with code

Get Back Here: Robust Imitation by Return-to-Distribution Planning

no code implementations2 May 2023 Geoffrey Cideron, Baruch Tabanpour, Sebastian Curi, Sertan Girgin, Leonard Hussenot, Gabriel Dulac-Arnold, Matthieu Geist, Olivier Pietquin, Robert Dadashi

We consider the Imitation Learning (IL) setup where expert data are not collected on the actual deployment environment but on a different version.

Imitation Learning

Gradient-Based Trajectory Optimization With Learned Dynamics

no code implementations9 Apr 2022 Bhavya Sukhija, Nathanael Köhler, Miguel Zamora, Simon Zimmermann, Sebastian Curi, Andreas Krause, Stelian Coros

In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car, and gives good performance in combination with trajectory optimization methods.

Constrained Policy Optimization via Bayesian World Models

1 code implementation ICLR 2022 Yarden As, Ilnura Usmanova, Sebastian Curi, Andreas Krause

Improving sample-efficiency and safety are crucial challenges when deploying reinforcement learning in high-stakes real world applications.

reinforcement-learning Reinforcement Learning (RL)

Risk-Averse Offline Reinforcement Learning

1 code implementation ICLR 2021 Núria Armengol Urpí, Sebastian Curi, Andreas Krause

We demonstrate empirically that in the presence of natural distribution-shifts, O-RAAC learns policies with good average performance.

reinforcement-learning Reinforcement Learning +1

Logistic Q-Learning

no code implementations21 Oct 2020 Joan Bas-Serrano, Sebastian Curi, Andreas Krause, Gergely Neu

We propose a new reinforcement learning algorithm derived from a regularized linear-programming formulation of optimal control in MDPs.

Q-Learning Reinforcement Learning (RL)

Efficient Model-Based Reinforcement Learning through Optimistic Policy Search and Planning

1 code implementation NeurIPS 2020 Sebastian Curi, Felix Berkenkamp, Andreas Krause

Based on this theoretical foundation, we show how optimistic exploration can be easily combined with state-of-the-art reinforcement learning algorithms and different probabilistic models.

Model-based Reinforcement Learning reinforcement-learning +2

Adaptive Sampling for Stochastic Risk-Averse Learning

1 code implementation NeurIPS 2020 Sebastian Curi, Kfir. Y. Levy, Stefanie Jegelka, Andreas Krause

In high-stakes machine learning applications, it is crucial to not only perform well on average, but also when restricted to difficult examples.

Point Processes

Structured Variational Inference in Unstable Gaussian Process State Space Models

1 code implementation16 Jul 2019 Silvan Melchior, Sebastian Curi, Felix Berkenkamp, Andreas Krause

Finally, we show experimentally that our learning algorithm performs well in stable and unstable real systems with hidden states.

Gaussian Processes State Space Models +1

Safe Contextual Bayesian Optimization for Sustainable Room Temperature PID Control Tuning

no code implementations28 Jun 2019 Marcello Fiducioso, Sebastian Curi, Benedikt Schumacher, Markus Gwerder, Andreas Krause

Furthermore, this successful attempt paves the way for further use at different levels of HVAC systems, with promising energy, operational, and commissioning costs savings, and it is a practical demonstration of the positive effects that Artificial Intelligence can have on environmental sustainability.

Bayesian Optimization

Online Variance Reduction with Mixtures

1 code implementation29 Mar 2019 Zalán Borsos, Sebastian Curi, Kfir. Y. Levy, Andreas Krause

Adaptive importance sampling for stochastic optimization is a promising approach that offers improved convergence through variance reduction.

Stochastic Optimization

Adaptive Input Estimation in Linear Dynamical Systems with Applications to Learning-from-Observations

no code implementations19 Jun 2018 Sebastian Curi, Kfir. Y. Levy, Andreas Krause

To this end, we introduce a novel estimation algorithm that explicitly trades off bias and variance to optimally reduce the overall estimation error.

Imitation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.