no code implementations • 12 Oct 2024 • Yarden As, Bhavya Sukhija, Lenart Treven, Carmelo Sferrazza, Stelian Coros, Andreas Krause
Under regularity assumptions on the constraints and dynamics, we show that ActSafe guarantees safety during learning while also obtaining a near-optimal policy in finite time.
1 code implementation • ICML Workshop on Aligning Reinforcement Learning Experimentalists and Theorists 2024 • Jonas Hübotter, Bhavya Sukhija, Lenart Treven, Yarden As, Andreas Krause
We analyze Safe BO under the lens of a generalization of active learning with concrete prediction targets where sampling is restricted to an accessible region of the domain, while prediction targets may lie outside this region.
no code implementations • 3 Jun 2024 • Bhavya Sukhija, Lenart Treven, Florian Dörfler, Stelian Coros, Andreas Krause
We study the problem of nonepisodic reinforcement learning (RL) for nonlinear dynamical systems, where the system dynamics are unknown and the RL agent has to learn from a single trajectory, i. e., without resets.
1 code implementation • 3 Jun 2024 • Lenart Treven, Bhavya Sukhija, Yarden As, Florian Dörfler, Andreas Krause
Finally, we propose OTaCoS, an efficient model-based algorithm for our setting.
no code implementations • 9 May 2024 • Yarden As, Bhavya Sukhija, Andreas Krause
A major challenge in deploying reinforcement learning in online tasks is ensuring that safety is maintained throughout the learning process.
no code implementations • 25 Mar 2024 • Jonas Rothfuss, Bhavya Sukhija, Lenart Treven, Florian Dörfler, Stelian Coros, Andreas Krause
We present SIM-FSVGD for learning robot dynamics from data.
no code implementations • 13 Feb 2024 • Jonas Hübotter, Bhavya Sukhija, Lenart Treven, Yarden As, Andreas Krause
We study the question: How can we select the right data for fine-tuning to a specific task?
2 code implementations • 13 Feb 2024 • Jonas Hübotter, Bhavya Sukhija, Lenart Treven, Yarden As, Andreas Krause
We study a generalization of classical active learning to real-world settings with concrete prediction targets where sampling is restricted to an accessible region of the domain, while prediction targets may lie outside this region.
no code implementations • 13 Nov 2023 • Arjun Bhardwaj, Jonas Rothfuss, Bhavya Sukhija, Yarden As, Marco Hutter, Stelian Coros, Andreas Krause
We introduce PACOH-RL, a novel model-based Meta-Reinforcement Learning (Meta-RL) algorithm designed to efficiently adapt control policies to changing dynamics.
1 code implementation • 12 Jun 2023 • Daniel Widmer, Dongho Kang, Bhavya Sukhija, Jonas Hübotter, Andreas Krause, Stelian Coros
This paper presents a data-driven strategy to streamline the deployment of model-based controllers in legged robotic hardware platforms.
1 code implementation • 2 Mar 2023 • Jonas Rothfuss, Bhavya Sukhija, Tobias Birchler, Parnian Kassraie, Andreas Krause
We study the problem of conservative off-policy evaluation (COPE) where given an offline dataset of environment interactions, collected by other agents, we seek to obtain a (tight) lower bound on a policy's performance.
no code implementations • 9 Apr 2022 • Bhavya Sukhija, Nathanael Köhler, Miguel Zamora, Simon Zimmermann, Sebastian Curi, Andreas Krause, Stelian Coros
In our hardware experiments, we demonstrate that our learned model can represent complex dynamics for both the Spot and Radio-controlled (RC) car, and gives good performance in combination with trajectory optimization methods.
1 code implementation • 24 Jan 2022 • Bhavya Sukhija, Matteo Turchetta, David Lindner, Andreas Krause, Sebastian Trimpe, Dominik Baumann
Learning optimal control policies directly on physical systems is challenging since even a single failure can lead to costly hardware damage.