no code implementations • 2 Mar 2023 • Jonas Rothfuss, Bhavya Sukhija, Tobias Birchler, Parnian Kassraie, Andreas Krause
We study the problem of conservative off-policy evaluation (COPE) where given an offline dataset of environment interactions, collected by other agents, we seek to obtain a (tight) lower bound on a policy's performance.
no code implementations • 14 Nov 2022 • Jonas Rothfuss, Martin Josifoski, Vincent Fortuin, Andreas Krause
Meta-Learning aims to accelerate the learning on new tasks by acquiring useful inductive biases from related data sources.
no code implementations • 2 Nov 2022 • Songyan Hou, Parnian Kassraie, Anastasis Kratsios, Jonas Rothfuss, Andreas Krause
Existing generalization bounds fail to explain crucial factors that drive generalization of modern neural networks.
no code implementations • 27 Oct 2022 • Felix Schur, Parnian Kassraie, Jonas Rothfuss, Andreas Krause
Our algorithm can be paired with any kernelized or linear bandit algorithm and guarantees oracle optimal performance, meaning that as more tasks are solved, the regret of LIBO on each task converges to the regret of the bandit algorithm with oracle knowledge of the true kernel.
no code implementations • 24 Oct 2022 • Krunoslav Lehman Pavasovic, Jonas Rothfuss, Andreas Krause
To circumvent these issues, we approach meta-learning through the lens of functional Bayesian neural network inference, which views the prior as a stochastic process and performs inference in the function space.
no code implementations • 3 Oct 2022 • Jonas Rothfuss, Christopher Koenig, Alisa Rupenyan, Andreas Krause
In the presence of unknown safety constraints, it is crucial to choose reliable model hyper-parameters to avoid safety violations.
1 code implementation • 3 Jun 2022 • Alexander Hägele, Jonas Rothfuss, Lars Lorch, Vignesh Ram Somnath, Bernhard Schölkopf, Andreas Krause
Inferring causal structures from experimentation is a central task in many domains.
1 code implementation • 25 May 2022 • Lars Lorch, Scott Sussex, Jonas Rothfuss, Andreas Krause, Bernhard Schölkopf
Rather than searching over structures, we train a variational inference model to directly predict the causal structure from observational or interventional data.
no code implementations • 1 Feb 2022 • Parnian Kassraie, Jonas Rothfuss, Andreas Krause
We demonstrate our approach on the kernelized bandit problem (a. k. a.~Bayesian optimization), where we establish regret bounds competitive with those given the true kernel.
1 code implementation • 14 Jun 2021 • Yashas Annadani, Jonas Rothfuss, Alexandre Lacoste, Nino Scherrer, Anirudh Goyal, Yoshua Bengio, Stefan Bauer
However, a crucial aspect to acting intelligently upon the knowledge about causal structure which has been inferred from finite data demands reasoning about its uncertainty.
no code implementations • NeurIPS 2021 • Jonas Rothfuss, Dominique Heyn, Jinfan Chen, Andreas Krause
When data are scarce meta-learning can improve a learner's accuracy by harnessing previous experience from related learning tasks.
2 code implementations • NeurIPS 2021 • Lars Lorch, Jonas Rothfuss, Bernhard Schölkopf, Andreas Krause
In this work, we propose a general, fully differentiable framework for Bayesian structure learning (DiBS) that operates in the continuous space of a latent probabilistic graph representation.
no code implementations • 10 Mar 2021 • Lorenz Kuhn, Clare Lyle, Aidan N. Gomez, Jonas Rothfuss, Yarin Gal
Existing generalization measures that aim to capture a model's simplicity based on parameter counts or norms fail to explain generalization in overparameterized deep neural networks.
no code implementations • 1 Jan 2021 • Jonas Rothfuss, Martin Josifoski, Andreas Krause
Bayesian deep learning is a promising approach towards improved uncertainty quantification and sample efficiency.
3 code implementations • ICML Workshop LifelongML 2020 • Jonas Rothfuss, Vincent Fortuin, Martin Josifoski, Andreas Krause
Meta-learning can successfully acquire useful inductive biases from data.
1 code implementation • 21 Jul 2019 • Jonas Rothfuss, Fabio Ferreira, Simon Boehm, Simon Walther, Maxim Ulrich, Tamim Asfour, Andreas Krause
To address this issue, we develop a model-agnostic noise regularization method for CDE that adds random perturbations to the data during training.
1 code implementation • 3 Mar 2019 • Jonas Rothfuss, Fabio Ferreira, Simon Walther, Maxim Ulrich
Given a set of empirical observations, conditional density estimation aims to capture the statistical relationship between a conditional variable $\mathbf{x}$ and a dependent variable $\mathbf{y}$ by modeling their conditional probability $p(\mathbf{y}|\mathbf{x})$.
6 code implementations • ICLR 2019 • Jonas Rothfuss, Dennis Lee, Ignasi Clavera, Tamim Asfour, Pieter Abbeel
Credit assignment in Meta-reinforcement learning (Meta-RL) is still poorly understood.
1 code implementation • 14 Sep 2018 • Ignasi Clavera, Jonas Rothfuss, John Schulman, Yasuhiro Fujita, Tamim Asfour, Pieter Abbeel
Finally, we demonstrate that our approach is able to match the asymptotic performance of model-free methods while requiring significantly less experience.
Model-based Reinforcement Learning
reinforcement-learning
+1
2 code implementations • 2 Jul 2018 • Fabio Ferreira, Jonas Rothfuss, Eren Erdal Aksoy, You Zhou, Tamim Asfour
We release two artificial datasets, Simulated Flying Shapes and Simulated Planar Manipulator that allow to test the learning ability of video processing systems.
1 code implementation • 12 Jan 2018 • Jonas Rothfuss, Fabio Ferreira, Eren Erdal Aksoy, You Zhou, Tamim Asfour
We present a novel deep neural network architecture for representing robot experiences in an episodic-like memory which facilitates encoding, recalling, and predicting action experiences.