You need to log in to edit.

You can create a new account if you don't have one.

Or, discuss a change on Slack.

You can create a new account if you don't have one.

Or, discuss a change on Slack.

no code implementations • 2 Mar 2023 • Jonas Rothfuss, Bhavya Sukhija, Tobias Birchler, Parnian Kassraie, Andreas Krause

We study the problem of conservative off-policy evaluation (COPE) where given an offline dataset of environment interactions, collected by other agents, we seek to obtain a (tight) lower bound on a policy's performance.

no code implementations • 14 Nov 2022 • Jonas Rothfuss, Martin Josifoski, Vincent Fortuin, Andreas Krause

Meta-Learning aims to accelerate the learning on new tasks by acquiring useful inductive biases from related data sources.

no code implementations • 2 Nov 2022 • Songyan Hou, Parnian Kassraie, Anastasis Kratsios, Jonas Rothfuss, Andreas Krause

Existing generalization bounds fail to explain crucial factors that drive generalization of modern neural networks.

no code implementations • 27 Oct 2022 • Felix Schur, Parnian Kassraie, Jonas Rothfuss, Andreas Krause

Our algorithm can be paired with any kernelized or linear bandit algorithm and guarantees oracle optimal performance, meaning that as more tasks are solved, the regret of LIBO on each task converges to the regret of the bandit algorithm with oracle knowledge of the true kernel.

no code implementations • 24 Oct 2022 • Krunoslav Lehman Pavasovic, Jonas Rothfuss, Andreas Krause

To circumvent these issues, we approach meta-learning through the lens of functional Bayesian neural network inference, which views the prior as a stochastic process and performs inference in the function space.

no code implementations • 3 Oct 2022 • Jonas Rothfuss, Christopher Koenig, Alisa Rupenyan, Andreas Krause

In the presence of unknown safety constraints, it is crucial to choose reliable model hyper-parameters to avoid safety violations.

1 code implementation • 3 Jun 2022 • Alexander Hägele, Jonas Rothfuss, Lars Lorch, Vignesh Ram Somnath, Bernhard Schölkopf, Andreas Krause

Inferring causal structures from experimentation is a central task in many domains.

1 code implementation • 25 May 2022 • Lars Lorch, Scott Sussex, Jonas Rothfuss, Andreas Krause, Bernhard Schölkopf

Rather than searching over structures, we train a variational inference model to directly predict the causal structure from observational or interventional data.

no code implementations • 1 Feb 2022 • Parnian Kassraie, Jonas Rothfuss, Andreas Krause

We demonstrate our approach on the kernelized bandit problem (a. k. a.~Bayesian optimization), where we establish regret bounds competitive with those given the true kernel.

1 code implementation • 14 Jun 2021 • Yashas Annadani, Jonas Rothfuss, Alexandre Lacoste, Nino Scherrer, Anirudh Goyal, Yoshua Bengio, Stefan Bauer

However, a crucial aspect to acting intelligently upon the knowledge about causal structure which has been inferred from finite data demands reasoning about its uncertainty.

no code implementations • NeurIPS 2021 • Jonas Rothfuss, Dominique Heyn, Jinfan Chen, Andreas Krause

When data are scarce meta-learning can improve a learner's accuracy by harnessing previous experience from related learning tasks.

2 code implementations • NeurIPS 2021 • Lars Lorch, Jonas Rothfuss, Bernhard Schölkopf, Andreas Krause

In this work, we propose a general, fully differentiable framework for Bayesian structure learning (DiBS) that operates in the continuous space of a latent probabilistic graph representation.

no code implementations • 10 Mar 2021 • Lorenz Kuhn, Clare Lyle, Aidan N. Gomez, Jonas Rothfuss, Yarin Gal

Existing generalization measures that aim to capture a model's simplicity based on parameter counts or norms fail to explain generalization in overparameterized deep neural networks.

no code implementations • 1 Jan 2021 • Jonas Rothfuss, Martin Josifoski, Andreas Krause

Bayesian deep learning is a promising approach towards improved uncertainty quantification and sample efficiency.

3 code implementations • ICML Workshop LifelongML 2020 • Jonas Rothfuss, Vincent Fortuin, Martin Josifoski, Andreas Krause

Meta-learning can successfully acquire useful inductive biases from data.

1 code implementation • 21 Jul 2019 • Jonas Rothfuss, Fabio Ferreira, Simon Boehm, Simon Walther, Maxim Ulrich, Tamim Asfour, Andreas Krause

To address this issue, we develop a model-agnostic noise regularization method for CDE that adds random perturbations to the data during training.

1 code implementation • 3 Mar 2019 • Jonas Rothfuss, Fabio Ferreira, Simon Walther, Maxim Ulrich

Given a set of empirical observations, conditional density estimation aims to capture the statistical relationship between a conditional variable $\mathbf{x}$ and a dependent variable $\mathbf{y}$ by modeling their conditional probability $p(\mathbf{y}|\mathbf{x})$.

6 code implementations • ICLR 2019 • Jonas Rothfuss, Dennis Lee, Ignasi Clavera, Tamim Asfour, Pieter Abbeel

Credit assignment in Meta-reinforcement learning (Meta-RL) is still poorly understood.

1 code implementation • 14 Sep 2018 • Ignasi Clavera, Jonas Rothfuss, John Schulman, Yasuhiro Fujita, Tamim Asfour, Pieter Abbeel

Finally, we demonstrate that our approach is able to match the asymptotic performance of model-free methods while requiring significantly less experience.

Model-based Reinforcement Learning
reinforcement-learning
**+1**

2 code implementations • 2 Jul 2018 • Fabio Ferreira, Jonas Rothfuss, Eren Erdal Aksoy, You Zhou, Tamim Asfour

We release two artificial datasets, Simulated Flying Shapes and Simulated Planar Manipulator that allow to test the learning ability of video processing systems.

1 code implementation • 12 Jan 2018 • Jonas Rothfuss, Fabio Ferreira, Eren Erdal Aksoy, You Zhou, Tamim Asfour

We present a novel deep neural network architecture for representing robot experiences in an episodic-like memory which facilitates encoding, recalling, and predicting action experiences.

Cannot find the paper you are looking for? You can
Submit a new open access paper.

Contact us on:
hello@paperswithcode.com
.
Papers With Code is a free resource with all data licensed under CC-BY-SA.