Search Results for author: Tim Genewein

Found 24 papers, 8 papers with code

A Nonparametric Conjugate Prior Distribution for the Maximizing Argument of a Noisy Function

no code implementations NeurIPS 2012 Pedro Ortega, Jordi Grau-Moya, Tim Genewein, David Balduzzi, Daniel Braun

We propose a novel Bayesian approach to solve stochastic optimization problems that involve finding extrema of noisy, nonlinear functions.

Stochastic Optimization

Abstraction in decision-makers with limited information processing capabilities

no code implementations16 Dec 2013 Tim Genewein, Daniel A. Braun

A distinctive property of human and animal intelligence is the ability to form abstractions by neglecting irrelevant information which allows to separate structure from noise.

Decision Making

Planning with Information-Processing Constraints and Model Uncertainty in Markov Decision Processes

no code implementations7 Apr 2016 Jordi Grau-Moya, Felix Leibfried, Tim Genewein, Daniel A. Braun

As limit cases, this generalized scheme includes standard value iteration with a known model, Bayesian MDP planning, and robust planning.

On Detecting Adversarial Perturbations

1 code implementation14 Feb 2017 Jan Hendrik Metzen, Tim Genewein, Volker Fischer, Bastian Bischoff

In this work, we propose to augment deep neural networks with a small "detector" subnetwork which is trained on the binary classification task of distinguishing genuine data from data containing adversarial perturbations.

Adversarial Attack Binary Classification +1

Variational Network Quantization

no code implementations ICLR 2018 Jan Achterhold, Jan Mathias Koehler, Anke Schmeink, Tim Genewein

In this paper, the preparation of a neural network for pruning and few-bit quantization is formulated as a variational inference problem.

Quantization Variational Inference

An information-theoretic on-line update principle for perception-action coupling

no code implementations16 Apr 2018 Zhen Peng, Tim Genewein, Felix Leibfried, Daniel A. Braun

Here we consider perception and action as two serial information channels with limited information-processing capacity.

The Power of Ensembles for Active Learning in Image Classification

no code implementations CVPR 2018 William H. Beluch, Tim Genewein, Andreas Nürnberger, Jan M. Köhler

To investigate why Monte-Carlo Dropout uncertainties perform worse, we explore potential differences in isolation in a series of experiments.

Active Learning General Classification +1

Sinkhorn AutoEncoders

2 code implementations ICLR 2019 Giorgio Patrini, Rianne van den Berg, Patrick Forré, Marcello Carioni, Samarth Bhargav, Max Welling, Tim Genewein, Frank Nielsen

We show that minimizing the p-Wasserstein distance between the generator and the true data distribution is equivalent to the unconstrained min-min optimization of the p-Wasserstein distance between the encoder aggregated posterior and the prior in latent space, plus a reconstruction error.

Probabilistic Programming

Group Pruning using a Bounded-Lp norm for Group Gating and Regularization

no code implementations9 Aug 2019 Chaithanya Kumar Mummadi, Tim Genewein, Dan Zhang, Thomas Brox, Volker Fischer

We achieve state-of-the-art pruning results for ResNet-50 with higher accuracy on ImageNet.

Meta-trained agents implement Bayes-optimal agents

no code implementations NeurIPS 2020 Vladimir Mikulik, Grégoire Delétang, Tom McGrath, Tim Genewein, Miljan Martic, Shane Legg, Pedro A. Ortega

Memory-based meta-learning is a powerful technique to build agents that adapt fast to any task within a target distribution.

Meta-Learning

Stochastic Approximation of Gaussian Free Energy for Risk-Sensitive Reinforcement Learning

no code implementations NeurIPS 2021 Grégoire Delétang, Jordi Grau-Moya, Markus Kunesch, Tim Genewein, Rob Brekelmans, Shane Legg, Pedro A Ortega

Since the Gaussian free energy is known to be a certainty-equivalent sensitive to the mean and the variance, the learning rule has applications in risk-sensitive decision-making.

Decision Making reinforcement-learning +1

Model-Free Risk-Sensitive Reinforcement Learning

no code implementations4 Nov 2021 Grégoire Delétang, Jordi Grau-Moya, Markus Kunesch, Tim Genewein, Rob Brekelmans, Shane Legg, Pedro A. Ortega

Since the Gaussian free energy is known to be a certainty-equivalent sensitive to the mean and the variance, the learning rule has applications in risk-sensitive decision-making.

Decision Making reinforcement-learning +1

Your Policy Regularizer is Secretly an Adversary

no code implementations23 Mar 2022 Rob Brekelmans, Tim Genewein, Jordi Grau-Moya, Grégoire Delétang, Markus Kunesch, Shane Legg, Pedro Ortega

Policy regularization methods such as maximum entropy regularization are widely used in reinforcement learning to improve the robustness of a learned policy.

Beyond Bayes-optimality: meta-learning what you know you don't know

no code implementations30 Sep 2022 Jordi Grau-Moya, Grégoire Delétang, Markus Kunesch, Tim Genewein, Elliot Catt, Kevin Li, Anian Ruoss, Chris Cundy, Joel Veness, Jane Wang, Marcus Hutter, Christopher Summerfield, Shane Legg, Pedro Ortega

This is in contrast to risk-sensitive agents, which additionally exploit the higher-order moments of the return, and ambiguity-sensitive agents, which act differently when recognizing situations in which they lack knowledge.

Decision Making Meta-Learning

Language Modeling Is Compression

1 code implementation19 Sep 2023 Grégoire Delétang, Anian Ruoss, Paul-Ambroise Duquenne, Elliot Catt, Tim Genewein, Christopher Mattern, Jordi Grau-Moya, Li Kevin Wenliang, Matthew Aitchison, Laurent Orseau, Marcus Hutter, Joel Veness

We show that large language models are powerful general-purpose predictors and that the compression viewpoint provides novel insights into scaling laws, tokenization, and in-context learning.

In-Context Learning Language Modelling

Learning Universal Predictors

1 code implementation26 Jan 2024 Jordi Grau-Moya, Tim Genewein, Marcus Hutter, Laurent Orseau, Grégoire Delétang, Elliot Catt, Anian Ruoss, Li Kevin Wenliang, Christopher Mattern, Matthew Aitchison, Joel Veness

Meta-learning has emerged as a powerful approach to train neural networks to learn new tasks quickly from limited data.

Meta-Learning

Grandmaster-Level Chess Without Search

no code implementations7 Feb 2024 Anian Ruoss, Grégoire Delétang, Sourabh Medapati, Jordi Grau-Moya, Li Kevin Wenliang, Elliot Catt, John Reid, Tim Genewein

Unlike traditional chess engines that rely on complex heuristics, explicit search, or a combination of both, we train a 270M parameter transformer model with supervised learning on a dataset of 10 million chess games.

Cannot find the paper you are looking for? You can Submit a new open access paper.