2 code implementations • NeurIPS 2019 • Lars Maaløe, Marco Fraccaro, Valentin Liévin, Ole Winther
In this paper we close the performance gap by constructing VAE models that can effectively utilize a deep hierarchy of stochastic variables and model complex covariance structures.
Ranked #18 on Image Generation on ImageNet 32x32 (bpd metric)
no code implementations • 28 Oct 2018 • Ulrich Paquet, Marco Fraccaro
This technical report presents pseudo-code for a Riemannian manifold Hamiltonian Monte Carlo (RMHMC) method to efficiently simulate samples from $N$-dimensional posterior distributions $p(x|y)$, where $x \in R^N$ is drawn from a Gaussian Process (GP) prior, and observations $y_n$ are independent given $x_n$.
no code implementations • ICML 2018 • Marco Fraccaro, Danilo Jimenez Rezende, Yori Zwols, Alexander Pritzel, S. M. Ali Eslami, Fabio Viola
In model-based reinforcement learning, generative and temporal models of environments can be leveraged to boost agent performance, either by tuning the agent's representations during training or via use as part of an explicit planning mechanism.
1 code implementation • NeurIPS 2017 • Marco Fraccaro, Simon Kamronn, Ulrich Paquet, Ole Winther
This paper takes a step towards temporal reasoning in a dynamically changing video, not in the pixel space that constitutes its frames, but in a latent space that describes the non-linear dynamics of the objects in its world.
no code implementations • 3 Apr 2017 • Lars Maaløe, Marco Fraccaro, Ole Winther
Deep generative models trained with large amounts of unlabelled data have proven to be powerful within the domain of unsupervised learning.
1 code implementation • NeurIPS 2016 • Marco Fraccaro, Søren Kaae Sønderby, Ulrich Paquet, Ole Winther
How can we efficiently propagate uncertainty in a latent state representation with recurrent neural networks?
no code implementations • 7 Apr 2016 • Marco Fraccaro, Ulrich Paquet, Ole Winther
The estimation of normalizing constants is a fundamental step in probabilistic model comparison.