Search Results for author: Will Grathwohl

Found 14 papers, 8 papers with code

Continuous diffusion for categorical data

no code implementations28 Nov 2022 Sander Dieleman, Laurent Sartran, Arman Roshannai, Nikolay Savinov, Yaroslav Ganin, Pierre H. Richemond, Arnaud Doucet, Robin Strudel, Chris Dyer, Conor Durkan, Curtis Hawthorne, Rémi Leblond, Will Grathwohl, Jonas Adler

Diffusion models have quickly become the go-to paradigm for generative modelling of perceptual signals (such as images and sound) through iterative refinement.

Language Modelling

Score-Based Diffusion meets Annealed Importance Sampling

no code implementations16 Aug 2022 Arnaud Doucet, Will Grathwohl, Alexander G. D. G. Matthews, Heiko Strathmann

To obtain an importance sampling estimate of the marginal likelihood, AIS introduces an extended target distribution to reweight the Markov chain proposal.

Directly Training Joint Energy-Based Models for Conditional Synthesis and Calibrated Prediction of Multi-Attribute Data

1 code implementation19 Jul 2021 Jacob Kelly, Richard Zemel, Will Grathwohl

We find our models are capable of both accurate, calibrated predictions and high-quality conditional synthesis of novel attribute combinations.

Oops I Took A Gradient: Scalable Sampling for Discrete Distributions

1 code implementation8 Feb 2021 Will Grathwohl, Kevin Swersky, Milad Hashemi, David Duvenaud, Chris J. Maddison

We propose a general and scalable approximate sampling strategy for probabilistic models with discrete variables.

Learning the Stein Discrepancy for Training and Evaluating Energy-Based Models without Sampling

1 code implementation ICML 2020 Will Grathwohl, Kuan-Chieh Wang, Jorn-Henrik Jacobsen, David Duvenaud, Richard Zemel

We estimate the Stein discrepancy between the data density $p(x)$ and the model density $q(x)$ defined by a vector function of the data.

Invertible Residual Networks

4 code implementations2 Nov 2018 Jens Behrmann, Will Grathwohl, Ricky T. Q. Chen, David Duvenaud, Jörn-Henrik Jacobsen

We show that standard ResNet architectures can be made invertible, allowing the same model to be used for classification, density estimation, and generation.

Density Estimation General Classification +1

FFJORD: Free-form Continuous Dynamics for Scalable Reversible Generative Models

7 code implementations ICLR 2019 Will Grathwohl, Ricky T. Q. Chen, Jesse Bettencourt, Ilya Sutskever, David Duvenaud

The result is a continuous-time invertible generative model with unbiased density estimation and one-pass sampling, while allowing unrestricted neural network architectures.

 Ranked #1 on Density Estimation on CIFAR-10 (NLL metric)

Density Estimation Image Generation +1

Disentangling Space and Time in Video with Hierarchical Variational Auto-encoders

no code implementations14 Dec 2016 Will Grathwohl, Aaron Wilson

In this paper we propose a probabilistic approach for learning separable representations of object identity and pose information using unsupervised video data.

Transfer Learning Variational Inference

Cannot find the paper you are looking for? You can Submit a new open access paper.