no code implementations • 25 Oct 2024 • Emiel Hoogeboom, Thomas Mensink, Jonathan Heek, Kay Lamerigts, Ruiqi Gao, Tim Salimans
Compared to pixel-space models that are trained end-to-end, latent models are perceived to be more efficient and to produce higher image quality at high resolution.
Ranked #1 on Image Generation on ImageNet 128x128
no code implementations • 6 Jun 2024 • Tim Salimans, Thomas Mensink, Jonathan Heek, Emiel Hoogeboom
We present a new method for making diffusion models faster to sample.
no code implementations • 27 May 2024 • Sirui Xie, Zhisheng Xiao, Diederik P Kingma, Tingbo Hou, Ying Nian Wu, Kevin Patrick Murphy, Tim Salimans, Ben Poole, Ruiqi Gao
We propose EM Distillation (EMD), a maximum likelihood-based approach that distills a diffusion model to a one-step generator model with minimal loss of perceptual quality.
no code implementations • 11 Mar 2024 • Jonathan Heek, Emiel Hoogeboom, Tim Salimans
By increasing the sample budget from a single step to 2-8 steps, we can train models more easily that generate higher quality samples, while retaining much of the sampling speed benefits.
no code implementations • 12 Feb 2024 • David Ruhe, Jonathan Heek, Tim Salimans, Emiel Hoogeboom
Diffusion models have recently been increasingly applied to temporal data such as video, fluid mechanics simulations, or climate data.
1 code implementation • 26 Jan 2023 • Emiel Hoogeboom, Jonathan Heek, Tim Salimans
Currently, applying diffusion models in pixel space of high resolution images is difficult.
Ranked #4 on Conditional Image Generation on ImageNet 128x128
2 code implementations • CVPR 2023 • Chenlin Meng, Robin Rombach, Ruiqi Gao, Diederik P. Kingma, Stefano Ermon, Jonathan Ho, Tim Salimans
For standard diffusion models trained on the pixel-space, our approach is able to generate images visually comparable to that of the original model using as few as 4 sampling steps on ImageNet 64x64 and CIFAR-10, achieving FID/IS scores comparable to that of the original model while being up to 256 times faster to sample from.
no code implementations • 5 Oct 2022 • Jonathan Ho, William Chan, Chitwan Saharia, Jay Whang, Ruiqi Gao, Alexey Gritsenko, Diederik P. Kingma, Ben Poole, Mohammad Norouzi, David J. Fleet, Tim Salimans
We present Imagen Video, a text-conditional video generation system based on a cascade of video diffusion models.
Ranked #1 on Video Generation on LAION-400M
no code implementations • 12 Sep 2022 • Emiel Hoogeboom, Tim Salimans
Recently, Rissanen et al., (2022) have presented a new type of diffusion process for generative modeling based on heat dissipation, or blurring, as an alternative to isotropic Gaussian diffusion.
10 code implementations • 26 Jul 2022 • Jonathan Ho, Tim Salimans
Classifier guidance is a recently introduced method to trade off mode coverage and sample fidelity in conditional diffusion models post training, in the same spirit as low temperature sampling or truncation in other types of generative models.
no code implementations • 17 Jun 2022 • Lucas Theis, Tim Salimans, Matthew D. Hoffman, Fabian Mentzer
Unlike modern compression schemes which rely on transform coding and quantization to restrict the transmitted information, DiffC relies on the efficient communication of pixels corrupted by Gaussian noise.
5 code implementations • 23 May 2022 • Chitwan Saharia, William Chan, Saurabh Saxena, Lala Li, Jay Whang, Emily Denton, Seyed Kamyar Seyed Ghasemipour, Burcu Karagol Ayan, S. Sara Mahdavi, Rapha Gontijo Lopes, Tim Salimans, Jonathan Ho, David J Fleet, Mohammad Norouzi
We present Imagen, a text-to-image diffusion model with an unprecedented degree of photorealism and a deep level of language understanding.
Ranked #16 on Text-to-Image Generation on MS COCO (using extra training data)
4 code implementations • 7 Apr 2022 • Jonathan Ho, Tim Salimans, Alexey Gritsenko, William Chan, Mohammad Norouzi, David J. Fleet
Generating temporally coherent high fidelity video is an important milestone in generative modeling research.
12 code implementations • ICLR 2022 • Tim Salimans, Jonathan Ho
Second, we present a method to distill a trained deterministic diffusion sampler, using many steps, into a new diffusion model that takes half as many sampling steps.
Ranked #45 on Image Generation on CIFAR-10
1 code implementation • NeurIPS 2021 • Diederik Kingma, Tim Salimans, Ben Poole, Jonathan Ho
In addition, we show that the continuous-time VLB is invariant to the noise schedule, except for the signal-to-noise ratio at its endpoints.
5 code implementations • 10 Nov 2021 • Chitwan Saharia, William Chan, Huiwen Chang, Chris A. Lee, Jonathan Ho, Tim Salimans, David J. Fleet, Mohammad Norouzi
We expect this standardized evaluation protocol to play a role in advancing image-to-image translation research.
Ranked #1 on Colorization on ImageNet ctest10k
3 code implementations • ICLR 2022 • Emiel Hoogeboom, Alexey A. Gritsenko, Jasmijn Bastings, Ben Poole, Rianne van den Berg, Tim Salimans
We introduce Autoregressive Diffusion Models (ARDMs), a model class encompassing and generalizing order-agnostic autoregressive models (Uria et al., 2014) and absorbing discrete diffusion (Austin et al., 2021), which we show are special cases of ARDMs under mild assumptions.
Ranked #10 on Image Generation on CIFAR-10 (bits/dimension metric)
no code implementations • 29 Sep 2021 • Jonathan Ho, Tim Salimans
Classifier guidance is a recently introduced method to trade off mode coverage and sample fidelity in conditional diffusion models post training, in the same spirit as low temperature sampling or truncation in other types of generative models.
4 code implementations • 1 Jul 2021 • Diederik P. Kingma, Tim Salimans, Ben Poole, Jonathan Ho
In addition, we show that the continuous-time VLB is invariant to the noise schedule, except for the signal-to-noise ratio at its endpoints.
Ranked #2 on Density Estimation on CIFAR-10
no code implementations • 30 May 2021 • Jonathan Ho, Chitwan Saharia, William Chan, David J. Fleet, Mohammad Norouzi, Tim Salimans
We show that cascaded diffusion models are capable of generating high fidelity images on the class-conditional ImageNet generation benchmark, without any assistance from auxiliary image classifiers to boost sample quality.
Ranked #8 on Image Generation on ImageNet 64x64
no code implementations • 19 Apr 2021 • Wenling Shang, Lasse Espeholt, Anton Raichuk, Tim Salimans
Empirically, agent-centric representation learning leads to the emergence of more complex cooperation strategies between agents as well as enhanced sample efficiency and generalization.
4 code implementations • 15 Apr 2021 • Chitwan Saharia, Jonathan Ho, William Chan, Tim Salimans, David J. Fleet, Mohammad Norouzi
We present SR3, an approach to image Super-Resolution via Repeated Refinement.
no code implementations • ICLR Workshop EBM 2021 • Tim Salimans, Jonathan Ho
Recent progress in training unnormalized models through denoising score matching with Langevin dynamics (SMLD) and denoising diffusion probabilistic modeling (DDPM) has made unnormalized models a competitive model class for generative modeling.
2 code implementations • NeurIPS 2020 • Alexey A. Gritsenko, Tim Salimans, Rianne van den Berg, Jasper Snoek, Nal Kalchbrenner
Speech synthesis is an important practical generative modeling problem that has seen great progress over the last few years, with likelihood-based autoregressive neural models now outperforming traditional concatenative systems.
no code implementations • ICLR 2021 • Rianne van den Berg, Alexey A. Gritsenko, Mostafa Dehghani, Casper Kaae Sønderby, Tim Salimans
Furthermore, we zoom in on the effect of gradient bias due to the straight-through estimator in integer discrete flows, and demonstrate that its influence is highly dependent on architecture choices and less prominent than previously thought.
2 code implementations • 26 Mar 2020 • Geoff French, Avital Oliver, Tim Salimans
Using it to provide perturbations for semi-supervised consistency regularization, we achieve a state-of-the-art result on ImageNet with 10% labeled data, with a top-5 error of 8. 76% and top-1 error of 26. 06%.
2 code implementations • 24 Mar 2020 • Casper Kaae Sønderby, Lasse Espeholt, Jonathan Heek, Mostafa Dehghani, Avital Oliver, Tim Salimans, Shreya Agrawal, Jason Hickey, Nal Kalchbrenner
Weather forecasting is a long standing scientific challenge with direct social and economic impact.
no code implementations • ICML 2020 • Jakub Swiatkowski, Kevin Roth, Bastiaan S. Veeling, Linh Tran, Joshua V. Dillon, Jasper Snoek, Stephan Mandt, Tim Salimans, Rodolphe Jenatton, Sebastian Nowozin
Variational Bayesian Inference is a popular methodology for approximating posterior distributions over Bayesian neural network weights.
1 code implementation • ICML 2020 • Florian Wenzel, Kevin Roth, Bastiaan S. Veeling, Jakub Świątkowski, Linh Tran, Stephan Mandt, Jasper Snoek, Tim Salimans, Rodolphe Jenatton, Sebastian Nowozin
In this work we cast doubt on the current understanding of Bayes posteriors in popular deep neural networks: we demonstrate through careful MCMC sampling that the posterior predictive induced by the Bayes posterior yields systematically worse predictions compared to simpler methods including point estimates obtained from SGD.
1 code implementation • 14 Jan 2020 • Linh Tran, Bastiaan S. Veeling, Kevin Roth, Jakub Swiatkowski, Joshua V. Dillon, Jasper Snoek, Stephan Mandt, Tim Salimans, Sebastian Nowozin, Rodolphe Jenatton
As a result, the diversity of the ensemble predictions, stemming from each member, is lost.
2 code implementations • 20 Dec 2019 • Jonathan Ho, Nal Kalchbrenner, Dirk Weissenborn, Tim Salimans
We propose Axial Transformers, a self-attention-based autoregressive model for images and other data organized as high dimensional tensors.
Ranked #29 on Image Generation on ImageNet 64x64 (Bits per dim metric)
1 code implementation • 13 Dec 2019 • Christopher Berner, Greg Brockman, Brooke Chan, Vicki Cheung, Przemysław Dębiak, Christy Dennison, David Farhi, Quirin Fischer, Shariq Hashme, Chris Hesse, Rafal Józefowicz, Scott Gray, Catherine Olsson, Jakub Pachocki, Michael Petrov, Henrique Pondé de Oliveira Pinto, Jonathan Raiman, Tim Salimans, Jeremy Schlatter, Jonas Schneider, Szymon Sidor, Ilya Sutskever, Jie Tang, Filip Wolski, Susan Zhang
On April 13th, 2019, OpenAI Five became the first AI system to defeat the world champions at an esports game.
no code implementations • 25 Sep 2019 • Jakub Świątkowski, Kevin Roth, Bastiaan S. Veeling, Linh Tran, Joshua V. Dillon, Jasper Snoek, Stephan Mandt, Tim Salimans, Rodolphe Jenatton, Sebastian Nowozin
Variational Bayesian Inference is a popular methodology for approximating posterior distributions in Bayesian neural networks.
1 code implementation • 9 May 2019 • Jaap H. Abbring, Tim Salimans
We present a method for computing the likelihood of a mixed hitting-time model that specifies durations as the first time a latent L\'evy process crosses a heterogeneous threshold.
no code implementations • 7 Apr 2019 • Thomas Anthony, Robert Nishihara, Philipp Moritz, Tim Salimans, John Schulman
Monte Carlo Tree Search (MCTS) algorithms perform simulation-based search to improve policies online.
no code implementations • ICLR Workshop DeepGenStruct 2019 • Alexey A. Gritsenko, Jasper Snoek, Tim Salimans
Normalising Flows (NFs) are a class of likelihood-based generative models that have recently gained popularity.
no code implementations • 8 Dec 2018 • Tim Salimans, Richard Chen
We propose a new method for learning from a single demonstration to solve hard exploration tasks like the Atari game Montezuma's Revenge.
12 code implementations • Preprint 2018 • Alec Radford, Karthik Narasimhan, Tim Salimans, Ilya Sutskever
We demonstrate that large gains on these tasks can be realized by generative pre-training of a language model on a diverse corpus of unlabeled text, followed by discriminative fine-tuning on each specific task.
Ranked #3 on Natural Language Inference on SciTail
2 code implementations • ICLR 2018 • Tim Salimans, Han Zhang, Alec Radford, Dimitris Metaxas
We present Optimal Transport GAN (OT-GAN), a variant of generative adversarial nets minimizing a new metric measuring the distance between the generator distribution and the data distribution.
23 code implementations • 10 Mar 2017 • Tim Salimans, Jonathan Ho, Xi Chen, Szymon Sidor, Ilya Sutskever
We explore the use of Evolution Strategies (ES), a class of black box optimization algorithms, as an alternative to popular MDP-based RL techniques such as Q-learning and Policy Gradients.
Ranked #1 on Atari Games on Atari 2600 Pong
7 code implementations • 19 Jan 2017 • Tim Salimans, Andrej Karpathy, Xi Chen, Diederik P. Kingma
1) We use a discretized logistic mixture likelihood on the pixels, rather than a 256-way softmax, which we find to speed up training.
Ranked #4 on Density Estimation on CIFAR-10
2 code implementations • NeurIPS 2016 • Durk P. Kingma, Tim Salimans, Rafal Jozefowicz, Xi Chen, Ilya Sutskever, Max Welling
The framework of normalizing flows provides a general strategy for flexible variational inference of posteriors over latent variables.
Ranked #44 on Image Generation on CIFAR-10 (bits/dimension metric)
no code implementations • 8 Nov 2016 • Xi Chen, Diederik P. Kingma, Tim Salimans, Yan Duan, Prafulla Dhariwal, John Schulman, Ilya Sutskever, Pieter Abbeel
Representation learning seeks to expose certain aspects of observed data in a learned representation that's amenable to downstream tasks like classification.
8 code implementations • 15 Jun 2016 • Diederik P. Kingma, Tim Salimans, Rafal Jozefowicz, Xi Chen, Ilya Sutskever, Max Welling
The framework of normalizing flows provides a general strategy for flexible variational inference of posteriors over latent variables.
46 code implementations • NeurIPS 2016 • Tim Salimans, Ian Goodfellow, Wojciech Zaremba, Vicki Cheung, Alec Radford, Xi Chen
We present a variety of new architectural features and training procedures that we apply to the generative adversarial networks (GANs) framework.
Ranked #15 on Conditional Image Generation on CIFAR-10 (Inception score metric)
Conditional Image Generation Semi-Supervised Image Classification
no code implementations • 28 Feb 2016 • Tim Salimans
To learn the parameters of the new model, we approximate the posterior of the latent variables with a variational auto-encoder.
9 code implementations • NeurIPS 2016 • Tim Salimans, Diederik P. Kingma
We present weight normalization: a reparameterization of the weight vectors in a neural network that decouples the length of those weight vectors from their direction.
12 code implementations • NeurIPS 2015 • Diederik P. Kingma, Tim Salimans, Max Welling
Our method allows inference of more flexibly parameterized posteriors; specifically, we propose variational dropout, a generalization of Gaussian dropout where the dropout rates are learned, often leading to better models.
no code implementations • 23 Oct 2014 • Tim Salimans, Diederik P. Kingma, Max Welling
Recent advances in stochastic gradient variational inference have made it possible to perform variational Bayesian inference with posterior approximations containing auxiliary random variables.
2 code implementations • 28 Jun 2012 • Tim Salimans, David A. Knowles
We propose a general algorithm for approximating nonstandard Bayesian posterior distributions.