no code implementations • 18 Mar 2024 • Axel Sauer, Frederic Boesel, Tim Dockhorn, Andreas Blattmann, Patrick Esser, Robin Rombach
Distillation methods, like the recently introduced adversarial diffusion distillation (ADD) aim to shift the model from many-shot to single-step inference, albeit at the cost of expensive and difficult optimization due to its reliance on a fixed pretrained DINOv2 discriminator.
1 code implementation • 5 Mar 2024 • Patrick Esser, Sumith Kulal, Andreas Blattmann, Rahim Entezari, Jonas Müller, Harry Saini, Yam Levi, Dominik Lorenz, Axel Sauer, Frederic Boesel, Dustin Podell, Tim Dockhorn, Zion English, Kyle Lacey, Alex Goodwin, Yannik Marek, Robin Rombach
Rectified flow is a recent generative model formulation that connects data and noise in a straight line.
2 code implementations • None 2023 • Andreas Blattmann, Tim Dockhorn, Sumith Kulal, Daniel Mendelevitch, Maciej Kilian, Dominik Lorenz, Yam Levi, Zion English, Vikram Voleti, Adam Letts, Varun Jampani, Robin Rombach
We then explore the impact of finetuning our base model on high-quality data and train a text-to-video model that is competitive with closed-source video generation.
3 code implementations • 4 Jul 2023 • Dustin Podell, Zion English, Kyle Lacey, Andreas Blattmann, Tim Dockhorn, Jonas Müller, Joe Penna, Robin Rombach
We present SDXL, a latent diffusion model for text-to-image synthesis.
2 code implementations • CVPR 2023 • Andreas Blattmann, Robin Rombach, Huan Ling, Tim Dockhorn, Seung Wook Kim, Sanja Fidler, Karsten Kreis
We first pre-train an LDM on images only; then, we turn the image generator into a video generator by introducing a temporal dimension to the latent space diffusion model and fine-tuning on encoded image sequences, i. e., videos.
Ranked #5 on Text-to-Video Generation on MSR-VTT (CLIP-FID metric)
no code implementations • 25 Nov 2022 • Karsten Kreis, Tim Dockhorn, Zihao Li, Ellen Zhong
The state-of-the-art method cryoDRGN uses a Variational Autoencoder (VAE) framework to learn a continuous distribution of protein structures from single particle cryo-EM imaging data.
1 code implementation • 18 Oct 2022 • Tim Dockhorn, Tianshi Cao, Arash Vahdat, Karsten Kreis
While modern machine learning models rely on increasingly large training datasets, data is often limited in privacy-sensitive domains.
1 code implementation • 11 Oct 2022 • Tim Dockhorn, Arash Vahdat, Karsten Kreis
Synthesis amounts to solving a differential equation (DE) defined by the learnt model.
Ranked #5 on Image Generation on AFHQV2
1 code implementation • ICLR 2022 • Tim Dockhorn, Arash Vahdat, Karsten Kreis
SGMs rely on a diffusion process that gradually perturbs the data towards a tractable distribution, while the generative model learns to denoise.
Ranked #24 on Image Generation on CIFAR-10
no code implementations • NeurIPS 2021 • Tim Dockhorn, YaoLiang Yu, Eyyüb Sari, Mahdi Zolnouri, Vahid Partovi Nia
BinaryConnect (BC) and its many variations have become the de facto standard for neural network quantization.
1 code implementation • 16 Jun 2020 • Tim Dockhorn, James A. Ritchie, Yao-Liang Yu, Iain Murray
Density deconvolution is the task of estimating a probability density function given only noise-corrupted samples.
1 code implementation • 15 Apr 2019 • Tim Dockhorn
Can neural networks learn to solve partial differential equations (PDEs)?