1 code implementation • 3 May 2023 • Heewoo Jun, Alex Nichol
We present Shap-E, a conditional generative model for 3D assets.
1 code implementation • 16 Dec 2022 • Alex Nichol, Heewoo Jun, Prafulla Dhariwal, Pamela Mishkin, Mark Chen
This is in stark contrast to state-of-the-art generative image models, which produce samples in a number of seconds or minutes.
6 code implementations • 13 Apr 2022 • Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, Mark Chen
Contrastive models like CLIP have been shown to learn robust representations of images that capture both semantics and style.
Ranked #27 on
Text-to-Image Generation
on COCO
(using extra training data)
Conditional Image Generation
Zero-Shot Text-to-Image Generation
2 code implementations • 20 Dec 2021 • Alex Nichol, Prafulla Dhariwal, Aditya Ramesh, Pranav Shyam, Pamela Mishkin, Bob McGrew, Ilya Sutskever, Mark Chen
Diffusion models have recently been shown to generate high-quality synthetic images, especially when paired with a guidance technique to trade off diversity for fidelity.
Ranked #32 on
Text-to-Image Generation
on COCO
(using extra training data)
13 code implementations • 7 Jul 2021 • Mark Chen, Jerry Tworek, Heewoo Jun, Qiming Yuan, Henrique Ponde de Oliveira Pinto, Jared Kaplan, Harri Edwards, Yuri Burda, Nicholas Joseph, Greg Brockman, Alex Ray, Raul Puri, Gretchen Krueger, Michael Petrov, Heidy Khlaaf, Girish Sastry, Pamela Mishkin, Brooke Chan, Scott Gray, Nick Ryder, Mikhail Pavlov, Alethea Power, Lukasz Kaiser, Mohammad Bavarian, Clemens Winter, Philippe Tillet, Felipe Petroski Such, Dave Cummings, Matthias Plappert, Fotios Chantzis, Elizabeth Barnes, Ariel Herbert-Voss, William Hebgen Guss, Alex Nichol, Alex Paino, Nikolas Tezak, Jie Tang, Igor Babuschkin, Suchir Balaji, Shantanu Jain, William Saunders, Christopher Hesse, Andrew N. Carr, Jan Leike, Josh Achiam, Vedant Misra, Evan Morikawa, Alec Radford, Matthew Knight, Miles Brundage, Mira Murati, Katie Mayer, Peter Welinder, Bob McGrew, Dario Amodei, Sam McCandlish, Ilya Sutskever, Wojciech Zaremba
We introduce Codex, a GPT language model fine-tuned on publicly available code from GitHub, and study its Python code-writing capabilities.
Ranked #1 on
Multi-task Language Understanding
on BBH-alg
16 code implementations • NeurIPS 2021 • Prafulla Dhariwal, Alex Nichol
Finally, we find that classifier guidance combines well with upsampling diffusion models, further improving FID to 3. 94 on ImageNet 256$\times$256 and 3. 85 on ImageNet 512$\times$512.
Ranked #1 on
Image Generation
on LSUN Bedroom 256 x 256
(FD metric)
9 code implementations • 18 Feb 2021 • Alex Nichol, Prafulla Dhariwal
Denoising diffusion probabilistic models (DDPM) are a class of generative models which have recently been shown to produce excellent samples.
Ranked #4 on
Image Generation
on CIFAR-10
(FD metric)
1 code implementation • 3 Mar 2020 • Alex Nichol
In this paper, I present VQ-DRAW, an algorithm for learning compact discrete representations of data.
3 code implementations • 10 Apr 2018 • Alex Nichol, Vicki Pfau, Christopher Hesse, Oleg Klimov, John Schulman
In this report, we present a new reinforcement learning (RL) benchmark based on the Sonic the Hedgehog (TM) video game franchise.
12 code implementations • 8 Mar 2018 • Alex Nichol, Joshua Achiam, John Schulman
This paper considers meta-learning problems, where there is a distribution of tasks, and we would like to obtain an agent that performs well (i. e., learns quickly) when presented with a previously unseen task sampled from this distribution.