no code implementations • 24 Jul 2023 • Izzeddin Gur, Hiroki Furuta, Austin Huang, Mustafa Safdari, Yutaka Matsuo, Douglas Eck, Aleksandra Faust
Pre-trained large language models (LLMs) have recently achieved better generalization and sample efficiency in autonomous web automation.
5 code implementations • Google Research 2022 • Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, Parker Schuh, Kensen Shi, Sasha Tsvyashchenko, Joshua Maynez, Abhishek Rao, Parker Barnes, Yi Tay, Noam Shazeer, Vinodkumar Prabhakaran, Emily Reif, Nan Du, Ben Hutchinson, Reiner Pope, James Bradbury, Jacob Austin, Michael Isard, Guy Gur-Ari, Pengcheng Yin, Toju Duke, Anselm Levskaya, Sanjay Ghemawat, Sunipa Dev, Henryk Michalewski, Xavier Garcia, Vedant Misra, Kevin Robinson, Liam Fedus, Denny Zhou, Daphne Ippolito, David Luan, Hyeontaek Lim, Barret Zoph, Alexander Spiridonov, Ryan Sepassi, David Dohan, Shivani Agrawal, Mark Omernick, Andrew M. Dai, Thanumalayan Sankaranarayana Pillai, Marie Pellat, Aitor Lewkowycz, Erica Moreira, Rewon Child, Oleksandr Polozov, Katherine Lee, Zongwei Zhou, Xuezhi Wang, Brennan Saeta, Mark Diaz, Orhan Firat, Michele Catasta, Jason Wei, Kathy Meier-Hellstern, Douglas Eck, Jeff Dean, Slav Petrov, Noah Fiedel
To further our understanding of the impact of scale on few-shot learning, we trained a 540-billion parameter, densely activated, Transformer language model, which we call Pathways Language Model PaLM.
Ranked #1 on
Natural Language Inference
on RTE
1 code implementation • ACL 2022 • Katherine Lee, Daphne Ippolito, Andrew Nystrom, Chiyuan Zhang, Douglas Eck, Chris Callison-Burch, Nicholas Carlini
As a result, over 1% of the unprompted output of language models trained on these datasets is copied verbatim from the training data.
no code implementations • 15 Apr 2021 • Dennis Lee, Natasha Jaques, Chase Kew, Jiaxing Wu, Douglas Eck, Dale Schuurmans, Aleksandra Faust
We then train agents to minimize the difference between the attention weights that they apply to the environment at each timestep, and the attention of other agents.
no code implementations • 1 Oct 2020 • Kamal Ndousse, Douglas Eck, Sergey Levine, Natasha Jaques
We analyze the reasons for this deficiency, and show that by imposing constraints on the training environment and introducing a model-based auxiliary loss we are able to obtain generalized social learning policies which enable agents to: i) discover complex skills that are not learned from single-agent training, and ii) adapt online to novel environments by taking cues from experts present in the new environment.
1 code implementation • ACL 2020 • Daphne Ippolito, David Grangier, Douglas Eck, Chris Callison-Burch
We propose a sentence-level language model which selects the next sentence in a story from a finite set of fluent alternatives.
2 code implementations • ACL 2020 • Daphne Ippolito, Daniel Duckworth, Chris Callison-Burch, Douglas Eck
Recent advancements in neural language modelling make it possible to rapidly generate vast amounts of human-sounding text.
no code implementations • WS 2019 • Daphne Ippolito, David Grangier, Chris Callison-Burch, Douglas Eck
Story infilling involves predicting words to go into a missing span from a story.
no code implementations • 14 May 2019 • Jon Gillick, Adam Roberts, Jesse Engel, Douglas Eck, David Bamman
We explore models for translating abstract musical ideas (scores, rhythms) into expressive performances using Seq2Seq and recurrent Variational Information Bottleneck (VIB) models.
2 code implementations • ICCV 2019 • Raphael Gontijo Lopes, David Ha, Douglas Eck, Jonathon Shlens
Dramatic advances in generative models have resulted in near photographic quality for artificially rendered faces, animals and other objects in the natural world.
4 code implementations • 18 Mar 2019 • Cheng-Zhi Anna Huang, Tim Cooijmans, Adam Roberts, Aaron Courville, Douglas Eck
Machine learning models of music typically break up the task of composition into a chronological process, composing a piece of music in a single pass from beginning to end.
Ranked #4 on
Music Modeling
on JSB Chorales
4 code implementations • ICLR 2019 • Curtis Hawthorne, Andriy Stasyuk, Adam Roberts, Ian Simon, Cheng-Zhi Anna Huang, Sander Dieleman, Erich Elsen, Jesse Engel, Douglas Eck
Generating musical audio directly with neural networks is notoriously difficult because it requires coherently modeling structure at many different timescales.
12 code implementations • ICLR 2019 • Cheng-Zhi Anna Huang, Ashish Vaswani, Jakob Uszkoreit, Noam Shazeer, Ian Simon, Curtis Hawthorne, Andrew M. Dai, Matthew D. Hoffman, Monica Dinculescu, Douglas Eck
This is impractical for long sequences such as musical compositions since their memory complexity for intermediate relative information is quadratic in the sequence length.
Ranked #3 on
Music Modeling
on JSB Chorales
5 code implementations • 10 Aug 2018 • Sageev Oore, Ian Simon, Sander Dieleman, Douglas Eck, Karen Simonyan
Music generation has generally been focused on either creating scores or interpreting them.
1 code implementation • 1 Jun 2018 • Ian Simon, Adam Roberts, Colin Raffel, Jesse Engel, Curtis Hawthorne, Douglas Eck
Discovering and exploring the underlying structure of multi-instrumental music using learning-based approaches remains an open problem.
7 code implementations • ICML 2018 • Adam Roberts, Jesse Engel, Colin Raffel, Curtis Hawthorne, Douglas Eck
The Variational Autoencoder (VAE) has proven to be an effective model for producing semantically meaningful latent representations for natural data.
no code implementations • 13 Feb 2018 • Natasha Jaques, Jennifer McCleary, Jesse Engel, David Ha, Fred Bertsch, Rosalind Picard, Douglas Eck
We use a Latent Constraints GAN (LC-GAN) to learn from the facial feedback of a small group of viewers, by optimizing the model to produce sketches that it predicts will lead to more positive facial expressions.
1 code implementation • 30 Oct 2017 • Curtis Hawthorne, Erich Elsen, Jialin Song, Adam Roberts, Ian Simon, Colin Raffel, Jesse Engel, Sageev Oore, Douglas Eck
We advance the state of the art in polyphonic piano music transcription by using a deep convolutional and recurrent neural network which is trained to jointly predict onsets and frames.
no code implementations • ICLR 2018 • Andrew Kyle Lampinen, David So, Douglas Eck, Fred Bertsch
GANs provide a framework for training generative models which mimic a data distribution.
19 code implementations • ICLR 2018 • David Ha, Douglas Eck
We present sketch-rnn, a recurrent neural network (RNN) able to construct stroke-based drawings of common objects.
5 code implementations • ICML 2017 • Jesse Engel, Cinjon Resnick, Adam Roberts, Sander Dieleman, Douglas Eck, Karen Simonyan, Mohammad Norouzi
Generative models in vision have seen rapid progress due to algorithmic improvements and the availability of high-quality image datasets.
2 code implementations • ICML 2017 • Colin Raffel, Minh-Thang Luong, Peter J. Liu, Ron J. Weiss, Douglas Eck
Recurrent neural network models with an attention mechanism have proven to be extremely effective on a wide variety of sequence-to-sequence problems.
Ranked #20 on
Speech Recognition
on TIMIT
no code implementations • ICML 2017 • Natasha Jaques, Shixiang Gu, Dzmitry Bahdanau, José Miguel Hernández-Lobato, Richard E. Turner, Douglas Eck
This paper proposes a general method for improving the structure and quality of sequences generated by a recurrent neural network (RNN), while maintaining information originally learned from data, as well as sample diversity.
no code implementations • NeurIPS 2009 • Douglas Eck, Yoshua Bengio, Aaron C. Courville
The Indian Buffet Process is a Bayesian nonparametric approach that models objects as arising from an infinite number of latent factors.
no code implementations • NeurIPS 2007 • Douglas Eck, Paul Lamere, Thierry Bertin-Mahieux, Stephen Green
Social tags are user-generated keywords associated with some resource on the Web.