no code implementations • 12 Sep 2024 • Vinicius Zambaldi, David La, Alexander E. Chu, Harshnira Patani, Amy E. Danson, Tristan O. C. Kwan, Thomas Frerix, Rosalia G. Schneider, David Saxton, Ashok Thillaisundaram, Zachary Wu, Isabel Moraes, Oskar Lange, Eliseo Papa, Gabriella Stanton, Victor Martin, Sukhdeep Singh, Lai H. Wong, Russ Bates, Simon A. Kohl, Josh Abramson, Andrew W. Senior, Yilmaz Alguel, Mary Y. Wu, Irene M. Aspalter, Katie Bentley, David L. V. Bauer, Peter Cherepanov, Demis Hassabis, Pushmeet Kohli, Rob Fergus, Jue Wang
Computational design of protein-binding proteins is a fundamental capability with broad utility in biomedical research and biotechnology.
no code implementations • 3 Sep 2024 • Nicholas Monath, Will Grathwohl, Michael Boratko, Rob Fergus, Andrew McCallum, Manzil Zaheer
In dense retrieval, deep encoders provide embeddings for both inputs and targets, and the softmax function is used to parameterize a distribution over a large number of candidate targets (e. g., textual passages for information retrieval).
no code implementations • 21 Aug 2024 • Anthony GX-Chen, Kenneth Marino, Rob Fergus
In the face of difficult exploration problems in reinforcement learning, we study whether giving an agent an object-centric mapping (describing a set of items and their attributes) allow for more efficient learning.
1 code implementation • 24 Jun 2024 • Shengbang Tong, Ellis Brown, Penghao Wu, Sanghyun Woo, Manoj Middepogu, Sai Charitha Akula, Jihan Yang, Shusheng Yang, Adithya Iyer, Xichen Pan, Austin Wang, Rob Fergus, Yann Lecun, Saining Xie
We introduce Cambrian-1, a family of multimodal LLMs (MLLMs) designed with a vision-centric approach.
no code implementations • 6 May 2024 • Nishant Yadav, Nicholas Monath, Manzil Zaheer, Rob Fergus, Andrew McCallum
Our method produces a high-quality approximation while requiring only a fraction of CE calls as compared to CUR-based methods, and allows for leveraging DE to initialize the embedding space while avoiding compute- and resource-intensive finetuning of DE via distillation.
1 code implementation • 12 Dec 2023 • Ulyana Piterbarg, Lerrel Pinto, Rob Fergus
On NetHack, an unsolved video game that requires long-horizon reasoning for decision-making, LMs tuned with diff history match state-of-the-art performance for neural agents while needing 1800x fewer training examples compared to prior work.
no code implementations • 20 Sep 2023 • Arun Ahuja, Kavya Kopparapu, Rob Fergus, Ishita Dasgupta
Hierarchical reinforcement learning has been a compelling approach for achieving goal directed behavior over long sequences of actions.
no code implementations • 31 Mar 2023 • Bogdan Mazoure, Jake Bruce, Doina Precup, Rob Fergus, Ankit Anand
In this work, we follow the hypothesis that exploration and representation learning can be improved by separately learning two different models from a single offline dataset.
3 code implementations • 22 Feb 2023 • Yilun Du, Conor Durkan, Robin Strudel, Joshua B. Tenenbaum, Sander Dieleman, Rob Fergus, Jascha Sohl-Dickstein, Arnaud Doucet, Will Grathwohl
In this work, we build upon these ideas using the score-based interpretation of diffusion models, and explore alternative ways to condition, modify, and reuse diffusion models for tasks involving compositional generation and guidance.
no code implementations • 1 Feb 2023 • Ishita Dasgupta, Christine Kaeser-Chen, Kenneth Marino, Arun Ahuja, Sheila Babayan, Felix Hill, Rob Fergus
On the other hand, Large Scale Language Models (LSLMs) have exhibited strong reasoning ability and the ability to to adapt to new tasks through in-context learning.
no code implementations • 29 Jan 2023 • Theodore Sumers, Kenneth Marino, Arun Ahuja, Rob Fergus, Ishita Dasgupta
Instruction-following agents must ground language into their observation and action spaces.
no code implementations • 27 Jan 2023 • Seungyeon Kim, Ankit Singh Rawat, Manzil Zaheer, Sadeep Jayasumana, Veeranjaneyulu Sadhanala, Wittawat Jitkrittum, Aditya Krishna Menon, Rob Fergus, Sanjiv Kumar
Large neural models (such as Transformers) achieve state-of-the-art performance for information retrieval (IR).
no code implementations • 31 Oct 2022 • Manzil Zaheer, Kenneth Marino, Will Grathwohl, John Schultz, Wendy Shang, Sheila Babayan, Arun Ahuja, Ishita Dasgupta, Christine Kaeser-Chen, Rob Fergus
A fundamental ability of an intelligent web-based agent is seeking out and acquiring new information.
no code implementations • 14 Aug 2022 • Manzil Zaheer, Ankit Singh Rawat, Seungyeon Kim, Chong You, Himanshu Jain, Andreas Veit, Rob Fergus, Sanjiv Kumar
In this paper, we propose the teacher-guided training (TGT) framework for training a high-quality compact model that leverages the knowledge acquired by pretrained generative models, while obviating the need to go through a large volume of data.
1 code implementation • NeurIPS 2021 • Roberta Raileanu, Maxwell Goldstein, Denis Yarats, Ilya Kostrikov, Rob Fergus
Deep reinforcement learning (RL) agents often fail to generalize beyond their training environments.
no code implementations • 29 Sep 2021 • Maxwell Goldstein, Leon Bottou, Rob Fergus
Contemporary ranking systems that are based on win/loss history, such as Elo or TrueSkill represent each player using a scalar estimate of ability (plus variance, in the latter case).
8 code implementations • ICLR 2022 • Denis Yarats, Rob Fergus, Alessandro Lazaric, Lerrel Pinto
We present DrQ-v2, a model-free reinforcement learning (RL) algorithm for visual continuous control.
no code implementations • 8 Jul 2021 • Andrew Jaegle, Yury Sulsky, Arun Ahuja, Jake Bruce, Rob Fergus, Greg Wayne
Imitation learning enables agents to reuse and adapt the hard-won expertise of others, offering a solution to several key challenges in learning behavior.
2 code implementations • 14 Mar 2021 • Ilya Kostrikov, Jonathan Tompson, Rob Fergus, Ofir Nachum
Many modern approaches to offline Reinforcement Learning (RL) utilize behavior regularization, typically augmenting a model-free actor critic algorithm with a penalty measuring divergence of the policy from the offline data.
1 code implementation • 22 Feb 2021 • Denis Yarats, Rob Fergus, Alessandro Lazaric, Lerrel Pinto
Unfortunately, in RL, representation learning is confounded with the exploratory experience of the agent -- learning a useful representation requires diverse data, while effective exploration is only possible with coherent representations.
2 code implementations • 20 Feb 2021 • Roberta Raileanu, Rob Fergus
Standard deep reinforcement learning algorithms use a shared representation for the policy and value function, especially when training directly from images.
1 code implementation • Proceedings of the National Academy of Sciences 2020 • Alexander Rives, Joshua Meier, Tom Sercu, Siddharth Goyal, Zeming Lin, Demi Guo, Myle Ott, C. Lawrence Zitnick, Jerry Ma, Rob Fergus
In the field of artificial intelligence, a combination of scale in data and model capacity enabled by unsupervised learning has led to major advances in representation learning and statistical generation.
1 code implementation • 6 Jul 2020 • Roberta Raileanu, Max Goldstein, Arthur Szlam, Rob Fergus
An ensemble of conventional RL policies is used to gather experience on training environments, from which embeddings of both policies and environments can be learned.
no code implementations • 29 Jun 2020 • Kenneth Marino, Rob Fergus, Arthur Szlam, Abhinav Gupta
This paper formulates hypothesis verification as an RL problem.
1 code implementation • NeurIPS 2021 • Roberta Raileanu, Max Goldstein, Denis Yarats, Ilya Kostrikov, Rob Fergus
Our agent outperforms other baselines specifically designed to improve generalization in RL.
4 code implementations • ICLR 2021 • Ilya Kostrikov, Denis Yarats, Rob Fergus
We propose a simple data augmentation technique that can be applied to standard model-free reinforcement learning algorithms, enabling robust learning directly from pixels without the need for auxiliary losses or pre-training.
Ranked #1 on Continuous Control on DeepMind Walker Walk (Images)
1 code implementation • ICLR 2020 • Yilun Du, Joshua Meier, Jerry Ma, Rob Fergus, Alexander Rives
We propose an energy-based model (EBM) of protein conformations that operates at atomic scale.
no code implementations • IJCNLP 2019 • Ethan Perez, Siddharth Karamcheti, Rob Fergus, Jason Weston, Douwe Kiela, Kyunghyun Cho
We propose a system that finds the strongest supporting evidence for a given answer to a question, using passage-based question-answering (QA) as a testbed.
3 code implementations • 2 Oct 2019 • Denis Yarats, Amy Zhang, Ilya Kostrikov, Brandon Amos, Joelle Pineau, Rob Fergus
A promising approach is to learn a latent representation together with the control policy.
no code implementations • 25 Sep 2019 • Kenneth Marino, Rob Fergus, Arthur Szlam, Abhinav Gupta
In order to train the agents, we exploit the underlying structure in the majority of hypotheses -- they can be formulated as triplets (pre-condition, action sequence, post-condition).
1 code implementation • 12 Sep 2019 • Ethan Perez, Siddharth Karamcheti, Rob Fergus, Jason Weston, Douwe Kiela, Kyunghyun Cho
We propose a system that finds the strongest supporting evidence for a given answer to a question, using passage-based question-answering (QA) as a testbed.
no code implementations • ICLR 2019 • Kenneth Marino, Abhinav Gupta, Rob Fergus, Arthur Szlam
The high-level policy is trained using a sparse, task-dependent reward, and operates by choosing which of the low-level policies to run at any given time.
no code implementations • 17 Jan 2019 • William F. Whitney, Rob Fergus
We propose an unsupervised variational model for disentangling video into independent factors, i. e. each factor's future can be predicted from its past without considering the others.
2 code implementations • 22 Nov 2018 • Sainbayar Sukhbaatar, Emily Denton, Arthur Szlam, Rob Fergus
In hierarchical reinforcement learning a major challenge is determining appropriate low-level policies.
Hierarchical Reinforcement Learning reinforcement-learning +1
no code implementations • 27 Sep 2018 • William Whitney, Rob Fergus
In complex simulated environments, model-based reinforcement learning methods typically lag the asymptotic performance of model-free approaches.
1 code implementation • 20 Mar 2018 • Ronan Riochet, Mario Ynocente Castro, Mathieu Bernard, Adam Lerer, Rob Fergus, Véronique Izard, Emmanuel Dupoux
In order to reach human performance on complexvisual tasks, artificial systems need to incorporate a sig-nificant amount of understanding of the world in termsof macroscopic objects, movements, forces, etc.
no code implementations • ICML 2018 • Amy Zhang, Adam Lerer, Sainbayar Sukhbaatar, Rob Fergus, Arthur Szlam
The tasks that an agent will need to solve often are not known during training.
1 code implementation • ICML 2018 • Roberta Raileanu, Emily Denton, Arthur Szlam, Rob Fergus
We consider the multi-agent reinforcement learning setting with imperfect information in which each agent is trying to maximize its own utility.
Multi-agent Reinforcement Learning reinforcement-learning +1
3 code implementations • ICML 2018 • Emily Denton, Rob Fergus
Sample generations are both varied and sharp, even many frames into the future, and compare favorably to those from existing approaches.
Ranked #3 on Video Prediction on KTH
no code implementations • CVPR 2018 • Ishan Misra, Ross Girshick, Rob Fergus, Martial Hebert, Abhinav Gupta, Laurens van der Maaten
We also show that our model asks questions that generalize to state-of-the-art VQA models and to novel test time distributions.
3 code implementations • ICLR 2018 • Sainbayar Sukhbaatar, Zeming Lin, Ilya Kostrikov, Gabriel Synnaeve, Arthur Szlam, Rob Fergus
When Bob is deployed on an RL task within the environment, this unsupervised training reduces the number of supervised episodes needed to learn, and in some cases converges to a higher reward.
7 code implementations • 19 Nov 2016 • Emily Denton, Sam Gross, Rob Fergus
We introduce a simple semi-supervised learning approach for images based on in-painting using an adversarial loss.
9 code implementations • NeurIPS 2016 • Sainbayar Sukhbaatar, Arthur Szlam, Rob Fergus
Many tasks in AI require the collaboration of multiple agents.
3 code implementations • 3 Mar 2016 • Adam Lerer, Sam Gross, Rob Fergus
Wooden blocks are a common toy for infants, allowing them to develop motor skills and gain intuition about the physical behavior of the world.
7 code implementations • 7 Dec 2015 • Bolei Zhou, Yuandong Tian, Sainbayar Sukhbaatar, Arthur Szlam, Rob Fergus
We describe a very simple bag-of-words baseline for visual question answering.
no code implementations • NeurIPS 2015 • Emily L. Denton, Soumith Chintala, Arthur Szlam, Rob Fergus
In this paper we introduce a generative model capable of producing high quality samples of natural images.
1 code implementation • 23 Nov 2015 • Wojciech Zaremba, Tomas Mikolov, Armand Joulin, Rob Fergus
We present an approach for learning simple algorithms such as copying, multi-digit addition and single digit multiplication directly from examples.
2 code implementations • 23 Nov 2015 • Sainbayar Sukhbaatar, Arthur Szlam, Gabriel Synnaeve, Soumith Chintala, Rob Fergus
This paper introduces MazeBase: an environment for simple 2D games, designed as a sandbox for machine learning approaches to reasoning and planning.
no code implementations • 20 Nov 2015 • Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, Manohar Paluri
Over the last few years deep learning methods have emerged as one of the most prominent approaches for video analysis.
1 code implementation • 18 Jun 2015 • Emily Denton, Soumith Chintala, Arthur Szlam, Rob Fergus
In this paper we introduce a generative parametric model capable of producing high quality samples of natural images.
no code implementations • CVPR 2015 • Yunchao Gong, Marcin Pawlowski, Fei Yang, Louis Brandy, Lubomir Bourdev, Rob Fergus
In addition, we propose an online clustering method based on binary k-means that is capable of clustering large photo stream on a single machine, and show applications to spam detection and trending photo discovery.
no code implementations • CVPR 2015 • Li Wan, David Eigen, Rob Fergus
In this paper, we propose a new model that combines these two approaches, obtaining the advantages of each.
no code implementations • ICCV 2015 • Kevin Tang, Manohar Paluri, Li Fei-Fei, Rob Fergus, Lubomir Bourdev
With the widespread availability of cellphones and cameras that have GPS capabilities, it is common for images being uploaded to the Internet today to have GPS coordinates associated with them.
44 code implementations • NeurIPS 2015 • Sainbayar Sukhbaatar, Arthur Szlam, Jason Weston, Rob Fergus
For the former our approach is competitive with Memory Networks, but with less supervision.
Ranked #6 on Question Answering on bAbi
no code implementations • CVPR 2015 • Ning Zhang, Manohar Paluri, Yaniv Taigman, Rob Fergus, Lubomir Bourdev
We explore the task of recognizing peoples' identities in photo albums in an unconstrained setting.
29 code implementations • ICCV 2015 • Du Tran, Lubomir Bourdev, Rob Fergus, Lorenzo Torresani, Manohar Paluri
We propose a simple, yet effective approach for spatiotemporal feature learning using deep 3-dimensional convolutional networks (3D ConvNets) trained on a large scale supervised video dataset.
Ranked #8 on Action Recognition on Sports-1M
Action Recognition In Videos Dynamic Facial Expression Recognition
no code implementations • 19 Nov 2014 • Li Wan, David Eigen, Rob Fergus
In this paper, we propose a new model that combines these two approaches, obtaining the advantages of each.
4 code implementations • ICCV 2015 • David Eigen, Rob Fergus
In this paper we address three different computer vision tasks using a single basic architecture: depth prediction, surface normal estimation, and semantic labeling.
Ranked #74 on Monocular Depth Estimation on NYU-Depth V2 (RMSE metric)
no code implementations • 2 Jul 2014 • Lubomir Bourdev, Fei Yang, Rob Fergus
We train the poselet model on top of PDF features and combine them with object-level CNNs for detection and bounding box prediction.
no code implementations • 9 Jun 2014 • Sainbayar Sukhbaatar, Joan Bruna, Manohar Paluri, Lubomir Bourdev, Rob Fergus
The availability of large labeled datasets has allowed Convolutional Network models to achieve impressive recognition results.
10 code implementations • NeurIPS 2014 • David Eigen, Christian Puhrsch, Rob Fergus
Predicting depth is an essential component in understanding the 3D geometry of a scene.
1 code implementation • NeurIPS 2014 • Wojciech Zaremba, Karol Kurach, Rob Fergus
In this paper we explore how machine learning techniques can be applied to the discovery of efficient mathematical identities.
no code implementations • NeurIPS 2014 • Emily Denton, Wojciech Zaremba, Joan Bruna, Yann Lecun, Rob Fergus
We present techniques for speeding up the test-time evaluation of large convolutional networks, designed for object recognition tasks.
12 code implementations • 21 Dec 2013 • Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, Rob Fergus
Deep neural networks are highly expressive models that have recently achieved state of the art performance on speech and visual recognition tasks.
4 code implementations • 21 Dec 2013 • Pierre Sermanet, David Eigen, Xiang Zhang, Michael Mathieu, Rob Fergus, Yann Lecun
This integrated framework is the winner of the localization task of the ImageNet Large Scale Visual Recognition Challenge 2013 (ILSVRC2013) and obtained very competitive results for the detection and classifications tasks.
no code implementations • 6 Dec 2013 • David Eigen, Jason Rolfe, Rob Fergus, Yann Lecun
A key challenge in designing convolutional network models is sizing them appropriately.
no code implementations • 16 Nov 2013 • Dilip Krishnan, Joan Bruna, Rob Fergus
Blind deconvolution has made significant progress in the past decade.
18 code implementations • 12 Nov 2013 • Matthew D. Zeiler, Rob Fergus
Large Convolutional Network models have recently demonstrated impressive classification performance on the ImageNet benchmark.
1 code implementation • ICML'13: Proceedings of the 30th International Conference on International Conference on Machine Learning - Volume 28 2013 • Li Wan, Matthew Zeiler, Sixin Zhang, Yann Lecun, Rob Fergus
When training with Dropout, a randomly selected subset of activations are set to zero within each layer.
Ranked #6 on Image Classification on MNIST
1 code implementation • 16 Jan 2013 • Matthew D. Zeiler, Rob Fergus
We introduce a simple and effective method for regularizing large convolutional neural networks.
Ranked #39 on Image Classification on SVHN
no code implementations • NeurIPS 2011 • Matthew D. Zeiler, Graham W. Taylor, Leonid Sigal, Iain Matthews, Rob Fergus
We present a type of Temporal Restricted Boltzmann Machine that defines a probability distribution over an output sequence conditional on an input sequence.
no code implementations • NeurIPS 2010 • Graham W. Taylor, Rob Fergus, George Williams, Ian Spiro, Christoph Bregler
We apply our method to challenging real-world data and show that it can generalize beyond hand localization to infer a more general notion of body pose.
no code implementations • NeurIPS 2009 • Rob Fergus, Yair Weiss, Antonio Torralba
With the advent of the Internet it is now possible to collect hundreds of millions of images.
no code implementations • NeurIPS 2009 • Dilip Krishnan, Rob Fergus
In this paper we describe a deconvolution approach that is several orders of magnitude faster than existing techniques that use hyper-Laplacian priors.
no code implementations • NeurIPS 2008 • Yair Weiss, Antonio Torralba, Rob Fergus
Semantic hashing seeks compact binary codes of datapoints so that the Hamming distance between codewords correlates with semantic similarity.