1 code implementation • 20 Nov 2022 • Micah Carroll, Orr Paradise, Jessy Lin, Raluca Georgescu, Mingfei Sun, David Bignell, Stephanie Milani, Katja Hofmann, Matthew Hausknecht, Anca Dragan, Sam Devlin
Randomly masking and predicting word tokens has been a successful approach in pre-training language models for a variety of downstream tasks.
1 code implementation • 15 Aug 2022 • Nolan Wagener, Andrey Kolobov, Felipe Vieira Frujeri, Ricky Loynd, Ching-An Cheng, Matthew Hausknecht
We demonstrate the utility of MoCapAct by using it to train a single hierarchical policy capable of tracking the entire MoCap dataset within dm_control and show the learned low-level component can be re-used to efficiently learn downstream high-level tasks.
no code implementations • 28 Apr 2022 • Micah Carroll, Jessy Lin, Orr Paradise, Raluca Georgescu, Mingfei Sun, David Bignell, Stephanie Milani, Katja Hofmann, Matthew Hausknecht, Anca Dragan, Sam Devlin
Randomly masking and predicting word tokens has been a successful approach in pre-training language models for a variety of downstream tasks.
no code implementations • 9 Mar 2022 • Nathaniel Weir, Xingdi Yuan, Marc-Alexandre Côté, Matthew Hausknecht, Romain Laroche, Ida Momennejad, Harm van Seijen, Benjamin Van Durme
Humans have the capability, aided by the expressive compositionality of their language, to learn quickly by demonstration.
no code implementations • 23 Feb 2022 • Matthew Hausknecht, Nolan Wagener
Dropout has long been a staple of supervised learning, but is rarely used in reinforcement learning.
no code implementations • 29 Mar 2021 • Sharada Mohanty, Jyotish Poonganam, Adrien Gaidon, Andrey Kolobov, Blake Wulfe, Dipam Chakraborty, Gražvydas Šemetulskis, João Schapke, Jonas Kubilius, Jurgis Pašukonis, Linas Klimas, Matthew Hausknecht, Patrick MacAlpine, Quang Nhat Tran, Thomas Tumiel, Xiaocheng Tang, Xinwei Chen, Christopher Hesse, Jacob Hilton, William Hebgen Guss, Sahika Genc, John Schulman, Karl Cobbe
We present the design of a centralized benchmark for Reinforcement Learning which can help measure Sample Efficiency and Generalization in Reinforcement Learning by doing end to end evaluation of the training and rollout phases of thousands of user submitted code bases in a scalable way.
no code implementations • NAACL 2021 • Shunyu Yao, Karthik Narasimhan, Matthew Hausknecht
Text-based games simulate worlds and interact with players using natural language.
no code implementations • ICLR 2021 • Mohit Shridhar, Xingdi Yuan, Marc-Alexandre Cote, Yonatan Bisk, Adam Trischler, Matthew Hausknecht
ALFWorld enables the creation of a new BUTLER agent whose abstract knowledge, learned in TextWorld, corresponds directly to concrete, visually grounded actions.
1 code implementation • 8 Oct 2020 • Mohit Shridhar, Xingdi Yuan, Marc-Alexandre Côté, Yonatan Bisk, Adam Trischler, Matthew Hausknecht
ALFWorld enables the creation of a new BUTLER agent whose abstract knowledge, learned in TextWorld, corresponds directly to concrete, visually grounded actions.
1 code implementation • EMNLP 2020 • Shunyu Yao, Rohan Rao, Matthew Hausknecht, Karthik Narasimhan
In this paper, we propose the Contextual Action Language Model (CALM) to generate a compact set of action candidates at each game state.
1 code implementation • 12 Jun 2020 • Prithviraj Ammanabrolu, Ethan Tien, Matthew Hausknecht, Mark O. Riedl
Text-based games are long puzzles or quests, characterized by a sequence of sparse and potentially deceptive rewards.
1 code implementation • ICLR 2020 • Prithviraj Ammanabrolu, Matthew Hausknecht
Interactive Fiction games are text-based simulations in which an agent interacts with the world purely through natural language.
no code implementations • ICML 2020 • Ricky Loynd, Roland Fernandez, Asli Celikyilmaz, Adith Swaminathan, Matthew Hausknecht
Transformers have increasingly outperformed gated RNNs in obtaining new state-of-the-art results on supervised tasks involving text sequences.
2 code implementations • ICML 2020 • Eric Zhan, Albert Tseng, Yisong Yue, Adith Swaminathan, Matthew Hausknecht
We study the problem of controllable generation of long-term sequential behaviors, where the goal is to calibrate to multiple behavior styles simultaneously.
2 code implementations • 11 Sep 2019 • Matthew Hausknecht, Prithviraj Ammanabrolu, Marc-Alexandre Côté, Xingdi Yuan
A hallmark of human intelligence is the ability to understand and communicate with language.
no code implementations • 5 Apr 2019 • Ishan Durugkar, Matthew Hausknecht, Adith Swaminathan, Patrick MacAlpine
Policy gradient algorithms typically combine discounted future rewards with an estimated value function, to compute the direction and magnitude of parameter updates.
no code implementations • 1 Apr 2019 • Jack W. Stokes, Rakshit Agrawal, Geoff McDonald, Matthew Hausknecht
We use the Convoluted Partitioning of Long Sequences (CPoLS) model, which processes Javascript files as byte sequences.
1 code implementation • 12 Feb 2019 • Matthew Hausknecht, Ricky Loynd, Greg Yang, Adith Swaminathan, Jason D. Williams
Interactive Fiction (IF) games are complex textual decision making problems.
2 code implementations • 29 Jun 2018 • Xingdi Yuan, Marc-Alexandre Côté, Alessandro Sordoni, Romain Laroche, Remi Tachet des Combes, Matthew Hausknecht, Adam Trischler
We propose a recurrent RL agent with an episodic exploration mechanism that helps discovering good policies in text-based game environments.
1 code implementation • 29 Jun 2018 • Marc-Alexandre Côté, Ákos Kádár, Xingdi Yuan, Ben Kybartas, Tavian Barnes, Emery Fine, James Moore, Ruo Yu Tao, Matthew Hausknecht, Layla El Asri, Mahmoud Adada, Wendy Tay, Adam Trischler
We introduce TextWorld, a sandbox learning environment for the training and evaluation of RL agents on text-based games.
no code implementations • ICLR 2018 • Rudy Bunel, Matthew Hausknecht, Jacob Devlin, Rishabh Singh, Pushmeet Kohli
Program synthesis is the task of automatically generating a program consistent with a specification.
no code implementations • ICLR 2018 • Ricky Loynd, Matthew Hausknecht, Lihong Li, Li Deng
Humans rely on episodic memory constantly, in remembering the name of someone they met 10 minutes ago, the plot of a movie as it unfolds, or where they parked the car.
no code implementations • NeurIPS 2017 • Jacob Devlin, Rudy Bunel, Rishabh Singh, Matthew Hausknecht, Pushmeet Kohli
In our first proposal, portfolio adaptation, a set of induction models is pretrained on a set of related tasks, and the best model is adapted towards the new task using transfer learning.
6 code implementations • 18 Sep 2017 • Marlos C. Machado, Marc G. Bellemare, Erik Talvitie, Joel Veness, Matthew Hausknecht, Michael Bowling
The Arcade Learning Environment (ALE) is an evaluation platform that poses the challenge of building AI agents with general competency across dozens of Atari 2600 games.
7 code implementations • 13 Nov 2015 • Matthew Hausknecht, Peter Stone
Recent work has shown that deep neural networks are capable of approximating both value functions and policies in reinforcement learning domains featuring continuous state and action spaces.
5 code implementations • 23 Jul 2015 • Matthew Hausknecht, Peter Stone
Deep Reinforcement Learning has yielded proficient controllers for complex tasks.
1 code implementation • CVPR 2015 • Joe Yue-Hei Ng, Matthew Hausknecht, Sudheendra Vijayanarasimhan, Oriol Vinyals, Rajat Monga, George Toderici
Convolutional neural networks (CNNs) have been extensively applied for image recognition problems giving state-of-the-art results on recognition, detection, segmentation and retrieval.
Ranked #5 on
Action Recognition
on Sports-1M