1 code implementation • 11 Dec 2024 • Chongyi Zheng, Jens Tuyls, Joanne Peng, Benjamin Eysenbach
Self-supervised learning has the potential of lifting several of the key challenges in reinforcement learning today, such as exploration, representation learning, and reward design.
no code implementations • 24 Jan 2024 • Alex Zhang, Khanh Nguyen, Jens Tuyls, Albert Lin, Karthik Narasimhan
In this work, we take initial steps in developing robust LWMs that can generalize to compositionally novel language descriptions.
1 code implementation • 18 Jul 2023 • Jens Tuyls, Dhruv Madeka, Kari Torkkola, Dean Foster, Karthik Narasimhan, Sham Kakade
Inspired by recent work in Natural Language Processing (NLP) where "scaling up" has resulted in increasingly more capable LLMs, we investigate whether carefully scaling up model and data size can bring similar improvements in the imitation learning setting for single-agent games.
1 code implementation • ICLR 2022 • Jens Tuyls, Shunyu Yao, Sham Kakade, Karthik Narasimhan
Text adventure games present unique challenges to reinforcement learning methods due to their combinatorially large action spaces and sparse rewards.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Junlin Wang, Jens Tuyls, Eric Wallace, Sameer Singh
Gradient-based analysis methods, such as saliency map visualizations and adversarial input perturbations, have found widespread use in interpreting neural NLP models due to their simplicity, flexibility, and most importantly, their faithfulness.
no code implementations • EMNLP (PrivateNLP) 2020 • Gavin Kerrigan, Dylan Slack, Jens Tuyls
Language modeling is a keystone task in natural language processing.
no code implementations • 3 Jul 2020 • Griffin Mooers, Jens Tuyls, Stephan Mandt, Michael Pritchard, Tom Beucler
While cloud-resolving models can explicitly simulate the details of small-scale storm formation and morphology, these details are often ignored by climate models for lack of computational resources.
1 code implementation • IJCNLP 2019 • Eric Wallace, Jens Tuyls, Junlin Wang, Sanjay Subramanian, Matt Gardner, Sameer Singh
Neural NLP models are increasingly accurate but are imperfect and opaque---they break in counterintuitive ways and leave end users puzzled at their behavior.