no code implementations • 28 Oct 2024 • Reece Shuttleworth, Jacob Andreas, Antonio Torralba, Pratyusha Sharma
Second, we show that LoRA models with intruder dimensions, despite achieving similar performance to full fine-tuning on the target task, become worse models of the pre-training distribution and adapt less robustly to multiple tasks sequentially.
no code implementations • 26 Feb 2024 • Haotian Fu, Pratyusha Sharma, Elias Stengel-Eskin, George Konidaris, Nicolas Le Roux, Marc-Alexandre Côté, Xingdi Yuan
We present an algorithm for skill discovery from expert demonstrations.
no code implementations • CVPR 2024 • Pratyusha Sharma, Tamar Rott Shaham, Manel Baradad, Stephanie Fu, Adrian Rodriguez-Munoz, Shivam Duggal, Phillip Isola, Antonio Torralba
Although LLM-generated images do not look like natural images, results on image generation and the ability of models to correct these generated images indicate that precise modeling of strings can teach language models about numerous aspects of the visual world.
1 code implementation • 21 Dec 2023 • Pratyusha Sharma, Jordan T. Ash, Dipendra Misra
Transformer-based Large Language Models (LLMs) have become a fixture in modern machine learning.
no code implementations • 13 Dec 2023 • Lionel Wong, Jiayuan Mao, Pratyusha Sharma, Zachary S. Siegel, Jiahai Feng, Noa Korneev, Joshua B. Tenenbaum, Jacob Andreas
Effective planning in the real world requires not only world knowledge, but the ability to leverage that knowledge to build the right representation of the task at hand.
1 code implementation • 29 Oct 2023 • Shikhar Murty, Pratyusha Sharma, Jacob Andreas, Christopher D. Manning
Recursion is a prominent feature of human language, and fundamentally challenging for self-attention due to the lack of an explicit recursive-state tracking mechanism.
no code implementations • 18 Oct 2023 • Shikhar Murty, Orr Paradise, Pratyusha Sharma
With large language models surpassing human performance on an increasing number of benchmarks, we must take a principled approach for targeted evaluation of model capabilities.
1 code implementation • 30 May 2023 • Shikhar Murty, Pratyusha Sharma, Jacob Andreas, Christopher D. Manning
When analyzing the relationship between model-internal properties and grokking, we find that optimal depth for grokking can be identified using the tree-structuredness metric of \citet{murty2023projections}.
1 code implementation • 3 Feb 2023 • Belinda Z. Li, William Chen, Pratyusha Sharma, Jacob Andreas
Language models trained on large text corpora encode rich distributional information about real-world environments and action sequences.
no code implementations • 2 Nov 2022 • Shikhar Murty, Pratyusha Sharma, Jacob Andreas, Christopher D. Manning
To evaluate this possibility, we describe an unsupervised and parameter-free method to \emph{functionally project} the behavior of any transformer into the space of tree-structured networks.
no code implementations • 11 Apr 2022 • Pratyusha Sharma, Balakumar Sundaralingam, Valts Blukis, Chris Paxton, Tucker Hermans, Antonio Torralba, Jacob Andreas, Dieter Fox
In this paper, we explore natural language as an expressive and flexible tool for robot correction.
no code implementations • ACL 2022 • Pratyusha Sharma, Antonio Torralba, Jacob Andreas
We evaluate this approach in the ALFRED household simulation environment, providing natural language annotations for only 10% of demonstrations.
no code implementations • CVPR 2021 • Yiyue Luo, Yunzhu Li, Michael Foshey, Wan Shou, Pratyusha Sharma, Tomas Palacios, Antonio Torralba, Wojciech Matusik
In this work, leveraging such tactile interactions, we propose a 3D human pose estimation approach using the pressure maps recorded by a tactile carpet as input.
no code implementations • 17 Apr 2021 • Jacob Andreas, Gašper Beguš, Michael M. Bronstein, Roee Diamant, Denley Delaney, Shane Gero, Shafi Goldwasser, David F. Gruber, Sarah de Haas, Peter Malkin, Roger Payne, Giovanni Petri, Daniela Rus, Pratyusha Sharma, Dan Tchernov, Pernille Tønnesen, Antonio Torralba, Daniel Vogt, Robert J. Wood
We posit that machine learning will be the cornerstone of future collection, processing, and analysis of multimodal streams of data in animal communication studies, including bioacoustic, behavioral, biological, and environmental data.
1 code implementation • NeurIPS 2019 • Pratyusha Sharma, Deepak Pathak, Abhinav Gupta
We study a generalized setup for learning from demonstration to build an agent that can manipulate novel objects in unseen scenarios by looking at only a single video of human demonstration from a third-person perspective.
no code implementations • 16 Oct 2018 • Pratyusha Sharma, Lekha Mohan, Lerrel Pinto, Abhinav Gupta
In order to make progress and capture the space of manipulation, we would need to collect a large-scale dataset of diverse tasks such as pouring, opening bottles, stacking objects etc.