no code implementations • 28 Nov 2024 • Ahmed Jaafar, Shreyas Sundara Raman, Yichen Wei, Sudarshan Harithas, Sofia Juliani, Anneke Wernerfelt, Benedict Quartey, Ifrah Idrees, Jason Xinyu Liu, Stefanie Tellex
Efficiently learning and executing long-horizon mobile manipulation (MoMa) tasks is crucial for advancing robotics in household and workplace settings.
no code implementations • 18 Oct 2024 • Rachel Ma, Lyndon Lam, Benjamin A. Spiegel, Aditya Ganeshan, Roma Patel, Ben Abbatematteo, David Paulius, Stefanie Tellex, George Konidaris
Our method allows a robot to generate a trajectory for a novel object based on a verb, which can then be used as input to a motion planner.
no code implementations • 17 Sep 2024 • Lance Ying, Jason Xinyu Liu, Shivam Aarya, Yizirui Fang, Stefanie Tellex, Joshua B. Tenenbaum, Tianmin Shu
We present a cognitively inspired model, Speech Instruction Following through Theory of Mind (SIFToM), to enable robots to pragmatically follow human instructions under diverse speech conditions by inferring the human's goal and joint plan as prior for speech perception and understanding.
no code implementations • 21 May 2024 • Vanya Cohen, Jason Xinyu Liu, Raymond Mooney, Stefanie Tellex, David Watkins
With large language models, robots can understand language more flexibly and more capable than ever before.
no code implementations • 1 Apr 2024 • Casey Kennington, Malihe Alikhani, Heather Pon-Barry, Katherine Atwell, Yonatan Bisk, Daniel Fried, Felix Gervits, Zhao Han, Mert Inan, Michael Johnston, Raj Korpan, Diane Litman, Matthew Marge, Cynthia Matuszek, Ross Mead, Shiwali Mohan, Raymond Mooney, Natalie Parde, Jivko Sinapov, Angela Stewart, Matthew Stone, Stefanie Tellex, Tom Williams
The ability to interact with machines using natural human language is becoming not just commonplace, but expected.
no code implementations • 18 Feb 2024 • Benedict Quartey, Eric Rosen, Stefanie Tellex, George Konidaris
We propose Language Instruction grounding for Motion Planning (LIMP), an approach that enables robots to verifiably follow expressive and complex open-ended instructions in real-world environments without prebuilt semantic maps.
no code implementations • 3 Oct 2023 • Ifrah Idrees, Tian Yun, Naveen Sharma, Yunxin Deng, Nakul Gopalan, George Konidaris, Stefanie Tellex
We propose a novel framework for plan and goal recognition in partially observable domains -- Dialogue for Goal Recognition (D4GR) enabling a robot to rectify its belief in human progress by asking clarification questions about noisy sensor data and sub-optimal human actions.
no code implementations • 18 Sep 2023 • ZiYi Yang, Shreyas S. Raman, Ankit Shah, Stefanie Tellex
Recent advancements in large language models (LLMs) have enabled a new research domain, LLM agents, for solving robotics and planning tasks by leveraging the world knowledge and general reasoning abilities of LLMs obtained during pretraining.
no code implementations • 13 Sep 2023 • Thao Nguyen, Vladislav Hrosinkov, Eric Rosen, Stefanie Tellex
In this work, we bridge the gap in realistic object search by posing the search problem as a partially observable Markov decision process (POMDP) where the object detector and visual sensor noise in the observation model is determined by a single Deep Neural Network conditioned on complex language descriptions.
1 code implementation • 6 Mar 2023 • Kaiyu Zheng, Anirudha Paul, Stefanie Tellex
GenMOS takes as input point cloud observations of the local region, object detection results, and localization of the robot's view pose, and outputs a 6D viewpoint to move to through online planning.
no code implementations • 22 Feb 2023 • Jason Xinyu Liu, ZiYi Yang, Ifrah Idrees, Sam Liang, Benjamin Schornstein, Stefanie Tellex, Ankit Shah
We propose Lang2LTL, a modular system and a software package that leverages large language models (LLMs) to ground temporal navigational commands to LTL specifications in environments without prior language data.
no code implementations • 17 Nov 2022 • Shreyas Sundara Raman, Vanya Cohen, Ifrah Idrees, Eric Rosen, Ray Mooney, Stefanie Tellex, David Paulius
Our improvements transfer to a Boston Dynamics Spot robot initialized with a set of skills (specified in language) and associated preconditions, where CAPE improves the correctness metric of the executed task plans by 76. 49% compared to SayCan.
no code implementations • 12 Aug 2022 • Rafael Rodriguez-Sanchez, Benjamin A. Spiegel, Jennifer Wang, Roma Patel, Stefanie Tellex, George Konidaris
We define precise syntax and grounding semantics for RLang, and provide a parser that grounds RLang programs to an algorithm-agnostic \textit{partial} world model and policy that can be exploited by an RL agent.
no code implementations • 25 Oct 2021 • Ifrah Idrees, Zahid Hasan, Steven P. Reiss, Stefanie Tellex
Robots equipped with situational awareness can help humans efficiently find their lost objects by leveraging spatial and temporal structure.
1 code implementation • 19 Oct 2021 • Kaiyu Zheng, Rohan Chitnis, Yoonchang Sung, George Konidaris, Stefanie Tellex
In realistic applications of object search, robots will need to locate target objects in complex environments while coping with unreliable sensors, especially for small or hard-to-detect objects.
no code implementations • 11 Oct 2021 • Eric Hsiung, Hiloni Mehta, Junchi Chu, Xinyu Liu, Roma Patel, Stefanie Tellex, George Konidaris
We compare our method of mapping natural language task specifications to intermediate contextual queries against state-of-the-art CopyNet models capable of translating natural language to LTL, by evaluating whether correct LTL for manipulation and navigation task specifications can be output, and show that our method outperforms the CopyNet model on unseen object references.
no code implementations • 28 Jul 2021 • Sreehari Rammohan, Shangqun Yu, Bowen He, Eric Hsiung, Eric Rosen, Stefanie Tellex, George Konidaris
Learning continuous control in high-dimensional sparse reward settings, such as robotic manipulation, is a challenging problem due to the number of samples often required to obtain accurate optimal value and policy estimates.
no code implementations • 22 Jul 2021 • Monica Roy, Kaiyu Zheng, Jason Liu, Stefanie Tellex
To this end, we introduce a new task, dialogue object search: A robot is tasked to search for a target object (e. g. fork) in a human environment (e. g., kitchen), while engaging in a "video call" with a remote human who has additional but inexact knowledge about the target's location.
no code implementations • 12 Jan 2021 • Ben Abbatematteo, Eric Rosen, Stefanie Tellex, George Konidaris
We propose using kinematic motion planning as a completely autonomous, sample efficient way to bootstrap motor skill learning for object manipulation.
1 code implementation • 4 Dec 2020 • Kaiyu Zheng, Deniz Bayazit, Rebecca Mathew, Ellie Pavlick, Stefanie Tellex
We propose SLOOP (Spatial Language Object-Oriented POMDP), a new framework for partially observable decision making with a probabilistic observation model for spatial language.
no code implementations • 17 Oct 2020 • Michael Fishman, Nishanth Kumar, Cameron Allen, Natasha Danas, Michael Littman, Stefanie Tellex, George Konidaris
Unfortunately, planning to solve any specific task using an open-scope model is computationally intractable - even for state-of-the-art methods - due to the many states and actions that are necessarily present in the model but irrelevant to that problem.
1 code implementation • ECCV 2020 • Atsunobu Kotani, Stefanie Tellex, James Tompkin
Instead, we introduce the Decoupled Style Descriptor (DSD) model for handwriting, which factors both character- and writer-level styles and allows our model to represent an overall greater space of styles.
1 code implementation • 23 Jun 2020 • Thao Nguyen, Nakul Gopalan, Roma Patel, Matt Corsaro, Ellie Pavlick, Stefanie Tellex
The model takes in a language command containing a verb, for example "Hand me something to cut," and RGB images of candidate objects and selects the object that best satisfies the task specified by the verb.
1 code implementation • 6 May 2020 • Kaiyu Zheng, Yoonchang Sung, George Konidaris, Stefanie Tellex
Robots operating in households must find objects on shelves, under tables, and in cupboards.
1 code implementation • 21 Apr 2020 • Kaiyu Zheng, Stefanie Tellex
In this paper, we present pomdp_py, a general purpose Partially Observable Markov Decision Process (POMDP) library written in Python and Cython.
2 code implementations • 23 Oct 2019 • Jonathan Chang, Nishanth Kumar, Sean Hastings, Aaron Gokaslan, Diego Romeres, Devesh Jha, Daniel Nikovski, George Konidaris, Stefanie Tellex
We demonstrate that our model trained on 33% of the possible goals is able to generalize to more than 90% of the targets in the scene for both simulation and robot experiments.
no code implementations • 30 May 2019 • Vanya Cohen, Benjamin Burchfiel, Thao Nguyen, Nakul Gopalan, Stefanie Tellex, George Konidaris
Our system is able to disambiguate between novel objects, observed via depth images, based on natural language descriptions.
2 code implementations • 28 May 2019 • Yoonseon Oh, Roma Patel, Thao Nguyen, Baichuan Huang, Ellie Pavlick, Stefanie Tellex
Often times, we specify tasks for a robot using temporal language that can also span different levels of abstraction.
no code implementations • 29 Nov 2017 • Thomas Kollar, Stefanie Tellex, Matthew Walter, Albert Huang, Abraham Bachrach, Sachi Hemachandra, Emma Brunskill, Ashis Banerjee, Deb Roy, Seth Teller, Nicholas Roy
Symbolic models capture linguistic structure but have not scaled successfully to handle the diverse language produced by untrained users.
1 code implementation • 20 Nov 2017 • Melrose Roderick, James Macglashan, Stefanie Tellex
The Deep Q-Network proposed by Mnih et al. [2015] has become a benchmark and building point for much deep reinforcement learning research.
no code implementations • 2 Oct 2017 • Melrose Roderick, Christopher Grimm, Stefanie Tellex
We examine the problem of learning and planning on high-dimensional domains with long horizons and sparse rewards.
no code implementations • 31 Jul 2017 • Lucas Lehnert, Stefanie Tellex, Michael L. Littman
One question central to Reinforcement Learning is how to learn a feature representation that supports algorithm scaling and re-use of learned information from different tasks.
1 code implementation • WS 2017 • Siddharth Karamcheti, Edward C. Williams, Dilip Arumugam, Mina Rhee, Nakul Gopalan, Lawson L. S. Wong, Stefanie Tellex
Robots operating alongside humans in diverse, stochastic environments must be able to accurately interpret natural language commands.
1 code implementation • 21 Apr 2017 • Dilip Arumugam, Siddharth Karamcheti, Nakul Gopalan, Lawson L. S. Wong, Stefanie Tellex
In this work, by grounding commands to all the tasks or subtasks available in a hierarchical planning framework, we arrive at a model capable of interpreting language at multiple levels of specificity ranging from coarse to more granular.