no code implementations • 20 Jun 2023 • Sriram Yenamandra, Arun Ramachandran, Karmesh Yadav, Austin Wang, Mukul Khanna, Theophile Gervet, Tsung-Yen Yang, Vidhi Jain, Alexander William Clegg, John Turner, Zsolt Kira, Manolis Savva, Angel Chang, Devendra Singh Chaplot, Dhruv Batra, Roozbeh Mottaghi, Yonatan Bisk, Chris Paxton
HomeRobot (noun): An affordable compliant robot that navigates homes and manipulates a wide range of objects in order to complete everyday tasks.
no code implementations • 4 Jun 2023 • Sam Powers, Abhinav Gupta, Chris Paxton
Robots in home environments need to be able to learn new skills continuously as data becomes available, becoming ever more capable over time while using as little real-world data as possible.
no code implementations • 6 May 2023 • Wenxuan Zhou, Bowen Jiang, Fan Yang, Chris Paxton, David Held
In this work, we introduce Hybrid Actor-Critic Maps for Manipulation (HACMan), a reinforcement learning approach for 6D non-prehensile manipulation of objects using point cloud observations.
no code implementations • 24 Apr 2023 • Benjamin Bolte, Austin Wang, Jimmy Yang, Mustafa Mukadam, Mrinal Kalakrishnan, Chris Paxton
In order for robots to follow open-ended instructions like "go open the brown cabinet over the sink", they require an understanding of both the scene geometry and the semantics of their environment.
no code implementations • 21 Apr 2023 • Priyam Parashar, Vidhi Jain, Xiaohan Zhang, Jay Vakil, Sam Powers, Yonatan Bisk, Chris Paxton
We see a 4x improvement over baseline in mobile manipulation setting.
no code implementations • ICCV 2023 • Jacob Krantz, Theophile Gervet, Karmesh Yadav, Austin Wang, Chris Paxton, Roozbeh Mottaghi, Dhruv Batra, Jitendra Malik, Stefan Lee, Devendra Singh Chaplot
Our modular method solves sub-tasks of exploration, goal instance re-identification, goal localization, and local navigation.
no code implementations • 8 Nov 2022 • Weiyu Liu, Yilun Du, Tucker Hermans, Sonia Chernova, Chris Paxton
StructDiffusion even improves the success rate of assembling physically-valid structures out of unseen objects by on average 16% over an existing multi-modal transformer model trained on specific structures.
2 code implementations • 11 Oct 2022 • Nur Muhammad Mahi Shafiullah, Chris Paxton, Lerrel Pinto, Soumith Chintala, Arthur Szlam
We propose CLIP-Fields, an implicit scene model that can be used for a variety of tasks, such as segmentation, instance identification, semantic search over space, and view localization.
no code implementations • 19 May 2022 • Yu-Wei Chao, Chris Paxton, Yu Xiang, Wei Yang, Balakumar Sundaralingam, Tao Chen, Adithyavairavan Murali, Maya Cakmak, Dieter Fox
We analyze the performance of a set of baselines and show a correlation with a real-world evaluation.
no code implementations • 11 Apr 2022 • Pratyusha Sharma, Balakumar Sundaralingam, Valts Blukis, Chris Paxton, Tucker Hermans, Antonio Torralba, Jacob Andreas, Dieter Fox
In this paper, we explore natural language as an expressive and flexible tool for robot correction.
no code implementations • 31 Mar 2022 • Wei Yang, Balakumar Sundaralingam, Chris Paxton, Iretiayo Akinola, Yu-Wei Chao, Maya Cakmak, Dieter Fox
However, how to responsively generate smooth motions to take an object from a human is still an open question.
no code implementations • 22 Feb 2022 • Hongtao Wu, Jikai Ye, Xin Meng, Chris Paxton, Gregory Chirikjian
We propose a visual foresight model for pick-and-place rearrangement manipulation which is able to learn efficiently.
1 code implementation • 3 Feb 2022 • Shuang Li, Xavier Puig, Chris Paxton, Yilun Du, Clinton Wang, Linxi Fan, Tao Chen, De-An Huang, Ekin Akyürek, Anima Anandkumar, Jacob Andreas, Igor Mordatch, Antonio Torralba, Yuke Zhu
Together, these results suggest that language modeling induces representations that are useful for modeling not just language, but also goals and plans; these representations can aid learning and generalization even outside of language processing.
no code implementations • CVPR 2022 • Ankit Goyal, Arsalan Mousavian, Chris Paxton, Yu-Wei Chao, Brian Okorn, Jia Deng, Dieter Fox
Accurate object rearrangement from vision is a crucial problem for a wide variety of real-world robotics applications in unstructured environments.
no code implementations • 9 Nov 2021 • Andreea Bobu, Chris Paxton, Wei Yang, Balakumar Sundaralingam, Yu-Wei Chao, Maya Cakmak, Dieter Fox
Second, we treat this low-dimensional concept as an automatic labeler to synthesize a large-scale high-dimensional data set with the simulator.
no code implementations • 19 Oct 2021 • Weiyu Liu, Chris Paxton, Tucker Hermans, Dieter Fox
Geometric organization of objects into semantically meaningful arrangements pervades the built world.
1 code implementation • 8 Sep 2021 • Wentao Yuan, Chris Paxton, Karthik Desingh, Dieter Fox
Sequential manipulation tasks require a robot to perceive the state of an environment and plan a sequence of actions leading to a desired goal state.
1 code implementation • 26 Aug 2021 • Chris Paxton, Chris Xie, Tucker Hermans, Dieter Fox
We further demonstrate the ability of our planner to generate and execute diverse manipulation plans through a set of real-world experiments with a variety of objects.
2 code implementations • 26 Jul 2021 • Jesse Thomason, Mohit Shridhar, Yonatan Bisk, Chris Paxton, Luke Zettlemoyer
We introduce several CLIP-based models for distinguishing objects and demonstrate that while recent advances in jointly modeling vision and language are useful for robotic language understanding, it is still the case that these image-based models are weaker at understanding the 3D nature of objects -- properties which play a key role in manipulation.
1 code implementation • 12 Jul 2021 • Valts Blukis, Chris Paxton, Dieter Fox, Animesh Garg, Yoav Artzi
Natural language provides an accessible and expressive interface to specify long-term tasks for robotic agents.
no code implementations • 2 Jun 2021 • Ahmed H. Qureshi, Arsalan Mousavian, Chris Paxton, Michael C. Yip, Dieter Fox
We propose NeRP (Neural Rearrangement Planning), a deep learning based approach for multi-step neural object rearrangement planning which works with never-before-seen objects, that is trained on simulation data, and generalizes to the real world.
no code implementations • 17 Nov 2020 • Shohin Mukherjee, Chris Paxton, Arsalan Mousavian, Adam Fishman, Maxim Likhachev, Dieter Fox
Zero-shot execution of unseen robotic tasks is important to allowing robots to perform a wide variety of tasks in human environments, but collecting the amounts of data necessary to train end-to-end policies in the real-world is often infeasible.
no code implementations • 17 Nov 2020 • Wei Yang, Chris Paxton, Arsalan Mousavian, Yu-Wei Chao, Maya Cakmak, Dieter Fox
We demonstrate the generalizability, usability, and robustness of our approach on a novel benchmark set of 26 diverse household objects, a user study with naive users (N=6) handing over a subset of 15 objects, and a systematic evaluation examining different ways of handing objects.
no code implementations • 12 Mar 2020 • Wei Yang, Chris Paxton, Maya Cakmak, Dieter Fox
In this paper, we propose an approach for human-to-robot handovers in which the robot meets the human halfway, by classifying the human's grasp of the object and quickly planning a trajectory accordingly to take the object from the human's hand according to their intent.
no code implementations • 8 Mar 2020 • Kei Kase, Chris Paxton, Hammad Mazhar, Tetsuya OGATA, Dieter Fox
On the other hand, symbolic planning methods such as STRIPS have long been able to solve new problems given only a domain definition and a symbolic goal, but these approaches often struggle on the real world robotic tasks due to the challenges of grounding these symbols from sensor data in a partially-observable world.
1 code implementation • 8 Dec 2019 • Adithyavairavan Murali, Arsalan Mousavian, Clemens Eppner, Chris Paxton, Dieter Fox
Grasping in cluttered environments is a fundamental but challenging robotic skill.
no code implementations • 13 Nov 2019 • De-An Huang, Yu-Wei Chao, Chris Paxton, Xinke Deng, Li Fei-Fei, Juan Carlos Niebles, Animesh Garg, Dieter Fox
We further show that by using the automatically inferred goal from the video demonstration, our robot is able to reproduce the same task in a real kitchen environment.
1 code implementation • 11 Nov 2019 • Caelan Reed Garrett, Chris Paxton, Tomás Lozano-Pérez, Leslie Pack Kaelbling, Dieter Fox
To solve multi-step manipulation tasks in the real world, an autonomous robot must take actions to observe its environment and react to unexpected observations.
no code implementations • 16 Oct 2019 • Junha Roh, Chris Paxton, Andrzej Pronobis, Ali Farhadi, Dieter Fox
Widespread adoption of self-driving cars will depend not only on their safety but largely on their ability to interact with human users.
1 code implementation • 25 Sep 2019 • Andrew Hundt, Benjamin Killeen, Nicholas Greene, Hongtao Wu, Heeyeon Kwon, Chris Paxton, Gregory D. Hager
We are able to create real stacks in 100% of trials with 61% efficiency and real rows in 100% of trials with 59% efficiency by directly loading the simulation-trained model on the real robot with no additional real-world fine-tuning.
no code implementations • 20 Mar 2019 • Chris Paxton, Yonatan Bisk, Jesse Thomason, Arunkumar Byravan, Dieter Fox
High-level human instructions often correspond to behaviors with multiple implicit steps.
3 code implementations • 27 Oct 2018 • Andrew Hundt, Varun Jain, Chia-Hung Lin, Chris Paxton, Gregory D. Hager
We show that a mild relaxation of the task and workspace constraints implicit in existing object grasping datasets can cause neural network based grasping algorithms to fail on even a simple block stacking task when executed under more realistic circumstances.
1 code implementation • 30 Mar 2018 • Chris Paxton, Yotam Barnoy, Kapil Katyal, Raman Arora, Gregory D. Hager
In this work, we propose a neural network architecture and associated planning algorithm that (1) learns a representation of the world useful for generating prospective futures after the application of high-level actions, (2) uses this generative model to simulate the result of sequences of high-level actions in a variety of environments, and (3) uses this same representation to evaluate these actions and perform tree search to find a sequence of high-level actions in a new environment.
no code implementations • 6 Mar 2018 • Kapil Katyal, Katie Popek, Chris Paxton, Joseph Moore, Kevin Wolfe, Philippe Burlina, Gregory D. Hager
In these situations, the robot's ability to reason about its future motion is often severely limited by sensor field of view (FOV).
no code implementations • 8 Nov 2017 • Chris Paxton, Kapil Katyal, Christian Rupprecht, Raman Arora, Gregory D. Hager
Ideally, we would combine the ability of machine learning to leverage big data for learning the semantics of a task, while using techniques from task planning to reliably generalize to new environment.
2 code implementations • 11 Oct 2017 • Felix Jonathan, Chris Paxton, Gregory D. Hager
Accurate knowledge of object poses is crucial to successful robotic manipulation tasks, and yet most current approaches only work in laboratory settings.
Robotics
no code implementations • 22 Mar 2017 • Chris Paxton, Vasumathi Raman, Gregory D. Hager, Marin Kobilarov
This paper investigates the ability of neural networks to learn both LTL constraints and control policies in order to generate task plans in complex environments.
Robotics