1 code implementation • 25 Feb 2023 • Tarasha Khurana, Peiyun Hu, David Held, Deva Ramanan
One promising self-supervised task is 3D point cloud forecasting from unannotated LiDAR sequences.
no code implementations • 24 Feb 2023 • Siddharth Ancha, Gaurav Pathak, Ji Zhang, Srinivasa Narasimhan, David Held
To navigate in an environment safely and autonomously, robots must accurately estimate where obstacles are and how they move.
no code implementations • 19 Feb 2023 • Zixuan Huang, Xingyu Lin, David Held
In this work, we propose a self-supervised method to finetune a mesh reconstruction model in the real world.
no code implementations • 21 Nov 2022 • Brian Okorn, Chuer Pan, Martial Hebert, David Held
While self-supervised learning has been used successfully for translational object keypoints, in this work, we show that naively applying relative supervision to the rotational group $SO(3)$ will often fail to converge due to the non-convexity of the rotational space.
no code implementations • 2 Nov 2022 • Wenxuan Zhou, David Held
Previous work in extrinsic dexterity usually has careful assumptions about contacts which impose restrictions on robot design, robot motions, and the variations of the physical parameters.
no code implementations • 27 Oct 2022 • Xingyu Lin, Carl Qi, Yunchu Zhang, Zhiao Huang, Katerina Fragkiadaki, Yunzhu Li, Chuang Gan, David Held
Effective planning of long-horizon deformable object manipulation requires suitable abstractions at both the spatial and temporal levels.
1 code implementation • 4 Oct 2022 • Tarasha Khurana, Peiyun Hu, Achal Dave, Jason Ziglar, David Held, Deva Ramanan
Self-supervised representations proposed for large-scale planning, such as ego-centric freespace, confound these two motions, making the representation difficult to use for downstream motion planners.
no code implementations • 19 Sep 2022 • Alberta Longhini, Marco Moletta, Alfredo Reichlin, Michael C. Welle, David Held, Zackory Erickson, Danica Kragic
In particular, we leverage a latent representation of elastic physical properties of cloth-like deformable objects which we explore through a pulling interaction.
no code implementations • 11 Aug 2022 • YuFei Wang, David Held, Zackory Erickson
Robotic manipulation of highly deformable cloth presents a promising opportunity to assist people with several daily tasks, such as washing dishes; folding laundry; or dressing, bathing, and hygiene assistance for individuals with severe motor impairments.
no code implementations • 6 Jun 2022 • Zixuan Huang, Xingyu Lin, David Held
We evaluate our system both on cloth flattening as well as on cloth canonicalization, in which the objective is to manipulate the cloth into a canonical pose.
no code implementations • 9 May 2022 • Ben Eisner, Harry Zhang, David Held
We propose a vision-based system that learns to predict the potential motions of the parts of a variety of articulated objects to guide downstream motion planning of the system to articulate the objects.
no code implementations • ICLR 2022 • Xingyu Lin, Zhiao Huang, Yunzhu Li, Joshua B. Tenenbaum, David Held, Chuang Gan
We consider the problem of sequential robotic manipulation of deformable objects using tools.
no code implementations • 3 Mar 2022 • Gautham Narayan Narasimhan, Kai Zhang, Ben Eisner, Xingyu Lin, David Held
Liquid state estimation is important for robotics tasks such as pouring; however, estimating the state of transparent liquids is a challenging problem.
no code implementations • 1 Feb 2022 • Jianren Wang, Haiming Gang, Siddharth Ancha, Yi-Ting Chen, David Held
However, these detectors usually require training on large amounts of annotated data that is expensive and time-consuming to collect.
no code implementations • 18 Jan 2022 • Qiao Gu, Brian Okorn, David Held
In this paper, we propose the OSSID framework, leveraging a slow zero-shot pose estimator to self-supervise the training of a fast detection algorithm.
1 code implementation • 21 Nov 2021 • Himangi Mittal, Brian Okorn, Arpit Jangid, David Held
The aim of this work is to learn to complete these partial point clouds, giving us a full understanding of the object's geometry using only partial observations.
1 code implementation • 10 Nov 2021 • Thomas Weng, Sujay Bajracharya, YuFei Wang, Khush Agrawal, David Held
We introduce FabricFlowNet (FFN), a cloth manipulation policy that leverages flow as both an input and as an action representation to improve performance.
no code implementations • 8 Jul 2021 • Siddharth Ancha, Gaurav Pathak, Srinivasa G. Narasimhan, David Held
We use light curtains to estimate the safety envelope of a scene: a hypothetical surface that separates the robot from all obstacles.
no code implementations • CVPR 2021 • Yaadhav Raaj, Siddharth Ancha, Robert Tamburo, David Held, Srinivasa G. Narasimhan
Active sensing through the use of Adaptive Depth Sensors is a nascent field, with potential in areas such as Advanced driver-assistance systems (ADAS).
1 code implementation • CVPR 2021 • Peiyun Hu, Aaron Huang, John Dolan, David Held, Deva Ramanan
Finally, we propose future freespace as an additional source of annotation-free supervision.
1 code implementation • 21 May 2021 • Xingyu Lin, YuFei Wang, Zixuan Huang, David Held
Robotic manipulation of cloth remains challenging for robotics due to the complex dynamics of the cloth, lack of a low-dimensional state representation, and self-occlusions.
1 code implementation • 28 Apr 2021 • Brian Okorn, Qiao Gu, Martial Hebert, David Held
We also demonstrate how our system can be used by quickly scanning and building a model of a novel object, which can immediately be used by our method for pose estimation.
1 code implementation • 16 Mar 2021 • Harshit Sikchi, Wenxuan Zhou, David Held
Current RL agents explore the environment without considering these constraints, which can lead to damage to the hardware or even other agents in the environment.
no code implementations • 17 Dec 2020 • Xia Chen, Jianren Wang, David Held, Martial Hebert
Visual data in autonomous driving perception, such as camera image and LiDAR point cloud, can be interpreted as a mixture of two aspects: semantic feature and geometric structure.
2 code implementations • 14 Nov 2020 • Wenxuan Zhou, Sujay Bajracharya, David Held
The goal of offline reinforcement learning is to learn a policy from a fixed dataset, without further interactions with the environment.
2 code implementations • 14 Nov 2020 • Xingyu Lin, YuFei Wang, Jake Olkin, David Held
Further, we evaluate a variety of algorithms on these tasks and highlight challenges for reinforcement learning algorithms, including dealing with a state representation that has a high intrinsic dimensionality and is partially observable.
1 code implementation • 13 Nov 2020 • YuFei Wang, Gautham Narayan Narasimhan, Xingyu Lin, Brian Okorn, David Held
Current image-based reinforcement learning (RL) algorithms typically operate on the whole image without performing object-level reasoning.
no code implementations • 9 Oct 2020 • Jianing Qian, Junyu Nan, Siddharth Ancha, Brian Okorn, David Held
Current state-of-the-art trackers often fail due to distractorsand large object appearance changes.
1 code implementation • 23 Aug 2020 • Harshit Sikchi, Wenxuan Zhou, David Held
In this work, we investigate a novel instantiation of H-step lookahead with a learned model and a terminal value function learned by a model-free off-policy algorithm, named Learning Off-Policy with Online Planning (LOOP).
no code implementations • 18 Aug 2020 • Xinshuo Weng, Jianren Wang, David Held, Kris Kitani
Additionally, 3D MOT datasets such as KITTI evaluate MOT methods in 2D space and standardized 3D MOT evaluation tools are missing for a fair comparison of 3D MOT methods.
1 code implementation • 18 Aug 2020 • Jianren Wang, Siddharth Ancha, Yi-Ting Chen, David Held
Instead, we propose leveraging vast unlabeled datasets by self-supervised metric learning of 3D object trackers, with a focus on data association.
1 code implementation • 13 Aug 2020 • Jianing Qian, Thomas Weng, Luxin Zhang, Brian Okorn, David Held
Our approach trains a network to segment the edges and corners of a cloth from a depth image, distinguishing such regions from wrinkles or folds.
Robotics
no code implementations • ECCV 2020 • Siddharth Ancha, Yaadhav Raaj, Peiyun Hu, Srinivasa G. Narasimhan, David Held
Most real-world 3D sensors such as LiDARs perform fixed scans of the entire environment, while being decoupled from the recognition system that processes the sensor data.
1 code implementation • 2 Jul 2020 • Brian Okorn, Mengyun Xu, Martial Hebert, David Held
Our first method, which regresses from deep learned features to an isotropic Bingham distribution, gives the best performance for orientation distribution estimation for non-symmetric objects.
no code implementations • 29 May 2020 • Thomas Weng, Amith Pallankize, Yimin Tang, Oliver Kroemer, David Held
State-of-the-art object grasping methods rely on depth sensing to plan robust grasps, but commercially available depth sensors fail to detect transparent and specular objects.
Robotics
2 code implementations • 27 Dec 2019 • Siddharth Ancha, Junyu Nan, David Held
For a reliable detection system, if a high confidence detection is made, we would want high certainty that the object has indeed been detected.
no code implementations • 10 Dec 2019 • Peiyun Hu, David Held, Deva Ramanan
We prove that if we score a segmentation by the worst objectness among its individual segments, there is an efficient algorithm that finds the optimal worst-case segmentation among an exponentially large number of candidate segmentations.
1 code implementation • CVPR 2020 • Peiyun Hu, Jason Ziglar, David Held, Deva Ramanan
On the NuScenes 3D detection benchmark, we show that, by adding an additional stream for visibility input, we can significantly improve the overall detection accuracy of a state-of-the-art 3D detector.
1 code implementation • NeurIPS 2019 • Xingyu Lin, Harjatin Baweja, George Kantor, David Held
Reinforcement learning is known to be sample inefficient, preventing its application to many real-world problems, especially with high dimensional observations like images.
1 code implementation • CVPR 2020 • Himangi Mittal, Brian Okorn, David Held
When interacting with highly dynamic environments, scene flow allows autonomous systems to reason about the non-rigid motion of multiple independent objects.
no code implementations • 29 Nov 2019 • Peng Yin, Jianing Qian, Yibo Cao, David Held, Howie Choset
In this paper, we introduce a fusion-based depth prediction method, called FusionMapping.
1 code implementation • 9 Jul 2019 • Xinshuo Weng, Jianren Wang, David Held, Kris Kitani
Additionally, 3D MOT datasets such as KITTI evaluate MOT methods in the 2D space and standardized 3D MOT evaluation tools are missing for a fair comparison of 3D MOT methods.
Ranked #3 on
3D Multi-Object Tracking
on KITTI
1 code implementation • 11 Jun 2019 • Siddhant Jain, Sowmya Munukutla, David Held
We propose a point cloud annotation framework that employs human-in-loop learning to enable the creation of large point cloud datasets with per-point annotations.
no code implementations • 20 May 2019 • Xingyu Lin, Harjatin Singh Baweja, David Held
However, if this policy is trained with reinforcement learning, then without a state estimator, it is hard to specify a reward function based on high-dimensional observations.
no code implementations • 15 Mar 2019 • Xingyu Lin, Pengsheng Guo, Carlos Florensa, David Held
Robots that are trained to perform a task in a fixed environment often fail when facing unexpected changes to the environment due to a lack of exploration.
1 code implementation • 27 Nov 2018 • Wentao Yuan, David Held, Christoph Mertz, Martial Hebert
Recently, neural networks operating on point clouds have shown superior performance on 3D understanding tasks such as shape classification and part segmentation.
5 code implementations • 2 Aug 2018 • Wentao Yuan, Tejas Khot, David Held, Christoph Mertz, Martial Hebert
Shape completion, the problem of estimating the complete geometry of objects from partial observations, lies at the core of many vision and robotics applications.
Ranked #4 on
Point Cloud Completion
on Completion3D
no code implementations • 17 Jul 2017 • Carlos Florensa, David Held, Markus Wulfmeier, Michael Zhang, Pieter Abbeel
The robot is trained in reverse, gradually learning to reach the goal from a set of start states increasingly far from the goal.
9 code implementations • ICML 2017 • Joshua Achiam, David Held, Aviv Tamar, Pieter Abbeel
For many applications of reinforcement learning it can be more convenient to specify both a reward function and constraints, rather than trying to design behavior through the reward function.
1 code implementation • ICML 2018 • Carlos Florensa, David Held, Xinyang Geng, Pieter Abbeel
Instead, we propose a method that allows an agent to automatically discover the range of tasks that it is capable of performing.
no code implementations • 15 May 2017 • David Held, Zoe McCarthy, Michael Zhang, Fred Shentu, Pieter Abbeel
Although learning-based methods have great potential for robotics, one concern is that a robot that updates its parameters might cause large amounts of damage before it learns the optimal policy.
no code implementations • 11 Feb 2017 • Sandy H. Huang, David Held, Pieter Abbeel, Anca D. Dragan
We show that certain approximate-inference models lead to the robot generating example behaviors that better enable users to anticipate what it will do in novel situations.
3 code implementations • 6 Apr 2016 • David Held, Sebastian Thrun, Silvio Savarese
We propose a method for offline training of neural networks that can track novel objects at test-time at 100 fps.
no code implementations • 29 Jul 2015 • David Held, Sebastian Thrun, Silvio Savarese
We show that feedforward neural networks outperform state-of-the-art methods for recognizing objects from novel viewpoints even when trained from just a single image per object.