1 code implementation • 24 Jan 2025 • Xiaohao Xu, Tianyi Zhang, Shibo Zhao, Xiang Li, Sibo Wang, Yongqi Chen, Ye Li, Bhiksha Raj, Matthew Johnson-Roberson, Sebastian Scherer, Xiaonan Huang
We aim to redefine robust ego-motion estimation and photorealistic 3D reconstruction by addressing a critical limitation: the reliance on noise-free data in existing models.
no code implementations • 29 Nov 2024 • Herbert Wright, Weiming Zhi, Matthew Johnson-Roberson, Tucker Hermans
However, this can be brittle to noisy real-world observations and objects not contained in the dataset, and cannot reason about their confidence.
no code implementations • 8 Oct 2024 • Xinyi Liu, Tianyi Zhang, Matthew Johnson-Roberson, Weiming Zhi
These regions are then projected to the camera's view as it moves over time and a cost is constructed.
no code implementations • 7 Oct 2024 • Ziwen Yuan, Tianyi Zhang, Matthew Johnson-Roberson, Weiming Zhi
Advances in photorealistic environment models have enabled robots to develop hyper-realistic reconstructions, which can be used to generate images that are intuitive for human inspection.
no code implementations • 26 Sep 2024 • Quanting Xie, So Yeon Min, Pengliang Ji, Yue Yang, Tianyi Zhang, Kedi Xu, Aarav Bajaj, Ruslan Salakhutdinov, Matthew Johnson-Roberson, Yonatan Bisk
There is no limit to how much a robot might explore and learn, but all of that knowledge needs to be searchable and actionable.
no code implementations • 14 Jul 2024 • Weiming Zhi, Haozhan Tang, Tianyi Zhang, Matthew Johnson-Roberson
We empirically evaluate our approach on a robot manipulator holding a diverse set of real-world objects.
no code implementations • 14 Jul 2024 • Tianyi Zhang, Weiming Zhi, Kaining Huang, Joshua Mangelson, Corina Barbalata, Matthew Johnson-Roberson
Water caustics are commonly observed in seafloor imaging data from shallow-water areas.
1 code implementation • 24 Jun 2024 • Xiaohao Xu, Tianyi Zhang, Sibo Wang, Xiang Li, Yongqi Chen, Ye Li, Bhiksha Raj, Matthew Johnson-Roberson, Xiaonan Huang
Embodied agents require robust navigation systems to operate in unstructured environments, making the robustness of Simultaneous Localization and Mapping (SLAM) models critical to embodied agent autonomy.
1 code implementation • 28 May 2024 • Xiaohao Xu, Ye Li, Tianyi Zhang, Jinrong Yang, Matthew Johnson-Roberson, Xiaonan Huang
Constructing large-scale labeled datasets for multi-modal perception model training in autonomous driving presents significant challenges.
no code implementations • 17 Apr 2024 • Weiming Zhi, Haozhan Tang, Tianyi Zhang, Matthew Johnson-Roberson
We demonstrate that JCR can build effective scene representations using a low-cost RGB camera attached to a manipulator, without prior calibration.
1 code implementation • 16 Mar 2024 • Tianyi Zhang, Kaining Huang, Weiming Zhi, Matthew Johnson-Roberson
Humans have the remarkable ability to construct consistent mental models of an environment, even under limited or varying levels of illumination.
1 code implementation • 12 Feb 2024 • Xiaohao Xu, Tianyi Zhang, Sibo Wang, Xiang Li, Yongqi Chen, Ye Li, Bhiksha Raj, Matthew Johnson-Roberson, Xiaonan Huang
To this end, we propose a novel, customizable pipeline for noisy data synthesis, aimed at assessing the resilience of multi-modal SLAM models against various perturbations.
no code implementations • 14 Dec 2023 • Yafei Hu, Quanting Xie, Vidhi Jain, Jonathan Francis, Jay Patrikar, Nikhil Keetha, Seungchan Kim, Yaqi Xie, Tianyi Zhang, Hao-Shu Fang, Shibo Zhao, Shayegan Omidshafiei, Dong-Ki Kim, Ali-akbar Agha-mohammadi, Katia Sycara, Matthew Johnson-Roberson, Dhruv Batra, Xiaolong Wang, Sebastian Scherer, Chen Wang, Zsolt Kira, Fei Xia, Yonatan Bisk
Motivated by the impressive open-set performance and content generation capabilities of web-scale, large-capacity pre-trained models (i. e., foundation models) in research fields such as Natural Language Processing (NLP) and Computer Vision (CV), we devote this survey to exploring (i) how these existing foundation models from NLP and CV can be applied to the field of general-purpose robotics, and also exploring (ii) what a robotics-specific foundation model would look like.
no code implementations • 19 Sep 2023 • Weiming Zhi, Tianyi Zhang, Matthew Johnson-Roberson
In this work, we tackle the problem of teaching a robot to approach a surface and then follow cyclic motion on it, where the cycle of the motion can be arbitrarily specified by a single user-provided sketch over an image from the robot's camera.
1 code implementation • 18 Sep 2023 • Quanting Xie, Tianyi Zhang, Kedi Xu, Matthew Johnson-Roberson, Yonatan Bisk
We introduce a new task OUTDOOR, a new mechanism for Large Language Models (LLMs) to accurately hallucinate possible futures, and a new computationally aware success metric for pushing research forward in this more complex domain.
no code implementations • 7 Sep 2023 • Weiming Zhi, Tianyi Zhang, Matthew Johnson-Roberson
Diagrammatic Teaching aims to teach robots novel skills by prompting the user to sketch out demonstration trajectories on 2D images of the scene, these are then synthesised as a generative model of motion trajectories in 3D task space.
1 code implementation • CVPR 2023 • Junming Zhang, Haomeng Zhang, Ram Vasudevan, Matthew Johnson-Roberson
Most real-world 3D measurements from depth sensors are incomplete, and to address this issue the point cloud completion task aims to predict the complete shapes of objects from partial observations.
1 code implementation • 6 Apr 2023 • Tianyi Zhang, Matthew Johnson-Roberson
The proposed technique integrates underwater light effects into a volume rendering framework with end-to-end differentiability.
1 code implementation • 18 Nov 2022 • Ming-Yuan Yu, Ram Vasudevan, Matthew Johnson-Roberson
LiDARs have been widely adopted to modern self-driving vehicles, providing 3D information of the scene and surrounding objects.
no code implementations • 2 Sep 2022 • Alexandra Carlson, Manikandasriram Srinivasan Ramanagopal, Nathan Tseng, Matthew Johnson-Roberson, Ram Vasudevan, Katherine A. Skinner
Recent advances in neural radiance fields (NeRFs) achieve state-of-the-art novel view synthesis and facilitate dense estimation of scene properties.
no code implementations • 22 Mar 2022 • Niankai Yang, Chao Shen, Matthew Johnson-Roberson, Jing Sun
In the first stage, the surge velocity, heave velocity, and pitch angle setpoints are optimized by minimizing the required vehicle propulsion energy under currents, and the line-of-sight (LOS) guidance law is used to generate the yaw angle setpoint that ensures path following.
1 code implementation • 9 Sep 2021 • Tianyi Zhang, Matthew Johnson-Roberson
Robot localization remains a challenging task in GPS denied environments.
no code implementations • 2 Jan 2021 • Junming Zhang, Ming-Yuan Yu, Ram Vasudevan, Matthew Johnson-Roberson
Point cloud analysis is an area of increasing interest due to the development of 3D sensors that are able to rapidly measure the depth of scenes accurately.
1 code implementation • 29 Jul 2020 • Yu Yao, Ella Atkins, Matthew Johnson-Roberson, Ram Vasudevan, Xiaoxiao Du
BiTraP estimates the goal (end-point) of trajectories and introduces a novel bi-directional decoder to improve longer-term trajectory prediction accuracy.
Ranked #2 on
Trajectory Prediction
on JAAD
1 code implementation • 9 Jul 2020 • Jun-ming Zhang, Weijia Chen, Yu-Ping Wang, Ram Vasudevan, Matthew Johnson-Roberson
This paper illustrates that this proposed method achieves state-of-the-art performance on shape classification, part segmentation and point cloud completion.
1 code implementation • 8 Jun 2020 • Manikandasriram Srinivasan Ramanagopal, Zixu Zhang, Ram Vasudevan, Matthew Johnson-Roberson
To address this problem, this paper formulates reversing the effect of thermal inertia at a single pixel as a Least Absolute Shrinkage and Selection Operator (LASSO) problem which we can solve rapidly using a quadratic programming solver.
1 code implementation • 1 Jun 2020 • Cyrus Anderson, Ram Vasudevan, Matthew Johnson-Roberson
Pedestrians and drivers interact closely in a wide range of environments.
Robotics
1 code implementation • 14 Apr 2020 • Gideon Billings, Eduardo Iscar, Matthew Johnson-Roberson
The design of optical systems for underwater vehicles is a complex process where the selection of cameras, lenses, housings, and operational parameters greatly influence the performance of the complete system.
no code implementations • 27 Feb 2020 • Gideon Billings, Matthew Johnson-Roberson
There has been much recent interest in deep learning methods for monocular image based object pose estimation.
2 code implementations • 5 Feb 2020 • Patrick Holmes, Shreyas Kousik, Bohao Zhang, Daphna Raz, Corina Barbalata, Matthew Johnson-Roberson, Ram Vasudevan
At runtime, in each receding-horizon planning iteration, ARMTD constructs a reachable set of the entire arm in workspace and intersects it with obstacles to generate sub-differentiable and provably-conservative collision-avoidance constraints on the trajectory parameters.
Robotics
no code implementations • 23 Sep 2019 • Alexandra Carlson, Ram Vasudevan, Matthew Johnson-Roberson
There have been impressive advances in the realm of image to image translation in transferring previously unseen visual effects into a dataset, specifically in day to night translation.
1 code implementation • 11 Sep 2019 • Cyrus Anderson, Ram Vasudevan, Matthew Johnson-Roberson
Highway driving places significant demands on human drivers and autonomous vehicles (AVs) alike due to high speeds and the complex interactions in dense traffic.
Robotics Signal Processing
no code implementations • 7 May 2019 • Jun-ming Zhang, Manikandasriram Srinivasan Ramanagopal, Ram Vasudevan, Matthew Johnson-Roberson
An accurate depth map of the environment is critical to the safe operation of autonomous robots and vehicles.
1 code implementation • 5 Mar 2019 • Cyrus Anderson, Xiaoxiao Du, Ram Vasudevan, Matthew Johnson-Roberson
Our work demonstrates the effectiveness and potential of using simulation as a substitution for human annotation efforts to train high-performing prediction algorithms such as the DNNs.
Robotics
1 code implementation • 5 Feb 2019 • Sean Vaskov, Utkarsh Sharma, Shreyas Kousik, Matthew Johnson-Roberson, Ramanarayan Vasudevan
Trajectory planning is challenging for autonomous cars since they operate in unpredictable environments with limited sensor horizons.
Systems and Control
1 code implementation • 18 Sep 2018 • Shreyas Kousik, Sean Vaskov, Fan Bu, Matthew Johnson-Roberson, Ram Vasudevan
At runtime, the FRS is used to map obstacles to the space of parameterized trajectories, which allows RTD to select a safe trajectory at every planning iteration.
Robotics Systems and Control
2 code implementations • 18 Sep 2018 • Gideon Billings, Matthew Johnson-Roberson
Autonomous robot manipulation involves estimating the translation and orientation of the object to be manipulated as a 6-degree-of-freedom (6D) pose.
no code implementations • 17 Sep 2018 • Alexandra Carlson, Katherine A. Skinner, Ram Vasudevan, Matthew Johnson-Roberson
This domain shift is especially exaggerated between synthetic and real datasets.
no code implementations • 13 Sep 2018 • Jun-ming Zhang, Katherine A. Skinner, Ram Vasudevan, Matthew Johnson-Roberson
Initial disparity estimates are refined with an embedding learned from the semantic segmentation branch of the network.
no code implementations • 11 Sep 2018 • Xiaoxiao Du, Ram Vasudevan, Matthew Johnson-Roberson
In applications such as autonomous driving, it is important to understand, infer, and anticipate the intention and future behavior of pedestrians.
no code implementations • 10 Sep 2018 • Wonhui Kim, Manikandasriram Srinivasan Ramanagopal, Charles Barto, Ming-Yuan Yu, Karl Rosaen, Nick Goumas, Ram Vasudevan, Matthew Johnson-Roberson
This paper presents a novel dataset titled PedX, a large-scale multimodal collection of pedestrians at complex urban intersections.
1 code implementation • 21 Mar 2018 • Alexandra Carlson, Katherine A. Skinner, Ram Vasudevan, Matthew Johnson-Roberson
Recent work has focused on generating synthetic imagery to increase the size and variability of training data for learning visual tasks in urban scenes.
1 code implementation • 30 Jun 2017 • Manikandasriram Srinivasan Ramanagopal, Cyrus Anderson, Ram Vasudevan, Matthew Johnson-Roberson
We show that a state-of-the-art detector, tracker, and our classifier trained only on synthetic data can identify valid errors on KITTI tracking dataset with an average precision of 0. 94.
2 code implementations • 28 Apr 2017 • Shreyas Kousik, Sean Vaskov, Matthew Johnson-Roberson, Ramanarayan Vasudevan
Path planning for autonomous vehicles in arbitrary environments requires a guarantee of safety, but this can be impractical to ensure in real-time when the vehicle is described with a high-fidelity model.
Systems and Control Robotics
1 code implementation • 23 Feb 2017 • Jie Li, Katherine A. Skinner, Ryan M. Eustice, Matthew Johnson-Roberson
Due to the depth-dependent water column effects inherent to underwater environments, we show that our end-to-end network implicitly learns a coarse depth estimate of the underwater scene from monocular underwater images.
2 code implementations • 6 Oct 2016 • Matthew Johnson-Roberson, Charles Barto, Rounak Mehta, Sharath Nittur Sridhar, Karl Rosaen, Ram Vasudevan
Deep learning has rapidly transformed the state of the art algorithms used to address a variety of problems in computer vision and robotics.