1 code implementation • 4 Dec 2024 • Fabian Schmidt, Markus Enzweiler, Abhinav Valada
Navigating outdoor environments with visual Simultaneous Localization and Mapping (SLAM) systems poses significant challenges due to dynamic scenes, lighting variations, and seasonal changes, requiring robust solutions.
Computational Efficiency Simultaneous Localization and Mapping
1 code implementation • 3 Dec 2024 • Fabian Schmidt, Constantin Blessing, Markus Enzweiler, Abhinav Valada
To address this gap, we present ROVER, a comprehensive benchmark dataset tailored for evaluating visual SLAM algorithms under diverse environmental conditions and spatial configurations.
no code implementations • 2 Dec 2024 • Christina Kassab, Matías Mattamala, Sacha Morin, Martin Büchner, Abhinav Valada, Liam Paull, Maurice Fallon
3D open-vocabulary scene graph methods are a promising map representation for embodied agents, however many current approaches are computationally expensive.
no code implementations • 4 Nov 2024 • Michael Kurenkov, Sajad Marvi, Julian Schmidt, Christoph B. Rist, Alessandro Canevaro, Hang Yu, Julian Jordan, Georg Schildbach, Abhinav Valada
The increasing interest in autonomous driving systems has highlighted the need for an in-depth analysis of human driving behavior in diverse scenarios.
1 code implementation • 26 Oct 2024 • Muhammad Zubair Irshad, Mauro Comi, Yen-Chen Lin, Nick Heppert, Abhinav Valada, Rares Ambrus, Zsolt Kira, Jonathan Tremblay
Neural Fields have emerged as a transformative approach for 3D scene representation in computer vision and robotics, enabling accurate inference of geometry, 3D semantics, and dynamics from posed 2D data.
no code implementations • 9 Oct 2024 • Rohit Mohan, Daniele Cattaneo, Florian Drews, Abhinav Valada
Multi-sensor fusion is crucial for accurate 3D object detection in autonomous driving, with cameras and LiDAR being the most commonly used sensors.
no code implementations • 2 Oct 2024 • Aron Distelzweig, Andreas Look, Eitan Kosman, Faris Janjoš, Jörg Wagner, Abhinav Valada
In autonomous driving, accurate motion prediction is essential for safe and efficient motion planning.
no code implementations • 23 Sep 2024 • Daniel Honerkamp, Harsh Mahesheka, Jan Ole von Hartz, Tim Welschehold, Abhinav Valada
In this work, we present MoMa-Teleop, a novel teleoperation method that delegates the base motions to a reinforcement learning agent, leaving the operator to focus fully on the task-relevant end-effector motions.
no code implementations • 18 Sep 2024 • Juana Valeria Hurtado, Riya Mohan, Abhinav Valada
Forecasting the semantics and 3D structure of scenes is essential for robots to navigate and plan actions safely.
no code implementations • 16 Sep 2024 • Aron Distelzweig, Eitan Kosman, Andreas Look, Faris Janjoš, Denesh K. Manivannan, Abhinav Valada
In this paper, we propose a novel sampling method applicable to trajectory prediction based on the predictions of multiple models.
no code implementations • 3 Aug 2024 • Fabian Schmidt, Constantin Blessing, Markus Enzweiler, Abhinav Valada
The findings highlight the importance of loop closing in improving localization accuracy while managing computational resources efficiently, offering valuable insights for optimizing Visual-Inertial SLAM systems for practical outdoor applications in mobile robotics.
no code implementations • 25 Jul 2024 • Julia Hindel, Daniele Cattaneo, Abhinav Valada
However, existing methods impose strict rigidity on old classes, reducing their effectiveness in learning new incremental classes.
Autonomous Driving Class-Incremental Semantic Segmentation +3
1 code implementation • 18 Jul 2024 • Jan Ole von Hartz, Tim Welschehold, Abhinav Valada, Joschka Boedecker
Task Parametrized Gaussian Mixture Models (TP-GMM) are a sample-efficient method for learning object-centric robot manipulation tasks.
no code implementations • 8 Jul 2024 • Martin Büchner, Simon Dorer, Abhinav Valada
This work introduces a novel approach to generating successor lane graphs from aerial imagery, utilizing the advanced capabilities of transformer models.
no code implementations • 3 Jun 2024 • Stephan Lochner, Daniel Honerkamp, Abhinav Valada, Andrew D. Straw
Bees are among the master navigators of the insect world.
1 code implementation • 29 May 2024 • Niclas Vödisch, Kürsat Petek, Markus Käppeler, Abhinav Valada, Wolfram Burgard
A key challenge for the widespread application of learning-based models for robotic perception is to significantly reduce the required amount of annotated training data while achieving accurate predictions.
no code implementations • 29 May 2024 • Nikhil Gosala, Kürsat Petek, B Ravi Kiran, Senthil Yogamani, Paulo Drews-Jr, Wolfram Burgard, Abhinav Valada
Our approach pretrains the network to independently reason about scene geometry and scene semantics using two disjoint neural pathways in an unsupervised manner and then finetunes it for the task of semantic BEV mapping using only a small fraction of labels in the BEV.
no code implementations • 2 May 2024 • Abdallah Ayad, Adrian Röfer, Nick Heppert, Abhinav Valada
We train Imagine2touch on two out of those shapes and validate it on the ood.
1 code implementation • 23 Apr 2024 • Sassan Mokhtar, Eugenio Chisari, Nick Heppert, Abhinav Valada
Precisely grasping and reconstructing articulated objects is key to enabling general robotic manipulation.
no code implementations • 19 Apr 2024 • Christopher Lang, Alexander Braun, Lars Schillingmann, Abhinav Valada
Unlike other LiDAR-based multi-task architectures, our proposed PAttFormer does not require separate feature encoders for multiple task-specific point cloud representations, resulting in a network that is 3x smaller and 1. 4x faster while achieving competitive performance on the nuScenes and KITTI benchmarks for autonomous driving perception.
no code implementations • 26 Mar 2024 • Abdelrhman Werby, Chenguang Huang, Martin Büchner, Abhinav Valada, Wolfram Burgard
Recent open-vocabulary robot mapping methods enrich dense geometric maps with pre-trained visual-language features.
no code implementations • 22 Mar 2024 • Adrian Röfer, Nick Heppert, Abdallah Ayman, Eugenio Chisari, Abhinav Valada
We frame this problem as the task of learning a low-dimensional visual-tactile embedding, wherein we encode a depth patch from which we decode the tactile signal.
no code implementations • 22 Mar 2024 • Nick Heppert, Max Argus, Tim Welschehold, Thomas Brox, Abhinav Valada
In the online trajectory generation stage, we first re-detect all objects, then warp the demonstration trajectory to the current scene and execute it on the robot.
1 code implementation • 18 Mar 2024 • Jonas Schramm, Niclas Vödisch, Kürsat Petek, B Ravi Kiran, Senthil Yogamani, Wolfram Burgard, Abhinav Valada
Semantic scene segmentation from a bird's-eye-view (BEV) perspective plays a crucial role in facilitating planning and decision-making for mobile robots.
no code implementations • 31 Jan 2024 • Daniele Cattaneo, Abhinav Valada
In this paper, we present CMRNext, a novel approach for camera-LIDAR matching that is independent of sensor-specific parameters, generalizable, and can be used in the wild for monocular localization in LiDAR maps and camera-LiDAR extrinsic calibration.
1 code implementation • 13 Dec 2023 • Eugenio Chisari, Nick Heppert, Tim Welschehold, Wolfram Burgard, Abhinav Valada
It consists of an RGB-D image encoder that leverages recent advances to detect objects and infer their pose and latent code, and a decoder to predict shape and grasps for each object in the scene.
no code implementations • 13 Nov 2023 • Maximilian Luz, Rohit Mohan, Ahmed Rida Sekkat, Oliver Sawade, Elmar Matthes, Thomas Brox, Abhinav Valada
Optical flow estimation is very challenging in situations with transparent or occluded objects.
no code implementations • 18 Oct 2023 • Rohit Mohan, Kiran Kumaraswamy, Juana Valeria Hurtado, Kürsat Petek, Abhinav Valada
Deep learning has led to remarkable strides in scene understanding with panoptic segmentation emerging as a key holistic scene interpretation task.
no code implementations • 6 Oct 2023 • Max Argus, Abhijeet Nayak, Martin Büchner, Silvio Galesso, Abhinav Valada, Thomas Brox
In this work, we present a framework that formulates the visual servoing task as graph traversal.
1 code implementation • 19 Sep 2023 • Markus Käppeler, Kürsat Petek, Niclas Vödisch, Wolfram Burgard, Abhinav Valada
Concurrently, recent breakthroughs in visual representation learning have sparked a paradigm shift leading to the advent of large foundation models that can be trained with completely unlabeled images.
no code implementations • 18 Sep 2023 • Abhijeet Nayak, Daniele Cattaneo, Abhinav Valada
RaLF is composed of radar and LiDAR feature encoders, a place recognition head that generates global descriptors, and a metric localization head that predicts the 3-DoF transformation between the radar scan and the map.
no code implementations • 16 Sep 2023 • David Unger, Nikhil Gosala, Varun Ravi Kumar, Shubhankar Borse, Abhinav Valada, Senthil Yogamani
Surround vision systems that are pretty common in new vehicles use the IPM principle to generate a BEV image and to show it on display to the driver.
no code implementations • 12 Sep 2023 • Ahmed Rida Sekkat, Rohit Mohan, Oliver Sawade, Elmar Matthes, Abhinav Valada
To address these limitations, we introduce AmodalSynthDrive, a synthetic multi-task multi-modal amodal perception dataset.
no code implementations • 10 Aug 2023 • D. Adriana Gómez-Rosal, Max Bergau, Georg K. J. Fischer, Andreas Wachaja, Johannes Gräter, Matthias Odenweller, Uwe Piechottka, Fabian Hoeflinger, Nikhil Gosala, Niklas Wetzel, Daniel Büscher, Abhinav Valada, Wolfram Burgard
In today's chemical plants, human field operators perform frequent integrity checks to guarantee high safety standards, and thus are possibly the first to encounter dangerous operating conditions.
no code implementations • 6 Aug 2023 • Rohit Mohan, José Arce, Sassan Mokhtar, Daniele Cattaneo, Abhinav Valada
Safety and efficiency are paramount in healthcare facilities where the lives of patients are at stake.
no code implementations • 12 Jul 2023 • Fabian Schmalstieg, Daniel Honerkamp, Tim Welschehold, Abhinav Valada
We present HIMOS, a hierarchical reinforcement learning approach that learns to compose exploration, navigation, and manipulation skills.
1 code implementation • 8 May 2023 • Jan Ole von Hartz, Eugenio Chisari, Tim Welschehold, Wolfram Burgard, Joschka Boedecker, Abhinav Valada
We employ our method to learn challenging multi-object robot manipulation tasks from wrist camera observations and demonstrate superior utility for policy learning compared to other representation learning techniques.
no code implementations • 25 Apr 2023 • Christopher Lang, Alexander Braun, Lars Schillingmann, Abhinav Valada
We hypothesize that this drawback results from formulating self-supervised objectives that are limited to single frames or frame pairs.
no code implementations • 31 Mar 2023 • Julia Hindel, Nikhil Gosala, Kevin Bregler, Abhinav Valada
Perception datasets for agriculture are limited both in quantity and diversity which hinders effective training of supervised learning approaches.
1 code implementation • CVPR 2023 • Nick Heppert, Muhammad Zubair Irshad, Sergey Zakharov, Katherine Liu, Rares Andrei Ambrus, Jeannette Bohg, Abhinav Valada, Thomas Kollar
We present CARTO, a novel approach for reconstructing multiple articulated objects from a single stereo RGB observation.
1 code implementation • 17 Mar 2023 • Niclas Vödisch, Kürsat Petek, Wolfram Burgard, Abhinav Valada
Operating a robot in the open world requires a high level of robustness with respect to previously unseen environments.
no code implementations • 6 Mar 2023 • Monish R. Nallapareddy, Kshitij Sirohi, Paulo L. J. Drews-Jr, Wolfram Burgard, Chih-Hong Cheng, Abhinav Valada
In this work, we propose EvCenterNet, a novel uncertainty-aware 2D object detection framework using evidential learning to directly estimate both classification and regression uncertainties.
no code implementations • 17 Feb 2023 • Christopher Lang, Alexander Braun, Lars Schillingmann, Karsten Haug, Abhinav Valada
Self-supervised feature learning enables perception systems to benefit from the vast raw data recorded by vehicle fleets worldwide.
no code implementations • CVPR 2023 • Martin Büchner, Jannik Zürn, Ion-George Todoran, Abhinav Valada, Wolfram Burgard
To overcome these challenges, we propose a novel bottom-up approach to lane graph estimation from aerial imagery that aggregates multiple overlapping graphs into a single consistent graph.
no code implementations • CVPR 2023 • Nikhil Gosala, Kürsat Petek, Paulo L. J. Drews-Jr, Wolfram Burgard, Abhinav Valada
Implicit supervision trains the model by enforcing spatial consistency of the scene over time based on FV semantic sequences, while explicit supervision exploits BEV pseudolabels generated from FV semantic annotations and self-supervised depth estimates.
no code implementations • 7 Jul 2022 • Laura Londoño, Juana Valeria Hurtado, Nora Hertz, Philipp Kellmeyer, Silja Voeneky, Abhinav Valada
In this work, we present the first survey on fairness in robot learning from an interdisciplinary perspective spanning technical, ethical, and legal challenges.
1 code implementation • 17 Jun 2022 • Daniel Honerkamp, Tim Welschehold, Abhinav Valada
Despite its importance in both industrial and service robotics, mobile manipulation remains a significant challenge as it requires a seamless integration of end-effector trajectory generation with navigation skills as well as reasoning over long-horizons.
no code implementations • 29 May 2022 • Rohit Mohan, Abhinav Valada
Amodal panoptic segmentation aims to connect the perception of the world to its cognitive understanding.
no code implementations • 17 May 2022 • Jan Ole von Hartz, Eugenio Chisari, Tim Welschehold, Abhinav Valada
In recent years, policy learning methods using either reinforcement or imitation have made significant progress.
no code implementations • 21 Mar 2022 • Martin Buchner, Abhinav Valada
We evaluate our approach using various sensor modalities and model configurations on the challenging nuScenes and KITTI datasets.
no code implementations • 15 Mar 2022 • Christopher Lang, Alexander Braun, Abhinav Valada
Object detection, for the most part, has been formulated in the euclidean space, where euclidean or spherical geodesic distances measure the similarity of an image region to an object class prototype.
no code implementations • CVPR 2022 • Rohit Mohan, Abhinav Valada
To enable robots to reason with this capability, we formulate and propose a novel task that we name amodal panoptic segmentation.
Ranked #1 on Amodal Panoptic Segmentation on BDD100K val
no code implementations • 15 Feb 2022 • Thomas Elsken, Arber Zela, Jan Hendrik Metzen, Benedikt Staffler, Thomas Brox, Abhinav Valada, Frank Hutter
The success of deep learning in recent years has lead to a rising demand for neural network architecture engineering.
no code implementations • 5 Feb 2022 • Laura Londoño, Adrian Röfer, Tim Welschehold, Abhinav Valada
As robotic systems become more and more capable of assisting humans in their everyday lives, we must consider the opportunities for these artificial agents to make their human collaborators feel unsafe or to treat them unfairly.
no code implementations • 21 Dec 2021 • Christopher Lang, Alexander Braun, Abhinav Valada
Object recognition for the most part has been approached as a one-hot problem that treats classes to be discrete and unrelated.
no code implementations • 9 Dec 2021 • Rohit Mohan, Abhinav Valada
In this technical report, we describe our EfficientLPT architecture that won the panoptic tracking challenge in the 7th AI Driving Olympics at NeurIPS 2021.
1 code implementation • 29 Nov 2021 • Abdelrahman Younes, Daniel Honerkamp, Tim Welschehold, Abhinav Valada
Audio-visual navigation combines sight and hearing to navigate to a sound-emitting source in an unmapped environment.
no code implementations • 25 Nov 2021 • Iman Nematollahi, Erick Rosete-Beas, Adrian Röfer, Tim Welschehold, Abhinav Valada, Wolfram Burgard
A core challenge for an autonomous agent acting in the real world is to adapt its repertoire of skills to cope with its noisy perception and dynamics.
no code implementations • 30 Sep 2021 • Borna Bešić, Nikhil Gosala, Daniele Cattaneo, Abhinav Valada
Unsupervised Domain Adaptation (UDA) techniques are thus essential to fill this domain gap and retain the performance of models on new sensor setups without the need for additional data labeling.
1 code implementation • 8 Sep 2021 • Whye Kit Fong, Rohit Mohan, Juana Valeria Hurtado, Lubing Zhou, Holger Caesar, Oscar Beijbom, Abhinav Valada
Panoptic scene understanding and tracking of dynamic agents are essential for robots and automated vehicles to navigate in urban environments.
Ranked #1 on Panoptic Segmentation on Panoptic nuScenes test
1 code implementation • 6 Aug 2021 • Nikhil Gosala, Abhinav Valada
Bird's-Eye-View (BEV) maps have emerged as one of the most powerful representations for scene understanding due to their ability to provide rich spatial context while being easy to interpret and process.
no code implementations • 20 May 2021 • Peter Jakob, Manav Madan, Tobias Schmid-Schirling, Abhinav Valada
Furthermore, we introduce the dices dataset, which consists of over 2000 grayscale images of falling dices from multiple perspectives, with 5\% of the images containing rare anomalies (e. g., drill holes, sawing, or scratches).
1 code implementation • 8 Mar 2021 • Daniele Cattaneo, Matteo Vaghi, Abhinav Valada
Loop closure detection is an essential component of Simultaneous Localization and Mapping (SLAM) systems, which reduces the drift accumulated over time.
no code implementations • CVPR 2021 • Francisco Rivera Valverde, Juana Valeria Hurtado, Abhinav Valada
In this work, we present the novel self-supervised MM-DistillNet framework consisting of multiple teachers that leverage diverse modalities including RGB, depth and thermal images, to simultaneously exploit complementary cues and distill knowledge into a single audio student network.
no code implementations • 16 Feb 2021 • Kshitij Sirohi, Rohit Mohan, Daniel Büscher, Wolfram Burgard, Abhinav Valada
Panoptic segmentation of point clouds is a crucial task that enables autonomous vehicles to comprehend their vicinity using their highly accurate and reliable LiDAR sensors.
no code implementations • 7 Jan 2021 • Juana Valeria Hurtado, Laura Londoño, Abhinav Valada
The exponentially increasing advances in robotics and machine learning are facilitating the transition of robots from being confined to controlled industrial spaces to performing novel everyday tasks in domestic and urban environments.
no code implementations • 23 Aug 2020 • Rohit Mohan, Abhinav Valada
In this technical report, we present key details of our winning panoptic segmentation architecture EffPS_b1bs4_RVC.
1 code implementation • 12 Aug 2020 • Borna Bešić, Abhinav Valada
Dynamic objects have a significant impact on the robot's perception of the environment which degrades the performance of essential tasks such as localization and mapping.
2 code implementations • 20 Apr 2020 • Daniele Cattaneo, Domenico Giorgio Sorrenti, Abhinav Valada
In this paper, we now take it a step further by introducing CMRNet++, which is a significantly more robust model that not only generalizes to new places effectively, but is also independent of the camera parameters.
no code implementations • 17 Apr 2020 • Juana Valeria Hurtado, Rohit Mohan, Wolfram Burgard, Abhinav Valada
In this paper, we introduce a novel perception task denoted as multi-object panoptic tracking (MOPT), which unifies the conventionally disjoint tasks of semantic segmentation, instance segmentation, and multi-object tracking.
2 code implementations • 5 Apr 2020 • Rohit Mohan, Abhinav Valada
Understanding the scene in which an autonomous robot operates is critical for its competent functioning.
Ranked #1 on Panoptic Segmentation on Indian Driving Dataset
no code implementations • 6 Dec 2019 • Jannik Zürn, Wolfram Burgard, Abhinav Valada
In this work, we propose a novel terrain classification framework leveraging an unsupervised proprioceptive classifier that learns from vehicle-terrain interaction sounds to self-supervise an exteroceptive classifier for pixel-wise semantic segmentation of images.
no code implementations • 4 Jun 2019 • Mayank Mittal, Rohit Mohan, Wolfram Burgard, Abhinav Valada
This problem is extremely challenging as pre-existing maps cannot be leveraged for navigation due to structural changes that may have occurred.
1 code implementation • 5 Mar 2019 • Federico Boniardi, Abhinav Valada, Rohit Mohan, Tim Caselitz, Wolfram Burgard
Indoor localization is one of the crucial enablers for deployment of service robots.
no code implementations • 15 Sep 2018 • Mayank Mittal, Abhinav Valada, Wolfram Burgard
However, these UAVs have to be able to autonomously land on debris piles in order to accurately locate the survivors.
Robotics
no code implementations • 21 Aug 2018 • Noha Radwan, Wolfram Burgard, Abhinav Valada
Learned representations from the traffic light recognition stream are fused with the estimated trajectories from the motion prediction stream to learn the crossing decision.
1 code implementation • 11 Aug 2018 • Abhinav Valada, Rohit Mohan, Wolfram Burgard
To address this limitation, we propose a mutimodal semantic segmentation framework that dynamically adapts the fusion of modality-specific features while being sensitive to the object category, spatial location and scene context in a self-supervised manner.
Ranked #1 on Scene Recognition on ScanNet
no code implementations • 23 Apr 2018 • Noha Radwan, Abhinav Valada, Wolfram Burgard
Semantic understanding and localization are fundamental enablers of robot autonomy that have for the most part been tackled as disjoint problems.
no code implementations • 2 Apr 2018 • Abhinav Valada, Wolfram Burgard
Terrain classification is a critical component of any autonomous mobile robot system operating in unknown real-world environments.
1 code implementation • 9 Mar 2018 • Abhinav Valada, Noha Radwan, Wolfram Burgard
We evaluate our proposed VLocNet on indoor as well as outdoor datasets and show that even our single task model exceeds the performance of state-of-the-art deep architectures for global localization, while achieving competitive performance for visual odometry estimation.