no code implementations • 3 Dec 2024 • Chenguang Huang, Shengchao Yan, Wolfram Burgard
Dynamic scene understanding remains a persistent challenge in robotic applications.
no code implementations • 4 Oct 2024 • Ksheeraja Raghavan, Samiran Gode, Ankit Shah, Surabhi Raghavan, Wolfram Burgard, Bhiksha Raj, Rita Singh
The data produced using the framework serves as a benchmark for anomaly detection applications, potentially enhancing the performance of models trained on audio data, particularly in handling out-of-distribution cases.
1 code implementation • 24 Sep 2024 • Yannik Blei, Michael Krawez, Nisarga Nilavadi, Tanja Katharina Kaiser, Wolfram Burgard
Nowadays, unmanned aerial vehicles (UAVs) are commonly used in search and rescue scenarios to gather information in the search area.
no code implementations • 5 Aug 2024 • Sai Prasanna, Daniel Honerkamp, Kshitij Sirohi, Tim Welschehold, Wolfram Burgard, Abhinav Valada
Embodied AI has made significant progress acting in unexplored environments.
1 code implementation • 18 Jul 2024 • Yue Yao, Shengchao Yan, Daniel Goehring, Wolfram Burgard, Joerg Reichardt
Within our OoD testing protocol, we further study two augmentation strategies of SotA models and their effects on model generalization.
1 code implementation • 29 May 2024 • Niclas Vödisch, Kürsat Petek, Markus Käppeler, Abhinav Valada, Wolfram Burgard
A key challenge for the widespread application of learning-based models for robotic perception is to significantly reduce the required amount of annotated training data while achieving accurate predictions.
no code implementations • 29 May 2024 • Nikhil Gosala, Kürsat Petek, B Ravi Kiran, Senthil Yogamani, Paulo Drews-Jr, Wolfram Burgard, Abhinav Valada
Our approach pretrains the network to independently reason about scene geometry and scene semantics using two disjoint neural pathways in an unsupervised manner and then finetunes it for the task of semantic BEV mapping using only a small fraction of labels in the BEV.
no code implementations • 26 Mar 2024 • Abdelrhman Werby, Chenguang Huang, Martin Büchner, Abhinav Valada, Wolfram Burgard
Recent open-vocabulary robot mapping methods enrich dense geometric maps with pre-trained visual-language features.
1 code implementation • 18 Mar 2024 • Jonas Schramm, Niclas Vödisch, Kürsat Petek, B Ravi Kiran, Senthil Yogamani, Wolfram Burgard, Abhinav Valada
Semantic scene segmentation from a bird's-eye-view (BEV) perspective plays a crucial role in facilitating planning and decision-making for mobile robots.
1 code implementation • 18 Mar 2024 • Shengchao Yan, Lukas König, Wolfram Burgard
Active traffic management with autonomous vehicles offers the potential for reduced congestion and improved traffic flow.
1 code implementation • 13 Dec 2023 • Eugenio Chisari, Nick Heppert, Tim Welschehold, Wolfram Burgard, Abhinav Valada
It consists of an RGB-D image encoder that leverages recent advances to detect objects and infer their pose and latent code, and a decoder to predict shape and grasps for each object in the scene.
no code implementations • 23 Oct 2023 • Iman Nematollahi, Kirill Yankov, Wolfram Burgard, Tim Welschehold
A long-standing challenge for a robotic manipulation system operating in real-world scenarios is adapting and generalizing its acquired motor skills to unseen environments.
1 code implementation • 9 Oct 2023 • Michael G. Adam, Sebastian Eger, Martin Piccolrovazzi, Maged Iskandar, Joern Vogel, Alexander Dietrich, Seongjien Bien, Jon Skerlj, Abdeldjallil Naceri, Eckehard Steinbach, Alin Albu-Schaeffer, Sami Haddadin, Wolfram Burgard
As labor shortage increases in the health sector, the demand for assistive robotics grows.
1 code implementation • 19 Sep 2023 • Markus Käppeler, Kürsat Petek, Niclas Vödisch, Wolfram Burgard, Abhinav Valada
Concurrently, recent breakthroughs in visual representation learning have sparked a paradigm shift leading to the advent of large foundation models that can be trained with completely unlabeled images.
no code implementations • 10 Aug 2023 • D. Adriana Gómez-Rosal, Max Bergau, Georg K. J. Fischer, Andreas Wachaja, Johannes Gräter, Matthias Odenweller, Uwe Piechottka, Fabian Hoeflinger, Nikhil Gosala, Niklas Wetzel, Daniel Büscher, Abhinav Valada, Wolfram Burgard
In today's chemical plants, human field operators perform frequent integrity checks to guarantee high safety standards, and thus are possibly the first to encounter dangerous operating conditions.
1 code implementation • 27 Jun 2023 • Jannik Zürn, Ingmar Posner, Wolfram Burgard
To overcome this limitation, we propose to use the motion patterns of traffic participants as lane graph annotations.
no code implementations • 20 Jun 2023 • Guangming Wang, Yu Zheng, Yanfeng Guo, Zhe Liu, Yixiang Zhu, Wolfram Burgard, Hesheng Wang
A popular approach to robot localization is based on image-to-point cloud registration, which combines illumination-invariant LiDAR-based mapping with economical image-based localization.
1 code implementation • 8 May 2023 • Jan Ole von Hartz, Eugenio Chisari, Tim Welschehold, Wolfram Burgard, Joschka Boedecker, Abhinav Valada
We employ our method to learn challenging multi-object robot manipulation tasks from wrist camera observations and demonstrate superior utility for policy learning compared to other representation learning techniques.
no code implementations • 21 Mar 2023 • Johan Vertens, Nicolai Dorka, Tim Welschehold, Michael Thompson, Wolfram Burgard
By training everything end-to-end with the loss of the dynamics model, we enforce the latent mapper to learn an update rule for the latent map that is useful for the subsequent dynamics model.
1 code implementation • 17 Mar 2023 • Niclas Vödisch, Kürsat Petek, Wolfram Burgard, Abhinav Valada
Operating a robot in the open world requires a high level of robustness with respect to previously unseen environments.
1 code implementation • 17 Mar 2023 • Nicolai Dorka, Tim Welschehold, Wolfram Burgard
Early stopping based on the validation set performance is a popular approach to find the right balance between under- and overfitting in the context of supervised learning.
Model-based Reinforcement Learning reinforcement-learning +2
no code implementations • 13 Mar 2023 • Chenguang Huang, Oier Mees, Andy Zeng, Wolfram Burgard
While interacting in the world is a multi-sensory experience, many robots continue to predominantly rely on visual perception to map and navigate in their environments.
no code implementations • 6 Mar 2023 • Monish R. Nallapareddy, Kshitij Sirohi, Paulo L. J. Drews-Jr, Wolfram Burgard, Chih-Hong Cheng, Abhinav Valada
In this work, we propose EvCenterNet, a novel uncertainty-aware 2D object detection framework using evidential learning to directly estimate both classification and regression uncertainties.
no code implementations • CVPR 2023 • Martin Büchner, Jannik Zürn, Ion-George Todoran, Abhinav Valada, Wolfram Burgard
To overcome these challenges, we propose a novel bottom-up approach to lane graph estimation from aerial imagery that aggregates multiple overlapping graphs into a single consistent graph.
no code implementations • CVPR 2023 • Nikhil Gosala, Kürsat Petek, Paulo L. J. Drews-Jr, Wolfram Burgard, Abhinav Valada
Implicit supervision trains the model by enforcing spatial consistency of the scene over time based on FV semantic sequences, while explicit supervision exploits BEV pseudolabels generated from FV semantic annotations and self-supervised depth estimates.
1 code implementation • 11 Oct 2022 • Chenguang Huang, Oier Mees, Andy Zeng, Wolfram Burgard
Grounding language to the visual observations of a navigating agent can be performed using off-the-shelf visual-language models pretrained on Internet-scale data (e. g., image captions).
1 code implementation • 10 Oct 2022 • Kshitij Sirohi, Sajad Marvi, Daniel Büscher, Wolfram Burgard
Current learning-based methods typically try to achieve maximum performance for this task, while neglecting a proper estimation of the associated uncertainties.
1 code implementation • 4 Oct 2022 • Oier Mees, Jessica Borja-Diaz, Wolfram Burgard
Recent works have shown that Large Language Models (LLMs) can be applied to ground natural language to a wide variety of robot skills.
Ranked #1 on Avg. sequence length on CALVIN
1 code implementation • 19 Sep 2022 • Erick Rosete-Beas, Oier Mees, Gabriel Kalweit, Joschka Boedecker, Wolfram Burgard
Concretely, we combine a low-level policy that learns latent skills via imitation learning and a high-level policy learned from offline reinforcement learning for skill-chaining the latent behavior priors.
1 code implementation • 19 Sep 2022 • Iman Nematollahi, Erick Rosete-Beas, Seyed Mahdi B. Azad, Raghu Rajan, Frank Hutter, Wolfram Burgard
To the best of our knowledge, our model is the first generative model that provides an RGB-D video prediction of the future for a static camera.
no code implementations • 12 Sep 2022 • Jannik Zürn, Sebastian Weber, Wolfram Burgard
Robustly classifying ground infrastructure such as roads and street crossings is an essential task for mobile robots operating alongside pedestrians.
no code implementations • 15 Jul 2022 • Johan Vertens, Wolfram Burgard
In this paper we propose USegScene, a framework for semantically guided unsupervised learning of depth, optical flow and ego-motion estimation for stereo camera images using convolutional neural networks.
1 code implementation • 29 Jun 2022 • Kshitij Sirohi, Sajad Marvi, Daniel Büscher, Wolfram Burgard
In this work, we introduce the novel task of uncertainty-aware panoptic segmentation, which aims to predict per-pixel semantic and instance segmentations, together with per-pixel uncertainty estimates.
2 code implementations • 13 Apr 2022 • Oier Mees, Lukas Hermann, Wolfram Burgard
We have open-sourced our implementation to facilitate future research in learning to perform many complex manipulation skills in a row specified with natural language.
1 code implementation • 1 Mar 2022 • Jessica Borja-Diaz, Oier Mees, Gabriel Kalweit, Lukas Hermann, Joschka Boedecker, Wolfram Burgard
Robots operating in human-centered environments should have the ability to understand how objects function: what can be done with each object, where this interaction may occur, and how the object is used to achieve a goal.
no code implementations • 30 Jan 2022 • Jannik Zürn, Wolfram Burgard
In extensive experiments carried out with a real-world dataset, we demonstrate that our approach provides accurate detections of moving vehicles and does not require manual annotations.
1 code implementation • 6 Dec 2021 • Oier Mees, Lukas Hermann, Erick Rosete-Beas, Wolfram Burgard
We show that a baseline model based on multi-context imitation learning performs poorly on CALVIN, suggesting that there is significant room for developing innovative agents that learn to relate human language to their world models with this benchmark.
no code implementations • 25 Nov 2021 • Iman Nematollahi, Erick Rosete-Beas, Adrian Röfer, Tim Welschehold, Abhinav Valada, Wolfram Burgard
A core challenge for an autonomous agent acting in the real world is to adapt its repertoire of skills to cope with its noisy perception and dynamics.
1 code implementation • 24 Nov 2021 • Nicolai Dorka, Tim Welschehold, Joschka Boedecker, Wolfram Burgard
Accurate value estimates are important for off-policy reinforcement learning.
no code implementations • 20 Oct 2021 • Kürsat Petek, Kshitij Sirohi, Daniel Büscher, Wolfram Burgard
Robust localization in dense urban scenarios using a low-cost sensor setup and sparse HD maps is highly relevant for the current advances in autonomous driving, but remains a challenging topic in research.
no code implementations • 11 Jun 2021 • Shengchao Yan, Tim Welschehold, Daniel Büscher, Wolfram Burgard
Our reinforcement learning agent learns a policy for a centralized controller to let connected autonomous vehicles at unsignalized intersections give up their right of way and yield to other vehicles to optimize traffic flow.
1 code implementation • 1 May 2021 • Jannik Zürn, Johan Vertens, Wolfram Burgard
Lane-level scene annotations provide invaluable data in autonomous vehicles for trajectory planning in complex environments such as urban areas and cities.
Ranked #2 on Lane Detection on nuScenes
no code implementations • 29 Apr 2021 • Artemij Amiranashvili, Max Argus, Lukas Hermann, Wolfram Burgard, Thomas Brox
Visual domain randomization in simulated environments is a widely used method to transfer policies trained in simulation to real robots.
1 code implementation • CVPR 2021 • Vitor Guizilini, Rares Ambrus, Wolfram Burgard, Adrien Gaidon
Estimating scene geometry from data obtained with cost-effective sensors is key for robots and self-driving cars.
1 code implementation • ICCV 2021 • Pavel Tokmakov, Jie Li, Wolfram Burgard, Adrien Gaidon
In this work, we introduce an end-to-end trainable approach for joint object detection and tracking that is capable of such reasoning.
no code implementations • 16 Feb 2021 • Kshitij Sirohi, Rohit Mohan, Daniel Büscher, Wolfram Burgard, Abhinav Valada
Panoptic segmentation of point clouds is a crucial task that enables autonomous vehicles to comprehend their vicinity using their highly accurate and reliable LiDAR sensors.
2 code implementations • 16 Feb 2021 • Oier Mees, Wolfram Burgard
Controlling robots to perform tasks via natural language is one of the most challenging topics in human-robot interaction.
no code implementations • 17 Nov 2020 • Nicolai Dorka, Johannes Meyer, Wolfram Burgard
Real-time object detection in videos using lightweight hardware is a crucial component of many robotic tasks.
1 code implementation • 21 Oct 2020 • Barbara Barros Carlos, Tommaso Sartor, Andrea Zanelli, Gianluca Frison, Wolfram Burgard, Moritz Diehl, Giuseppe Oriolo
The advances in computer processor technology have enabled the application of nonlinear model predictive control (NMPC) to agile systems, such as quadrotors.
Robotics Systems and Control Systems and Control Optimization and Control
no code implementations • 17 Sep 2020 • Lukas Enderich, Fabian Timm, Wolfram Burgard
Deep neural networks (DNNs) are usually over-parameterized to increase the likelihood of getting adequate initial weights by random initialization.
no code implementations • 29 Aug 2020 • Andreas Bühler, Adrien Gaidon, Andrei Cramariuc, Rares Ambrus, Guy Rosman, Wolfram Burgard
In this work, we propose a behavioral cloning approach that can safely leverage imperfect perception without being conservative.
1 code implementation • 15 Aug 2020 • Igor Vasiljevic, Vitor Guizilini, Rares Ambrus, Sudeep Pillai, Wolfram Burgard, Greg Shakhnarovich, Adrien Gaidon
Self-supervised learning has emerged as a powerful tool for depth and ego-motion estimation, leading to state-of-the-art results on benchmark datasets.
no code implementations • 3 Aug 2020 • Kuan-Hui Lee, Matthew Kliemann, Adrien Gaidon, Jie Li, Chao Fang, Sudeep Pillai, Wolfram Burgard
In autonomous driving, accurately estimating the state of surrounding obstacles is critical for safe and robust path planning.
no code implementations • 2 Aug 2020 • Iman Nematollahi, Oier Mees, Lukas Hermann, Wolfram Burgard
A key challenge for an agent learning to interact with the world is to reason about physical properties of objects and to foresee their dynamics under the effect of applied forces.
no code implementations • 19 Jul 2020 • Gabriel L. Oliveira, Senthil Yogamani, Wolfram Burgard, Thomas Brox
In order to further improve the architecture we introduce a weight function which aims to re-balance classes to increase the attention of the networks to under-represented objects.
1 code implementation • 6 Jul 2020 • Artemij Amiranashvili, Nicolai Dorka, Wolfram Burgard, Vladlen Koltun, Thomas Brox
Imitation learning is a powerful family of techniques for learning sensorimotor coordination in immersive environments.
1 code implementation • 19 May 2020 • Andreas Eitel, Nico Hauff, Wolfram Burgard
To achieve this, we fine-tune an existing DeepMask network for instance segmentation on the self-labeled training data acquired by the robot.
no code implementations • 17 Apr 2020 • Juana Valeria Hurtado, Rohit Mohan, Wolfram Burgard, Abhinav Valada
In this paper, we introduce a novel perception task denoted as multi-object panoptic tracking (MOPT), which unifies the conventionally disjoint tasks of semantic segmentation, instance segmentation, and multi-object tracking.
no code implementations • 10 Mar 2020 • Johan Vertens, Jannik Zürn, Wolfram Burgard
We avoid the expensive annotation of nighttime images by leveraging an existing daytime RGB-dataset and propose a teacher-student training approach that transfers the dataset's knowledge to the nighttime domain.
no code implementations • 9 Mar 2020 • Shengchao Yan, Jingwei Zhang, Daniel Büscher, Wolfram Burgard
In this paper we present an approach to learning policies for signal controllers using deep reinforcement learning aiming for optimized traffic flow.
no code implementations • 19 Feb 2020 • Lukas Enderich, Fabian Timm, Wolfram Burgard
We propose SYMOG (symmetric mixture of Gaussian modes), which significantly decreases the complexity of DNNs through low-bit fixed-point quantization.
2 code implementations • 23 Jan 2020 • Oier Mees, Alp Emek, Johan Vertens, Wolfram Burgard
One particular requirement for such robots is that they are able to understand spatial relations and can place objects in accordance with the spatial relations expressed by their user.
no code implementations • 6 Dec 2019 • Jannik Zürn, Wolfram Burgard, Abhinav Valada
In this work, we propose a novel terrain classification framework leveraging an unsupervised proprioceptive classifier that learns from vehicle-terrain interaction sounds to self-supervise an exteroceptive classifier for pixel-wise semantic segmentation of images.
no code implementations • 23 Oct 2019 • Lukas Luft, Alexander Schaefer, Tobias Schubert, Wolfram Burgard
A popular class of lidar-based grid mapping algorithms computes for each map cell the probability that it reflects an incident laser beam.
1 code implementation • 23 Oct 2019 • Alexander Schaefer, Daniel Büscher, Johan Vertens, Lukas Luft, Wolfram Burgard
Due to their ubiquity and long-term stability, pole-like objects are well suited to serve as landmarks for vehicle localization in urban environments.
no code implementations • 23 Oct 2019 • Alexander Schaefer, Lukas Luft, Wolfram Burgard
Most robot mapping techniques for lidar sensors tessellate the environment into pixels or voxels and assume uniformity of the environment within them.
no code implementations • 23 Oct 2019 • Alexander Schaefer, Lukas Luft, Wolfram Burgard
However, many common lidar models perform poorly in unstructured, unpredictable environments, they lack a consistent physical model for both mapping and localization, and they do not exploit all the information the sensor provides, e. g. out-of-range measurements.
1 code implementation • 23 Oct 2019 • Alexander Schaefer, Johan Vertens, Daniel Büscher, Wolfram Burgard
Whether it is object detection, model reconstruction, laser odometry, or point cloud registration: Plane extraction is a vital component of many robotic systems.
1 code implementation • 21 Oct 2019 • Oier Mees, Markus Merklinger, Gabriel Kalweit, Wolfram Burgard
Our method learns a general skill embedding independently from the task context by using an adversarial loss.
1 code implementation • 17 Oct 2019 • Lukas Hermann, Max Argus, Andreas Eitel, Artemij Amiranashvili, Wolfram Burgard, Thomas Brox
We propose Adaptive Curriculum Generation from Demonstrations (ACGD) for reinforcement learning in the presence of sparse rewards.
1 code implementation • 17 Oct 2019 • Oier Mees, Maxim Tatarchenko, Thomas Brox, Wolfram Burgard
We present a convolutional neural network for joint 3D shape prediction and viewpoint estimation from a single input image.
3D Object Reconstruction From A Single Image 3D Reconstruction +3
no code implementations • 3 Sep 2019 • Henrich Kolkhorst, Wolfram Burgard, Michael Tangermann
Robot motions in the presence of humans should not only be feasible and safe, but also conform to human preferences.
no code implementations • 16 Jul 2019 • Lukas Enderich, Fabian Timm, Lars Rosenbaum, Wolfram Burgard
Due to their high computational complexity, deep neural networks are still limited to powerful processing units.
2 code implementations • 24 Jun 2019 • Daniele Cattaneo, Matteo Vaghi, Augusto Luis Ballardini, Simone Fontana, Domenico Giorgio Sorrenti, Wolfram Burgard
In this paper we present CMRNet, a realtime approach based on a Convolutional Neural Network to localize an RGB image of a scene in a map built from LiDAR data.
1 code implementation • 17 Jun 2019 • Ayush Dewan, Wolfram Burgard
To make the predictions from the DCNN temporally consistent, we propose a Bayes filter based method.
no code implementations • 4 Jun 2019 • Mayank Mittal, Rohit Mohan, Wolfram Burgard, Abhinav Valada
This problem is extremely challenging as pre-existing maps cannot be leveraged for navigation due to structural changes that may have occurred.
no code implementations • 18 Mar 2019 • Jingwei Zhang, Niklas Wetzel, Nicolai Dorka, Joschka Boedecker, Wolfram Burgard
Many state-of-the-art methods use intrinsic motivation to complement the sparse extrinsic reward signal, giving the agent more opportunities to receive feedback during exploration.
1 code implementation • 5 Mar 2019 • Federico Boniardi, Abhinav Valada, Rohit Mohan, Tim Caselitz, Wolfram Burgard
Indoor localization is one of the crucial enablers for deployment of service robots.
no code implementations • 20 Sep 2018 • Ayush Dewan, Tim Caselitz, Wolfram Burgard
Our proposed architecture consists of a Siamese network for learning a feature descriptor and a metric learning network for matching the descriptors.
no code implementations • 15 Sep 2018 • Mayank Mittal, Abhinav Valada, Wolfram Burgard
However, these UAVs have to be able to autonomously land on debris piles in order to accurately locate the survivors.
Robotics
no code implementations • 21 Aug 2018 • Noha Radwan, Wolfram Burgard, Abhinav Valada
Learned representations from the traffic light recognition stream are fused with the estimated trajectories from the motion prediction stream to learn the crossing decision.
1 code implementation • 11 Aug 2018 • Abhinav Valada, Rohit Mohan, Wolfram Burgard
To address this limitation, we propose a mutimodal semantic segmentation framework that dynamically adapts the fusion of modality-specific features while being sensitive to the object category, spatial location and scene context in a self-supervised manner.
Ranked #1 on Scene Recognition on ScanNet
no code implementations • 4 May 2018 • Martin Völker, Jiří Hammer, Robin T. Schirrmeister, Joos Behncke, Lukas D. J. Fiederer, Andreas Schulze-Bonhage, Petr Marusič, Wolfram Burgard, Tonio Ball
Deep learning techniques have revolutionized the field of machine learning and were recently successfully applied to various classification problems in noninvasive electroencephalography (EEG).
no code implementations • 23 Apr 2018 • Noha Radwan, Abhinav Valada, Wolfram Burgard
Semantic understanding and localization are fundamental enablers of robot autonomy that have for the most part been tackled as disjoint problems.
no code implementations • 18 Apr 2018 • Niko Sünderhauf, Oliver Brock, Walter Scheirer, Raia Hadsell, Dieter Fox, Jürgen Leitner, Ben Upcroft, Pieter Abbeel, Wolfram Burgard, Michael Milford, Peter Corke
In this paper we discuss a number of robotics-specific learning, reasoning, and embodiment challenges for deep learning.
Robotics
no code implementations • 2 Apr 2018 • Abhinav Valada, Wolfram Burgard
Terrain classification is a critical component of any autonomous mobile robot system operating in unknown real-world environments.
no code implementations • 2 Apr 2018 • Oleksii Zhelo, Jingwei Zhang, Lei Tai, Ming Liu, Wolfram Burgard
A video of our experimental results can be found at https://goo. gl/pWbpcF.
1 code implementation • 9 Mar 2018 • Abhinav Valada, Noha Radwan, Wolfram Burgard
We evaluate our proposed VLocNet on indoor as well as outdoor datasets and show that even our single task model exceeds the performance of state-of-the-art deep architectures for global localization, while achieving competitive performance for visual odometry estimation.
1 code implementation • 7 Mar 2018 • Christian Zimmermann, Tim Welschehold, Christian Dornhege, Wolfram Burgard, Thomas Brox
We propose an approach to estimate 3D human pose in real world units from a single RGBD image and show that it exceeds performance of monocular 3D pose estimation approaches from color as well as pose estimation exclusively from depth.
Ranked #14 on 3D Human Pose Estimation on Total Capture
no code implementations • 1 Feb 2018 • Jingwei Zhang, Lei Tai, Peng Yun, Yufeng Xiong, Ming Liu, Joschka Boedecker, Wolfram Burgard
In this paper, we deal with the reality gap from a novel perspective, targeting transferring Deep Reinforcement Learning (DRL) policies learned in simulated environments to the real-world domain for visual control tasks.
no code implementations • 16 Nov 2017 • Joos Behncke, Robin Tibor Schirrmeister, Wolfram Burgard, Tonio Ball
Analysis of brain signals from a human interacting with a robot may help identifying robot errors, but accuracies of such analyses have still substantial space for improvement.
no code implementations • 25 Oct 2017 • Martin Völker, Robin T. Schirrmeister, Lukas D. J. Fiederer, Wolfram Burgard, Tonio Ball
We recorded high-density EEG in a flanker task experiment (31 subjects) and an online BCI control paradigm (4 subjects).
1 code implementation • 6 Oct 2017 • Lei Tai, Jingwei Zhang, Ming Liu, Wolfram Burgard
Experiments show that our GAIL-based approach greatly improves the safety and efficiency of the behavior of mobile robots from pure behavior cloning.
no code implementations • 14 Sep 2017 • Florian Kraemer, Alexander Schaefer, Andreas Eitel, Johan Vertens, Wolfram Burgard
Agricultural robots are expected to increase yields in a sustainable way and automate precision tasks, such as weeding and plant monitoring.
no code implementations • 4 Aug 2017 • Dominik Welke, Joos Behncke, Marina Hader, Robin Tibor Schirrmeister, Andreas Schönau, Boris Eßmann, Oliver Müller, Wolfram Burgard, Tonio Ball
Our findings suggest that non-invasive recordings of brain responses elicited when observing robots indeed contain decodable information about the correctness of the robot's action and the type of observed robot.
no code implementations • 2 Aug 2017 • Andres Vasquez, Marina Kollmitz, Andreas Eitel, Wolfram Burgard
In this paper, we propose a depth-based perception pipeline that estimates the position and velocity of people in the environment and categorizes them according to the mobility aids they use: pedestrian, person in wheelchair, person in a wheelchair with a person pushing them, person with crutches and person using a walker.
no code implementations • 25 Jul 2017 • Andreas Eitel, Nico Hauff, Wolfram Burgard
We present a novel neural network-based approach that separates unknown objects in clutter by selecting favourable push actions.
no code implementations • 20 Jul 2017 • Felix Burget, Lukas Dominique Josef Fiederer, Daniel Kuhner, Martin Völker, Johannes Aldinger, Robin Tibor Schirrmeister, Chau Do, Joschka Boedecker, Bernhard Nebel, Tonio Ball, Wolfram Burgard
As our results demonstrate, our system is capable of adapting to frequent changes in the environment and reliably completing given tasks within a reasonable amount of time.
1 code implementation • 18 Jul 2017 • Oier Mees, Andreas Eitel, Wolfram Burgard
Object detection is an essential task for autonomous robots operating in dynamic and changing environments.
1 code implementation • 4 Jul 2017 • Philipp Jund, Andreas Eitel, Nichola Abdo, Wolfram Burgard
To operate intelligently in domestic environments, robots require the ability to understand arbitrary spatial relations between objects and to generalize them to objects of varying sizes and shapes.
1 code implementation • 29 Jun 2017 • Jingwei Zhang, Lei Tai, Ming Liu, Joschka Boedecker, Wolfram Burgard
We present an approach for agents to learn representations of a global map from sensor data, to aid their exploration in new environments.
no code implementations • 27 Jun 2017 • Gabriel L. Oliveira, Noha Radwan, Wolfram Burgard, Thomas Brox
Compared to LiDAR-based localization methods, which provide high accuracy but rely on expensive sensors, visual localization approaches only require a camera and thus are more cost-effective while their accuracy and reliability typically is inferior to LiDAR-based methods.
no code implementations • 26 Jun 2017 • Ayush Dewan, Gabriel L. Oliveira, Wolfram Burgard
To learn the distinction between movable and non-movable points in the environment, we introduce an approach based on deep neural network and for detecting the dynamic points, we estimate pointwise motion.
5 code implementations • 15 Mar 2017 • Robin Tibor Schirrmeister, Jost Tobias Springenberg, Lukas Dominique Josef Fiederer, Martin Glasstetter, Katharina Eggensperger, Michael Tangermann, Frank Hutter, Wolfram Burgard, Tonio Ball
PLEASE READ AND CITE THE REVISED VERSION at Human Brain Mapping: http://onlinelibrary. wiley. com/doi/10. 1002/hbm. 23730/full Code available here: https://github. com/robintibor/braindecode
1 code implementation • 6 Mar 2017 • Oier Mees, Nichola Abdo, Mladen Mazuran, Wolfram Burgard
Human-centered environments are rich with a wide variety of spatial relations between everyday objects.
1 code implementation • 21 Dec 2016 • Lei Tai, Jingwei Zhang, Ming Liu, Joschka Boedecker, Wolfram Burgard
We carry out our discussions on the two main paradigms for learning control with deep networks: deep reinforcement learning and imitation learning.
no code implementations • 16 Dec 2016 • Jingwei Zhang, Jost Tobias Springenberg, Joschka Boedecker, Wolfram Burgard
We propose a successor feature based deep reinforcement learning algorithm that can learn to transfer knowledge from previously mastered navigation tasks to new problem instances.
2 code implementations • 17 Nov 2016 • Philipp Jund, Nichola Abdo, Andreas Eitel, Wolfram Burgard
In this paper, we address this issue and present a dataset consisting of 5, 000 images covering 25 different classes of groceries, with at least 97 images per class.
no code implementations • 13 Apr 2016 • Michael Herman, Tobias Gindele, Jörg Wagner, Felix Schmitt, Wolfram Burgard
Inverse Reinforcement Learning (IRL) describes the problem of learning an unknown reward function of a Markov Decision Process (MDP) from observed behavior of an agent.
2 code implementations • 24 Jul 2015 • Andreas Eitel, Jost Tobias Springenberg, Luciano Spinello, Martin Riedmiller, Wolfram Burgard
Robust object recognition is a crucial ingredient of many, if not all, real-world robotics applications.
no code implementations • 2 Apr 2015 • Bahram Behzadian, Pratik Agarwal, Wolfram Burgard, Gian Diego Tipaldi
In this paper, we address the localization problem when the map of the environment is not present beforehand, and the robot relies on a hand-drawn map from a non-expert user.
no code implementations • 14 Mar 2015 • Pratik Agarwal, Wolfram Burgard, Luciano Spinello
In this paper, we present a novel approach that instead uses geotagged panoramas from the Google Street View as a source of global positioning.
no code implementations • 2 Feb 2015 • Miguel Heredia, Felix Endres, Wolfram Burgard, Rafael Sanz
In this paper we present a novel approach to global localization using an RGB-D camera in maps of visual features.