no code implementations • 4 May 2022 • Dorian Henning, Tristan Laidlow, Stefan Leutenegger
Through a series of experiments on video sequences of human motion captured by a moving monocular camera, we demonstrate that BodySLAM improves estimates of all human body parameters and camera poses when compared to estimating these separately.
no code implementations • 18 Feb 2022 • Stefan Leutenegger
Robust and accurate state estimation remains a challenge in robotics, Augmented, and Virtual Reality (AR/VR), even as Visual-Inertial Simultaneous Localisation and Mapping (VI-SLAM) getting commoditised.
no code implementations • ICCV 2021 • Zoe Landgraf, Raluca Scona, Tristan Laidlow, Stephen James, Stefan Leutenegger, Andrew J. Davison
At test time, our model can generate 3D shape and instance segmentation from a single depth view, probabilistically sampling proposals for the occluded region from the learned latent space.
no code implementations • ICCV 2021 • Shuaifeng Zhi, Tristan Laidlow, Stefan Leutenegger, Andrew J. Davison
Semantic labelling is highly correlated with geometry and radiance reconstruction, as scene entities with similar shape and appearance are more likely to come from similar classes.
no code implementations • 31 Aug 2020 • Binbin Xu, Andrew J. Davison, Stefan Leutenegger
Dense image alignment from RGB-D images remains a critical issue for real-world applications, especially under challenging lighting conditions and in a wide baseline setting.
1 code implementation • CVPR 2020 • Joseph Ortiz, Mark Pupilli, Stefan Leutenegger, Andrew J. Davison
Graph processors such as Graphcore's Intelligence Processing Unit (IPU) are part of the major new wave of novel computer architecture for AI, and have a general design with massively parallel computation, distributed on-chip memory and very high inter-core communication bandwidth which allows breakthrough performance for message passing algorithms on arbitrary graphs.
no code implementations • 24 Feb 2020 • Zoe Landgraf, Fabian Falck, Michael Bloesch, Stefan Leutenegger, Andrew Davison
Generally capable Spatial AI systems must build persistent scene representations where geometric models are combined with meaningful semantic labels.
no code implementations • 18 Feb 2020 • Ujwal Bonde, Pablo F. Alcantarilla, Stefan Leutenegger
Our approach is distinct from previous works in panoptic segmentation that rely on a combination of a semantic segmentation network with a computationally costly instance segmentation network based on bounding box proposals, such as Mask R-CNN, to guide the prediction of instance labels using a Mixture-of-Expert (MoE) approach.
4 code implementations • 17 Apr 2019 • Guillermo Gallego, Tobi Delbruck, Garrick Orchard, Chiara Bartolozzi, Brian Taba, Andrea Censi, Stefan Leutenegger, Andrew Davison, Joerg Conradt, Kostas Daniilidis, Davide Scaramuzza
Event cameras offer attractive properties compared to traditional cameras: high temporal resolution (in the order of microseconds), very high dynamic range (140 dB vs. 60 dB), low power consumption, and high pixel bandwidth (on the order of kHz) resulting in reduced motion blur.
no code implementations • CVPR 2019 • Shuaifeng Zhi, Michael Bloesch, Stefan Leutenegger, Andrew J. Davison
Systems which incrementally create 3D semantic maps from image sequences must store and update representations of both geometry and semantic entities.
no code implementations • 3 Mar 2019 • Andrea Nicastro, Ronald Clark, Stefan Leutenegger
Detailed 3D reconstruction is an important challenge with application to robotics, augmented and virtual reality, which has seen impressive progress throughout the past years.
no code implementations • 19 Dec 2018 • Binbin Xu, Wenbin Li, Dimos Tzoumanikas, Michael Bloesch, Andrew Davison, Stefan Leutenegger
It can provide robust camera tracking in dynamic environments and at the same time, continuously estimate geometric, semantic, and motion properties for arbitrary objects in the scene.
no code implementations • ECCV 2018 • Ronald Clark, Michael Bloesch, Jan Czarnowski, Stefan Leutenegger, Andrew J. Davison
In this paper, we propose LS-Net, a neural nonlinear least squares optimization algorithm which learns to effectively optimize these cost functions even in the presence of adversities.
no code implementations • 3 Sep 2018 • Wenbin Li, Sajad Saeedi, John McCormac, Ronald Clark, Dimos Tzoumanikas, Qing Ye, Yuzhong Huang, Rui Tang, Stefan Leutenegger
Datasets have gained an enormous amount of popularity in the computer vision community, from training and evaluation of Deep Learning-based methods to benchmarking Simultaneous Localization and Mapping (SLAM).
no code implementations • ECCV 2018 • Ronald Clark, Michael Bloesch, Jan Czarnowski, Stefan Leutenegger, Andrew J. Davison
In this paper, we propose a neural nonlinear least squares optimization algorithm which learns to effectively optimize these cost functions even in the presence of adversities.
no code implementations • 25 Aug 2018 • John McCormac, Ronald Clark, Michael Bloesch, Andrew J. Davison, Stefan Leutenegger
Reconstructed objects are stored in an optimisable 6DoF pose graph which is our only persistent map representation.
no code implementations • 27 Jul 2018 • Mickey Li, Noyan Songur, Pavel Orlov, Stefan Leutenegger, A. Aldo Faisal
Incorporating the physical environment is essential for a complete understanding of human behavior in unconstrained every-day tasks.
1 code implementation • CVPR 2018 • Michael Bloesch, Jan Czarnowski, Ronald Clark, Stefan Leutenegger, Andrew J. Davison
Our approach is suitable for use in a keyframe-based monocular dense SLAM system: While each keyframe with a code can produce a depth map, the code can be optimised efficiently jointly with pose variables and together with the codes of overlapping keyframes to attain global consistency.
2 code implementations • 3 Apr 2018 • Michael Bloesch, Jan Czarnowski, Ronald Clark, Stefan Leutenegger, Andrew J. Davison
Our approach is suitable for use in a keyframe-based monocular dense SLAM system: While each keyframe with a code can produce a depth map, the code can be optimised efficiently jointly with pose variables and together with the codes of overlapping keyframes to attain global consistency.
no code implementations • ICCV 2017 • John McCormac, Ankur Handa, Stefan Leutenegger, Andrew J. Davison
We compare the semantic segmentation performance of network weights produced from pre-training on RGB images from our dataset against generic VGG-16 ImageNet weights.
no code implementations • 29 Aug 2017 • Jan Czarnowski, Stefan Leutenegger, Andrew Davison
We argue that robust dense SLAM systems can make valuable use of the layers of features coming from a standard CNN as a pyramid of `semantic texture' which is suitable for dense alignment while being much more robust to nuisance factors such as lighting than raw RGB values.
1 code implementation • 15 Dec 2016 • John McCormac, Ankur Handa, Stefan Leutenegger, Andrew J. Davison
We introduce SceneNet RGB-D, expanding the previous work of SceneNet to enable large scale photorealistic rendering of indoor scene trajectories.
no code implementations • 16 Sep 2016 • John McCormac, Ankur Handa, Andrew Davison, Stefan Leutenegger
This not only produces a useful semantic 3D map, but we also show on the NYUv2 dataset that fusing multiple predictions leads to an improvement even in the 2D semantic labelling over baseline single frame predictions.
no code implementations • 7 Aug 2016 • Edward Johns, Stefan Leutenegger, Andrew J. Davison
With this, it is possible to achieve grasping robust to the gripper's pose uncertainty, by smoothing the grasp function with the pose uncertainty function.
no code implementations • CVPR 2016 • Patrick Bardow, Andrew J. Davison, Stefan Leutenegger
In a series of examples, we demonstrate the successful operation of our framework, including in situations where conventional cameras heavily suffer from dynamic range limitations or motion blur.
no code implementations • CVPR 2016 • Edward Johns, Stefan Leutenegger, Andrew J. Davison
A multi-view image sequence provides a much richer capacity for object recognition than from a single image.
no code implementations • 18 May 2015 • Michael Milford, Hanme Kim, Michael Mangan, Stefan Leutenegger, Tom Stone, Barbara Webb, Andrew Davison
Event-based cameras offer much potential to the fields of robotics and computer vision, in part due to their large dynamic range and extremely high "frame rates".