no code implementations • 16 Dec 2024 • Riku Murai, Eric Dexheimer, Andrew J. Davison
We present a real-time monocular dense SLAM system designed bottom-up from MASt3R, a two-view 3D reconstruction and matching prior.
no code implementations • 4 Apr 2024 • Eric Dexheimer, Andrew J. Davison
We present COMO, a real-time monocular mapping and odometry system that encodes dense geometry via a compact set of 3D anchor points.
no code implementations • 22 Mar 2024 • Aalok Patwardhan, Callum Rhodes, Gwangbin Bae, Andrew J. Davison
Given a sequence of images, we can use the per-frame rotation estimates and their uncertainty to perform multi-frame optimisation, achieving robustness and temporal consistency.
1 code implementation • CVPR 2024 • Gwangbin Bae, Andrew J. Davison
Despite the growing demand for accurate surface normal estimation models, existing methods use general-purpose dense prediction models, adopting the same inductive biases as other tasks.
1 code implementation • CVPR 2024 • Xin Kong, Shikun Liu, Xiaoyang Lyu, Marwan Taher, Xiaojuan Qi, Andrew J. Davison
We introduce EscherNet, a multi-view conditioned diffusion model for view synthesis.
no code implementations • 4 Jan 2024 • Marwan Taher, Ignacio Alzugaray, Andrew J. Davison
Accurate 3D object pose estimation is key to enabling many robotic applications that involve challenging object interactions.
1 code implementation • CVPR 2024 • Hidenobu Matsuki, Riku Murai, Paul H. J. Kelly, Andrew J. Davison
We present the first application of 3D Gaussian Splatting in monocular SLAM, the most fundamental but the hardest setup for Visual SLAM.
no code implementations • CVPR 2024 • Kirill Mazur, Gwangbin Bae, Andrew J. Davison
We address this issue with a new image representation which we call a SuperPrimitive.
no code implementations • CVPR 2023 • Eric Dexheimer, Andrew J. Davison
We propose learning a depth covariance function with applications to geometric vision tasks.
1 code implementation • CVPR 2023 • Xin Kong, Shikun Liu, Marwan Taher, Andrew J. Davison
We present vMAP, an object-level dense SLAM system using neural field representations.
no code implementations • 31 Oct 2022 • Iain Haughton, Edgar Sucar, Andre Mouton, Edward Johns, Andrew J. Davison
Neural fields can be trained from scratch to represent the shape and appearance of 3D scenes efficiently.
no code implementations • 6 Oct 2022 • Kirill Mazur, Edgar Sucar, Andrew J. Davison
General scene understanding for robotics requires flexible semantic representation, so that novel objects and structures which may not have been known at training time can be identified, segmented and grouped.
no code implementations • 9 Aug 2022 • Binbin Xu, Andrew J. Davison, Stefan Leutenegger
In this paper, we propose a novel object-level mapping system that can simultaneously segment, track, and reconstruct objects in dynamic scenes.
1 code implementation • 22 Mar 2022 • Aalok Patwardhan, Riku Murai, Andrew J. Davison
Precise coordinated planning over a forward time window enables safe and highly efficient motion when many robots must work together in tight spaces, but this would normally require centralised control of all devices which is difficult to scale.
no code implementations • 15 Mar 2022 • Tristan Laidlow, Andrew J. Davison
Human-made environments contain a lot of structure, and we seek to take advantage of this by enabling the use of quadric surfaces as features in SLAM systems.
1 code implementation • 22 Feb 2022 • Kentaro Wada, Stephen James, Andrew J. Davison
Robots need the capability of placing objects in arbitrary, specific poses to rearrange the world and achieve various valuable tasks.
1 code implementation • 11 Feb 2022 • Kentaro Wada, Stephen James, Andrew J. Davison
We evaluate our methods using the YCB objects in both simulation and the real world, achieving safe object extraction from piles.
no code implementations • 7 Feb 2022 • Riku Murai, Joseph Ortiz, Sajad Saeedi, Paul H. J. Kelly, Andrew J. Davison
We show that a distributed network of robots or other devices which make measurements of each other can collaborate to globally localise via efficient ad-hoc peer to peer communication.
1 code implementation • 7 Feb 2022 • Shikun Liu, Stephen James, Andrew J. Davison, Edward Johns
Unlike previous methods where task relationships are assumed to be fixed, Auto-Lambda is a gradient-based meta learning framework which explores continuous, dynamic task relationships via task-specific weightings, and can optimise any choice of combination of tasks through the formulation of a meta-loss; where the validation loss automatically influences task weightings throughout training.
Ranked #3 on
Robot Manipulation
on RLBench
(Succ. Rate (10 tasks, 100 demos/task) metric)
no code implementations • 29 Nov 2021 • Shuaifeng Zhi, Edgar Sucar, Andre Mouton, Iain Haughton, Tristan Laidlow, Andrew J. Davison
ILabel's underlying model is a multilayer perceptron (MLP) trained from scratch in real-time to learn a joint neural scene representation.
no code implementations • 13 Sep 2021 • Joseph Ortiz, Talfan Evans, Edgar Sucar, Andrew J. Davison
Scene graphs represent the key components of a scene in a compact and semantically rich way, but are difficult to build during incremental SLAM operation because of the challenges of robustly identifying abstract scene elements and optimising continually changing, complex graphs.
no code implementations • 19 Jul 2021 • Hidenobu Matsuki, Raluca Scona, Jan Czarnowski, Andrew J. Davison
In this paper we propose a dense mapping framework to complement sparse visual SLAM systems which takes as input the camera poses, keyframes and sparse points produced by the SLAM system and predicts a dense depth image for every keyframe.
no code implementations • 5 Jul 2021 • Joseph Ortiz, Talfan Evans, Andrew J. Davison
In this article, we present a visual introduction to Gaussian Belief Propagation (GBP), an approximate probabilistic inference algorithm that operates by passing messages between the nodes of arbitrarily structured factor graphs.
1 code implementation • CVPR 2022 • Stephen James, Kentaro Wada, Tristan Laidlow, Andrew J. Davison
We present a coarse-to-fine discretisation method that enables the use of discrete reinforcement learning approaches in place of unstable and data-inefficient actor-critic methods in continuous robotics domains.
Ranked #12 on
Robot Manipulation
on RLBench
1 code implementation • 31 May 2021 • Stephen James, Andrew J. Davison
Despite the success of reinforcement learning methods, they have yet to have their breakthrough moment when applied to a broad range of robotic manipulation tasks.
2 code implementations • ICLR 2022 • Shikun Liu, Shuaifeng Zhi, Edward Johns, Andrew J. Davison
We present ReCo, a contrastive learning framework designed at a regional level to assist learning in semantic segmentation.
no code implementations • ICCV 2021 • Zoe Landgraf, Raluca Scona, Tristan Laidlow, Stephen James, Stefan Leutenegger, Andrew J. Davison
At test time, our model can generate 3D shape and instance segmentation from a single depth view, probabilistically sampling proposals for the occluded region from the learned latent space.
no code implementations • ICCV 2021 • Shuaifeng Zhi, Tristan Laidlow, Stefan Leutenegger, Andrew J. Davison
Semantic labelling is highly correlated with geometry and radiance reconstruction, as scene entities with similar shape and appearance are more likely to come from similar classes.
3 code implementations • ICCV 2021 • Edgar Sucar, Shikun Liu, Joseph Ortiz, Andrew J. Davison
We show for the first time that a multilayer perceptron (MLP) can serve as the only scene representation in a real-time SLAM system for a handheld RGB-D camera.
2 code implementations • 15 Feb 2021 • Daniel Lenton, Stephen James, Ronald Clark, Andrew J. Davison
Spatial memory, or the ability to remember and recall specific locations and objects, is central to autonomous agents' ability to carry out tasks in real environments.
no code implementations • 3 Nov 2020 • Dhruv Batra, Angel X. Chang, Sonia Chernova, Andrew J. Davison, Jia Deng, Vladlen Koltun, Sergey Levine, Jitendra Malik, Igor Mordatch, Roozbeh Mottaghi, Manolis Savva, Hao Su
In the rearrangement task, the goal is to bring a given physical environment into a specified state.
1 code implementation • 31 Aug 2020 • Binbin Xu, Andrew J. Davison, Stefan Leutenegger
Dense image alignment from RGB-D images remains a critical issue for real-world applications, especially under challenging lighting conditions and in a wide baseline setting.
1 code implementation • CVPR 2020 • Kentaro Wada, Edgar Sucar, Stephen James, Daniel Lenton, Andrew J. Davison
Robots and other smart devices need efficient object-based scene representations from their on-board vision systems to reason about contact, physics and occlusion.
1 code implementation • CVPR 2020 • Joseph Ortiz, Mark Pupilli, Stefan Leutenegger, Andrew J. Davison
Graph processors such as Graphcore's Intelligence Processing Unit (IPU) are part of the major new wave of novel computer architecture for AI, and have a general design with massively parallel computation, distributed on-chip memory and very high inter-core communication bandwidth which allows breakthrough performance for message passing algorithms on arbitrary graphs.
1 code implementation • 14 Jan 2020 • Jan Czarnowski, Tristan Laidlow, Ronald Clark, Andrew J. Davison
The ability to estimate rich geometry and camera motion from monocular imagery is fundamental to future interactive robotics and augmented reality applications.
1 code implementation • 4 Nov 2019 • Alessandro Bonardi, Stephen James, Andrew J. Davison
But is there a way to remove the need for real world human demonstrations during training?
no code implementations • 30 Oct 2019 • Andrew J. Davison, Joseph Ortiz
We argue the case for Gaussian Belief Propagation (GBP) as a strong algorithmic framework for the distributed, generic and incremental probabilistic estimation we need in Spatial AI as we aim at high performance smart robots and devices which operate within the constraints of real products.
2 code implementations • 26 Sep 2019 • Stephen James, Zicong Ma, David Rovick Arrojo, Andrew J. Davison
We present a challenging new benchmark and learning-environment for robot learning: RLBench.
1 code implementation • 26 Jun 2019 • Stephen James, Marc Freese, Andrew J. Davison
PyRep is a toolkit for robot learning research, built on top of the virtual robotics experimentation platform (V-REP).
no code implementations • CVPR 2019 • Shuaifeng Zhi, Michael Bloesch, Stefan Leutenegger, Andrew J. Davison
Systems which incrementally create 3D semantic maps from image sequences must store and update representations of both geometry and semantic entities.
4 code implementations • NeurIPS 2019 • Shikun Liu, Andrew J. Davison, Edward Johns
The loss for the label-generation network incorporates the loss of the multi-task network, and so this interaction between the two networks can be seen as a form of meta learning with a double gradient.
3 code implementations • 8 Oct 2018 • Stephen James, Michael Bloesch, Andrew J. Davison
Despite this, most robot learning approaches have focused on learning a single task, from scratch, with a limited notion of generalisation, and no way of leveraging the knowledge to learn other tasks more efficiently.
no code implementations • ECCV 2018 • Ronald Clark, Michael Bloesch, Jan Czarnowski, Stefan Leutenegger, Andrew J. Davison
In this paper, we propose LS-Net, a neural nonlinear least squares optimization algorithm which learns to effectively optimize these cost functions even in the presence of adversities.
no code implementations • 25 Aug 2018 • John McCormac, Ronald Clark, Michael Bloesch, Andrew J. Davison, Stefan Leutenegger
Reconstructed objects are stored in an optimisable 6DoF pose graph which is our only persistent map representation.
2 code implementations • 20 Aug 2018 • Sajad Saeedi, Bruno Bodin, Harry Wagstaff, Andy Nisbet, Luigi Nardi, John Mawer, Nicolas Melot, Oscar Palomar, Emanuele Vespa, Tom Spink, Cosmin Gorgovan, Andrew Webb, James Clarkson, Erik Tomusk, Thomas Debrunner, Kuba Kaszyk, Pablo Gonzalez-de-Aledo, Andrey Rodchenko, Graham Riley, Christos Kotselidis, Björn Franke, Michael F. P. O'Boyle, Andrew J. Davison, Paul H. J. Kelly, Mikel Luján, Steve Furber
Visual understanding of 3D environments in real-time, at low power, is a huge computational challenge.
1 code implementation • 20 Jun 2018 • Jan Matas, Stephen James, Andrew J. Davison
Moreover, due to the large amount of data needed to learn these end-to-end solutions, an emerging trend is to learn control policies in simulation and then transfer them over to the real world.
Deep Reinforcement Learning
Deformable Object Manipulation
+3
1 code implementation • CVPR 2018 • Michael Bloesch, Jan Czarnowski, Ronald Clark, Stefan Leutenegger, Andrew J. Davison
Our approach is suitable for use in a keyframe-based monocular dense SLAM system: While each keyframe with a code can produce a depth map, the code can be optimised efficiently jointly with pose variables and together with the codes of overlapping keyframes to attain global consistency.
3 code implementations • 3 Apr 2018 • Michael Bloesch, Jan Czarnowski, Ronald Clark, Stefan Leutenegger, Andrew J. Davison
Our approach is suitable for use in a keyframe-based monocular dense SLAM system: While each keyframe with a code can produce a depth map, the code can be optimised efficiently jointly with pose variables and together with the codes of overlapping keyframes to attain global consistency.
no code implementations • 29 Mar 2018 • Andrew J. Davison
We discuss and predict the evolution of Simultaneous Localisation and Mapping (SLAM) into a general geometric and semantic `Spatial AI' perception capability for intelligent embodied devices.
4 code implementations • CVPR 2019 • Shikun Liu, Edward Johns, Andrew J. Davison
Our design, the Multi-Task Attention Network (MTAN), consists of a single shared network containing a global feature pool, together with a soft-attention module for each task.
no code implementations • ICCV 2017 • John McCormac, Ankur Handa, Stefan Leutenegger, Andrew J. Davison
We compare the semantic segmentation performance of network weights produced from pre-training on RGB images from our dataset against generic VGG-16 ImageNet weights.
1 code implementation • 7 Jul 2017 • Stephen James, Andrew J. Davison, Edward Johns
End-to-end control for robot manipulation and grasping is emerging as an attractive alternative to traditional pipelined approaches.
no code implementations • 2 Feb 2017 • Luigi Nardi, Bruno Bodin, Sajad Saeedi, Emanuele Vespa, Andrew J. Davison, Paul H. J. Kelly
In this paper we investigate an emerging application, 3D scene understanding, likely to be significant in the mobile space in the near future.
1 code implementation • 15 Dec 2016 • John McCormac, Ankur Handa, Stefan Leutenegger, Andrew J. Davison
We introduce SceneNet RGB-D, expanding the previous work of SceneNet to enable large scale photorealistic rendering of indoor scene trajectories.
no code implementations • 7 Aug 2016 • Edward Johns, Stefan Leutenegger, Andrew J. Davison
With this, it is possible to achieve grasping robust to the gripper's pose uncertainty, by smoothing the grasp function with the pose uncertainty function.
no code implementations • CVPR 2016 • Patrick Bardow, Andrew J. Davison, Stefan Leutenegger
In a series of examples, we demonstrate the successful operation of our framework, including in situations where conventional cameras heavily suffer from dynamic range limitations or motion blur.
no code implementations • CVPR 2016 • Edward Johns, Stefan Leutenegger, Andrew J. Davison
A multi-view image sequence provides a much richer capacity for object recognition than from a single image.
no code implementations • 29 Apr 2016 • Chourmouzios Tsiotsios, Maria E. Angelopoulou, Andrew J. Davison, Tae-Kyun Kim
Backscatter corresponds to a complex term with several unknown variables, and makes the problem of normal estimation hard.
no code implementations • 15 Sep 2015 • M. Zeeshan Zia, Luigi Nardi, Andrew Jack, Emanuele Vespa, Bruno Bodin, Paul H. J. Kelly, Andrew J. Davison
SLAM has matured significantly over the past few years, and is beginning to appear in serious commercial products.
3 code implementations • 8 Oct 2014 • Luigi Nardi, Bruno Bodin, M. Zeeshan Zia, John Mawer, Andy Nisbet, Paul H. J. Kelly, Andrew J. Davison, Mikel Luján, Michael F. P. O'Boyle, Graham Riley, Nigel Topham, Steve Furber
Real-time dense computer vision and SLAM offer great potential for a new level of scene modelling, tracking and real environmental interaction for many types of robot, but their high computational requirements mean that use on mass market embedded platforms is challenging.
no code implementations • CVPR 2014 • Chourmouzios Tsiotsios, Maria E. Angelopoulou, Tae-Kyun Kim, Andrew J. Davison
We compare our method with previous approaches through extensive experimental results, where a variety of objects are imaged in a big water tank whose turbidity is systematically increased, and show reconstruction quality which degrades little relative to clean water results even with a very significant scattering level.
no code implementations • CVPR 2013 • Renato F. Salas-Moreno, Richard A. Newcombe, Hauke Strasdat, Paul H. J. Kelly, Andrew J. Davison
We present the major advantages of a new 'object oriented' 3D SLAM paradigm, which takes full advantage in the loop of prior knowledge that many scenes consist of repeated, domain-specific objects and structures.
no code implementations • ISMAR 2011 • Richard A. Newcombe, Shahram Izadi, Otmar Hilliges, David Molyneaux, David Kim, Andrew J. Davison, Pushmeet Kohli, Jamie Shotton, Steve Hodges, Andrew Fitzgibbon
We present a system for accurate real-time mapping of complex and arbitrary indoor scenes in variable lighting conditions, using only a moving low-cost depth camera and commodity graphics hardware.