Search Results for author: Andrew J. Davison

Found 62 papers, 29 papers with code

COMO: Compact Mapping and Odometry

no code implementations4 Apr 2024 Eric Dexheimer, Andrew J. Davison

We present COMO, a real-time monocular mapping and odometry system that encodes dense geometry via a compact set of 3D anchor points.

U-ARE-ME: Uncertainty-Aware Rotation Estimation in Manhattan Environments

no code implementations22 Mar 2024 Aalok Patwardhan, Callum Rhodes, Gwangbin Bae, Andrew J. Davison

Given a sequence of images, we can use the per-frame rotation estimates and their uncertainty to perform multi-frame optimisation, achieving robustness and temporal consistency.

Rethinking Inductive Biases for Surface Normal Estimation

1 code implementation1 Mar 2024 Gwangbin Bae, Andrew J. Davison

Despite the growing demand for accurate surface normal estimation models, existing methods use general-purpose dense prediction models, adopting the same inductive biases as other tasks.

Surface Normal Estimation

Fit-NGP: Fitting Object Models to Neural Graphics Primitives

no code implementations4 Jan 2024 Marwan Taher, Ignacio Alzugaray, Andrew J. Davison

Accurate 3D object pose estimation is key to enabling many robotic applications that involve challenging object interactions.

Object Pose Estimation

Gaussian Splatting SLAM

no code implementations11 Dec 2023 Hidenobu Matsuki, Riku Murai, Paul H. J. Kelly, Andrew J. Davison

We present the first application of 3D Gaussian Splatting in monocular SLAM, the most fundamental but the hardest setup for Visual SLAM.

3D Reconstruction Novel View Synthesis +1

Learning a Depth Covariance Function

no code implementations CVPR 2023 Eric Dexheimer, Andrew J. Davison

We propose learning a depth covariance function with applications to geometric vision tasks.

Depth Completion Visual Odometry

vMAP: Vectorised Object Mapping for Neural Field SLAM

1 code implementation CVPR 2023 Xin Kong, Shikun Liu, Marwan Taher, Andrew J. Davison

We present vMAP, an object-level dense SLAM system using neural field representations.

Object

Real-time Mapping of Physical Scene Properties with an Autonomous Robot Experimenter

no code implementations31 Oct 2022 Iain Haughton, Edgar Sucar, Andre Mouton, Edward Johns, Andrew J. Davison

Neural fields can be trained from scratch to represent the shape and appearance of 3D scenes efficiently.

Feature-Realistic Neural Fusion for Real-Time, Open Set Scene Understanding

no code implementations6 Oct 2022 Kirill Mazur, Edgar Sucar, Andrew J. Davison

General scene understanding for robotics requires flexible semantic representation, so that novel objects and structures which may not have been known at training time can be identified, segmented and grouped.

Scene Understanding

Learning to Complete Object Shapes for Object-level Mapping in Dynamic Scenes

no code implementations9 Aug 2022 Binbin Xu, Andrew J. Davison, Stefan Leutenegger

In this paper, we propose a novel object-level mapping system that can simultaneously segment, track, and reconstruct objects in dynamic scenes.

Instance Segmentation Object +2

Distributing Collaborative Multi-Robot Planning with Gaussian Belief Propagation

1 code implementation22 Mar 2022 Aalok Patwardhan, Riku Murai, Andrew J. Davison

Precise coordinated planning over a forward time window enables safe and highly efficient motion when many robots must work together in tight spaces, but this would normally require centralised control of all devices which is difficult to scale.

Simultaneous Localisation and Mapping with Quadric Surfaces

no code implementations15 Mar 2022 Tristan Laidlow, Andrew J. Davison

Human-made environments contain a lot of structure, and we seek to take advantage of this by enabling the use of quadric surfaces as features in SLAM systems.

ReorientBot: Learning Object Reorientation for Specific-Posed Placement

1 code implementation22 Feb 2022 Kentaro Wada, Stephen James, Andrew J. Davison

Robots need the capability of placing objects in arbitrary, specific poses to rearrange the world and achieve various valuable tasks.

Motion Planning Object +2

SafePicking: Learning Safe Object Extraction via Object-Level Mapping

1 code implementation11 Feb 2022 Kentaro Wada, Stephen James, Andrew J. Davison

We evaluate our methods using the YCB objects in both simulation and the real world, achieving safe object extraction from piles.

Motion Planning Object +2

Auto-Lambda: Disentangling Dynamic Task Relationships

1 code implementation7 Feb 2022 Shikun Liu, Stephen James, Andrew J. Davison, Edward Johns

Unlike previous methods where task relationships are assumed to be fixed, Auto-Lambda is a gradient-based meta learning framework which explores continuous, dynamic task relationships via task-specific weightings, and can optimise any choice of combination of tasks through the formulation of a meta-loss; where the validation loss automatically influences task weightings throughout training.

Ranked #3 on Robot Manipulation on RLBench (Succ. Rate (10 tasks, 100 demos/task) metric)

Auxiliary Learning Meta-Learning +2

A Robot Web for Distributed Many-Device Localisation

no code implementations7 Feb 2022 Riku Murai, Joseph Ortiz, Sajad Saeedi, Paul H. J. Kelly, Andrew J. Davison

We show that a distributed network of robots or other devices which make measurements of each other can collaborate to globally localise via efficient ad-hoc peer to peer communication.

ILabel: Interactive Neural Scene Labelling

no code implementations29 Nov 2021 Shuaifeng Zhi, Edgar Sucar, Andre Mouton, Iain Haughton, Tristan Laidlow, Andrew J. Davison

ILabel's underlying model is a multilayer perceptron (MLP) trained from scratch in real-time to learn a joint neural scene representation.

Semantic Segmentation

Incremental Abstraction in Distributed Probabilistic SLAM Graphs

no code implementations13 Sep 2021 Joseph Ortiz, Talfan Evans, Edgar Sucar, Andrew J. Davison

Scene graphs represent the key components of a scene in a compact and semantically rich way, but are difficult to build during incremental SLAM operation because of the challenges of robustly identifying abstract scene elements and optimising continually changing, complex graphs.

CodeMapping: Real-Time Dense Mapping for Sparse SLAM using Compact Scene Representations

no code implementations19 Jul 2021 Hidenobu Matsuki, Raluca Scona, Jan Czarnowski, Andrew J. Davison

In this paper we propose a dense mapping framework to complement sparse visual SLAM systems which takes as input the camera poses, keyframes and sparse points produced by the SLAM system and predicts a dense depth image for every keyframe.

3D Reconstruction Depth Estimation +1

A visual introduction to Gaussian Belief Propagation

no code implementations5 Jul 2021 Joseph Ortiz, Talfan Evans, Andrew J. Davison

In this article, we present a visual introduction to Gaussian Belief Propagation (GBP), an approximate probabilistic inference algorithm that operates by passing messages between the nodes of arbitrarily structured factor graphs.

Coarse-to-Fine Q-attention: Efficient Learning for Visual Robotic Manipulation via Discretisation

1 code implementation CVPR 2022 Stephen James, Kentaro Wada, Tristan Laidlow, Andrew J. Davison

We present a coarse-to-fine discretisation method that enables the use of discrete reinforcement learning approaches in place of unstable and data-inefficient actor-critic methods in continuous robotics domains.

Continuous Control Q-Learning +2

Q-attention: Enabling Efficient Learning for Vision-based Robotic Manipulation

1 code implementation31 May 2021 Stephen James, Andrew J. Davison

Despite the success of reinforcement learning methods, they have yet to have their breakthrough moment when applied to a broad range of robotic manipulation tasks.

reinforcement-learning Reinforcement Learning (RL) +1

SIMstack: A Generative Shape and Instance Model for Unordered Object Stacks

no code implementations ICCV 2021 Zoe Landgraf, Raluca Scona, Tristan Laidlow, Stephen James, Stefan Leutenegger, Andrew J. Davison

At test time, our model can generate 3D shape and instance segmentation from a single depth view, probabilistically sampling proposals for the occluded region from the learned latent space.

Instance Segmentation Segmentation +2

In-Place Scene Labelling and Understanding with Implicit Scene Representation

no code implementations ICCV 2021 Shuaifeng Zhi, Tristan Laidlow, Stefan Leutenegger, Andrew J. Davison

Semantic labelling is highly correlated with geometry and radiance reconstruction, as scene entities with similar shape and appearance are more likely to come from similar classes.

Denoising Super-Resolution

iMAP: Implicit Mapping and Positioning in Real-Time

3 code implementations ICCV 2021 Edgar Sucar, Shikun Liu, Joseph Ortiz, Andrew J. Davison

We show for the first time that a multilayer perceptron (MLP) can serve as the only scene representation in a real-time SLAM system for a handheld RGB-D camera.

End-to-End Egospheric Spatial Memory

2 code implementations15 Feb 2021 Daniel Lenton, Stephen James, Ronald Clark, Andrew J. Davison

Spatial memory, or the ability to remember and recall specific locations and objects, is central to autonomous agents' ability to carry out tasks in real environments.

General Reinforcement Learning Imitation Learning +3

Deep Probabilistic Feature-metric Tracking

1 code implementation31 Aug 2020 Binbin Xu, Andrew J. Davison, Stefan Leutenegger

Dense image alignment from RGB-D images remains a critical issue for real-world applications, especially under challenging lighting conditions and in a wide baseline setting.

Object Tracking

MoreFusion: Multi-object Reasoning for 6D Pose Estimation from Volumetric Fusion

1 code implementation CVPR 2020 Kentaro Wada, Edgar Sucar, Stephen James, Daniel Lenton, Andrew J. Davison

Robots and other smart devices need efficient object-based scene representations from their on-board vision systems to reason about contact, physics and occlusion.

6D Pose Estimation Object

Bundle Adjustment on a Graph Processor

1 code implementation CVPR 2020 Joseph Ortiz, Mark Pupilli, Stefan Leutenegger, Andrew J. Davison

Graph processors such as Graphcore's Intelligence Processing Unit (IPU) are part of the major new wave of novel computer architecture for AI, and have a general design with massively parallel computation, distributed on-chip memory and very high inter-core communication bandwidth which allows breakthrough performance for message passing algorithms on arbitrary graphs.

DeepFactors: Real-Time Probabilistic Dense Monocular SLAM

1 code implementation14 Jan 2020 Jan Czarnowski, Tristan Laidlow, Ronald Clark, Andrew J. Davison

The ability to estimate rich geometry and camera motion from monocular imagery is fundamental to future interactive robotics and augmented reality applications.

Learning One-Shot Imitation from Humans without Humans

1 code implementation4 Nov 2019 Alessandro Bonardi, Stephen James, Andrew J. Davison

But is there a way to remove the need for real world human demonstrations during training?

Imitation Learning Meta-Learning

FutureMapping 2: Gaussian Belief Propagation for Spatial AI

no code implementations30 Oct 2019 Andrew J. Davison, Joseph Ortiz

We argue the case for Gaussian Belief Propagation (GBP) as a strong algorithmic framework for the distributed, generic and incremental probabilistic estimation we need in Spatial AI as we aim at high performance smart robots and devices which operate within the constraints of real products.

PyRep: Bringing V-REP to Deep Robot Learning

1 code implementation26 Jun 2019 Stephen James, Marc Freese, Andrew J. Davison

PyRep is a toolkit for robot learning research, built on top of the virtual robotics experimentation platform (V-REP).

Imitation Learning reinforcement-learning +1

SceneCode: Monocular Dense Semantic Reconstruction using Learned Encoded Scene Representations

no code implementations CVPR 2019 Shuaifeng Zhi, Michael Bloesch, Stefan Leutenegger, Andrew J. Davison

Systems which incrementally create 3D semantic maps from image sequences must store and update representations of both geometry and semantic entities.

Self-Supervised Generalisation with Meta Auxiliary Learning

4 code implementations NeurIPS 2019 Shikun Liu, Andrew J. Davison, Edward Johns

The loss for the label-generation network incorporates the loss of the multi-task network, and so this interaction between the two networks can be seen as a form of meta learning with a double gradient.

Auxiliary Learning Meta-Learning +1

Task-Embedded Control Networks for Few-Shot Imitation Learning

3 code implementations8 Oct 2018 Stephen James, Michael Bloesch, Andrew J. Davison

Despite this, most robot learning approaches have focused on learning a single task, from scratch, with a limited notion of generalisation, and no way of leveraging the knowledge to learn other tasks more efficiently.

Few-Shot Imitation Learning Imitation Learning +1

LS-Net: Learning to Solve Nonlinear Least Squares for Monocular Stereo

no code implementations ECCV 2018 Ronald Clark, Michael Bloesch, Jan Czarnowski, Stefan Leutenegger, Andrew J. Davison

In this paper, we propose LS-Net, a neural nonlinear least squares optimization algorithm which learns to effectively optimize these cost functions even in the presence of adversities.

Fusion++: Volumetric Object-Level SLAM

no code implementations25 Aug 2018 John McCormac, Ronald Clark, Michael Bloesch, Andrew J. Davison, Stefan Leutenegger

Reconstructed objects are stored in an optimisable 6DoF pose graph which is our only persistent map representation.

Loop Closure Detection Object

Sim-to-Real Reinforcement Learning for Deformable Object Manipulation

1 code implementation20 Jun 2018 Jan Matas, Stephen James, Andrew J. Davison

Moreover, due to the large amount of data needed to learn these end-to-end solutions, an emerging trend is to learn control policies in simulation and then transfer them over to the real world.

Deformable Object Manipulation Object +2

CodeSLAM — Learning a Compact, Optimisable Representation for Dense Visual SLAM

1 code implementation CVPR 2018 Michael Bloesch, Jan Czarnowski, Ronald Clark, Stefan Leutenegger, Andrew J. Davison

Our approach is suitable for use in a keyframe-based monocular dense SLAM system: While each keyframe with a code can produce a depth map, the code can be optimised efficiently jointly with pose variables and together with the codes of overlapping keyframes to attain global consistency.

CodeSLAM - Learning a Compact, Optimisable Representation for Dense Visual SLAM

3 code implementations3 Apr 2018 Michael Bloesch, Jan Czarnowski, Ronald Clark, Stefan Leutenegger, Andrew J. Davison

Our approach is suitable for use in a keyframe-based monocular dense SLAM system: While each keyframe with a code can produce a depth map, the code can be optimised efficiently jointly with pose variables and together with the codes of overlapping keyframes to attain global consistency.

FutureMapping: The Computational Structure of Spatial AI Systems

no code implementations29 Mar 2018 Andrew J. Davison

We discuss and predict the evolution of Simultaneous Localisation and Mapping (SLAM) into a general geometric and semantic `Spatial AI' perception capability for intelligent embodied devices.

End-to-End Multi-Task Learning with Attention

4 code implementations CVPR 2019 Shikun Liu, Edward Johns, Andrew J. Davison

Our design, the Multi-Task Attention Network (MTAN), consists of a single shared network containing a global feature pool, together with a soft-attention module for each task.

Multi-Task Learning

SceneNet RGB-D: Can 5M Synthetic Images Beat Generic ImageNet Pre-Training on Indoor Segmentation?

no code implementations ICCV 2017 John McCormac, Ankur Handa, Stefan Leutenegger, Andrew J. Davison

We compare the semantic segmentation performance of network weights produced from pre-training on RGB images from our dataset against generic VGG-16 ImageNet weights.

16k Instance Segmentation +7

Transferring End-to-End Visuomotor Control from Simulation to Real World for a Multi-Stage Task

1 code implementation7 Jul 2017 Stephen James, Andrew J. Davison, Edward Johns

End-to-end control for robot manipulation and grasping is emerging as an attractive alternative to traditional pipelined approaches.

Robotic Grasping Robot Manipulation

Algorithmic Performance-Accuracy Trade-off in 3D Vision Applications Using HyperMapper

no code implementations2 Feb 2017 Luigi Nardi, Bruno Bodin, Sajad Saeedi, Emanuele Vespa, Andrew J. Davison, Paul H. J. Kelly

In this paper we investigate an emerging application, 3D scene understanding, likely to be significant in the mobile space in the near future.

Active Learning Scene Understanding

SceneNet RGB-D: 5M Photorealistic Images of Synthetic Indoor Trajectories with Ground Truth

1 code implementation15 Dec 2016 John McCormac, Ankur Handa, Stefan Leutenegger, Andrew J. Davison

We introduce SceneNet RGB-D, expanding the previous work of SceneNet to enable large scale photorealistic rendering of indoor scene trajectories.

3D Reconstruction Depth Estimation +7

Deep Learning a Grasp Function for Grasping under Gripper Pose Uncertainty

no code implementations7 Aug 2016 Edward Johns, Stefan Leutenegger, Andrew J. Davison

With this, it is possible to achieve grasping robust to the gripper's pose uncertainty, by smoothing the grasp function with the pose uncertainty function.

Simultaneous Optical Flow and Intensity Estimation From an Event Camera

no code implementations CVPR 2016 Patrick Bardow, Andrew J. Davison, Stefan Leutenegger

In a series of examples, we demonstrate the successful operation of our framework, including in situations where conventional cameras heavily suffer from dynamic range limitations or motion blur.

Optical Flow Estimation

Effective Backscatter Approximation for Photometry in Murky Water

no code implementations29 Apr 2016 Chourmouzios Tsiotsios, Maria E. Angelopoulou, Andrew J. Davison, Tae-Kyun Kim

Backscatter corresponds to a complex term with several unknown variables, and makes the problem of normal estimation hard.

Image Restoration

Comparative Design Space Exploration of Dense and Semi-Dense SLAM

no code implementations15 Sep 2015 M. Zeeshan Zia, Luigi Nardi, Andrew Jack, Emanuele Vespa, Bruno Bodin, Paul H. J. Kelly, Andrew J. Davison

SLAM has matured significantly over the past few years, and is beginning to appear in serious commercial products.

Benchmarking

Introducing SLAMBench, a performance and accuracy benchmarking methodology for SLAM

3 code implementations8 Oct 2014 Luigi Nardi, Bruno Bodin, M. Zeeshan Zia, John Mawer, Andy Nisbet, Paul H. J. Kelly, Andrew J. Davison, Mikel Luján, Michael F. P. O'Boyle, Graham Riley, Nigel Topham, Steve Furber

Real-time dense computer vision and SLAM offer great potential for a new level of scene modelling, tracking and real environmental interaction for many types of robot, but their high computational requirements mean that use on mass market embedded platforms is challenging.

Benchmarking

Backscatter Compensated Photometric Stereo with 3 Sources

no code implementations CVPR 2014 Chourmouzios Tsiotsios, Maria E. Angelopoulou, Tae-Kyun Kim, Andrew J. Davison

We compare our method with previous approaches through extensive experimental results, where a variety of objects are imaged in a big water tank whose turbidity is systematically increased, and show reconstruction quality which degrades little relative to clean water results even with a very significant scattering level.

SLAM++: Simultaneous Localisation and Mapping at the Level of Objects

no code implementations CVPR 2013 Renato F. Salas-Moreno, Richard A. Newcombe, Hauke Strasdat, Paul H. J. Kelly, Andrew J. Davison

We present the major advantages of a new 'object oriented' 3D SLAM paradigm, which takes full advantage in the loop of prior knowledge that many scenes consist of repeated, domain-specific objects and structures.

3D Object Recognition Descriptive +2

KinectFusion: Real-Time Dense Surface Mapping and Tracking

no code implementations ISMAR 2011 Richard A. Newcombe, Shahram Izadi, Otmar Hilliges, David Molyneaux, David Kim, Andrew J. Davison, Pushmeet Kohli, Jamie Shotton, Steve Hodges, Andrew Fitzgibbon

We present a system for accurate real-time mapping of complex and arbitrary indoor scenes in variable lighting conditions, using only a moving low-cost depth camera and commodity graphics hardware.

Cannot find the paper you are looking for? You can Submit a new open access paper.