Search Results for author: Andrew Markham

Found 71 papers, 34 papers with code

VINet: Visual-Inertial Odometry as a Sequence-to-Sequence Learning Problem

no code implementations29 Jan 2017 Ronald Clark, Sen Wang, Hongkai Wen, Andrew Markham, Niki Trigoni

In this paper we present an on-manifold sequence-to-sequence learning approach to motion estimation using visual and inertial sensors.

Motion Estimation

VidLoc: A Deep Spatio-Temporal Model for 6-DoF Video-Clip Relocalization

no code implementations CVPR 2017 Ronald Clark, Sen Wang, Andrew Markham, Niki Trigoni, Hongkai Wen

Machine learning techniques, namely convolutional neural networks (CNN) and regression forests, have recently shown great promise in performing 6-DoF localization of monocular images.

Autonomous Driving Indoor Localization

3D Object Reconstruction from a Single Depth View with Adversarial Learning

2 code implementations26 Aug 2017 Bo Yang, Hongkai Wen, Sen Wang, Ronald Clark, Andrew Markham, Niki Trigoni

In this paper, we propose a novel 3D-RecGAN approach, which reconstructs the complete 3D structure of a given object from a single arbitrary depth view using generative adversarial networks.

3D Object Reconstruction Object

Learning from lions: inferring the utility of agents from their trajectories

no code implementations7 Sep 2017 Adam D. Cobb, Andrew Markham, Stephen J. Roberts

We build a model using Gaussian processes to infer a spatio-temporal vector field from observed agent trajectories.

Decision Making Gaussian Processes

IONet: Learning to Cure the Curse of Drift in Inertial Odometry

no code implementations30 Jan 2018 Changhao Chen, Xiaoxuan Lu, Andrew Markham, Niki Trigoni

Inertial sensors play a pivotal role in indoor localization, which in turn lays the foundation for pervasive personal applications.

Indoor Localization

Dense 3D Object Reconstruction from a Single Depth View

2 code implementations1 Feb 2018 Bo Yang, Stefano Rosa, Andrew Markham, Niki Trigoni, Hongkai Wen

Unlike existing work which typically requires multiple views of the same object or class labels to recover the full 3D geometry, the proposed 3D-RecGAN++ only takes the voxel grid representation of a depth view of the object as input, and is able to generate the complete 3D occupancy grid with a high resolution of 256^3 by recovering the occluded/missing regions.

3D Object Reconstruction Object

Defo-Net: Learning Body Deformation using Generative Adversarial Networks

1 code implementation16 Apr 2018 Zhihua Wang, Stefano Rosa, Linhai Xie, Bo Yang, Sen Wang, Niki Trigoni, Andrew Markham

Modelling the physical properties of everyday objects is a fundamental prerequisite for autonomous robots.

Robotics

3D-PhysNet: Learning the Intuitive Physics of Non-Rigid Object Deformations

1 code implementation25 Apr 2018 Zhihua Wang, Stefano Rosa, Bo Yang, Sen Wang, Niki Trigoni, Andrew Markham

This is further confounded by the fact that shape information about encountered objects in the real world is often impaired by occlusions, noise and missing regions e. g. a robot manipulating an object will only be able to observe a partial view of the entire solid.

Robust Attentional Aggregation of Deep Feature Sets for Multi-view 3D Reconstruction

1 code implementation2 Aug 2018 Bo Yang, Sen Wang, Andrew Markham, Niki Trigoni

However, GRU based approaches are unable to consistently estimate 3D shapes given different permutations of the same set of input images as the recurrent unit is permutation variant.

3D Object Reconstruction 3D Reconstruction +1

Neural Allocentric Intuitive Physics Prediction from Real Videos

no code implementations7 Sep 2018 Zhihua Wang, Stefano Rosa, Yishu Miao, Zihang Lai, Linhai Xie, Andrew Markham, Niki Trigoni

In this framework, real images are first converted to a synthetic domain representation that reduces complexity arising from lighting and texture.

GANVO: Unsupervised Deep Monocular Visual Odometry and Depth Estimation with Generative Adversarial Networks

no code implementations16 Sep 2018 Yasin Almalioglu, Muhamad Risqi U. Saputra, Pedro P. B. de Gusmao, Andrew Markham, Niki Trigoni

In the last decade, supervised deep learning approaches have been extensively employed in visual odometry (VO) applications, which is not feasible in environments where labelled data is not abundant.

Depth Estimation Monocular Visual Odometry +1

OxIOD: The Dataset for Deep Inertial Odometry

no code implementations20 Sep 2018 Changhao Chen, Peijun Zhao, Chris Xiaoxuan Lu, Wei Wang, Andrew Markham, Niki Trigoni

Advances in micro-electro-mechanical (MEMS) techniques enable inertial measurements units (IMUs) to be small, cheap, energy efficient, and widely used in smartphones, robots, and drones.

Transferring Physical Motion Between Domains for Neural Inertial Tracking

no code implementations4 Oct 2018 Changhao Chen, Yishu Miao, Chris Xiaoxuan Lu, Phil Blunsom, Andrew Markham, Niki Trigoni

Inertial information processing plays a pivotal role in ego-motion awareness for mobile agents, as inertial measurements are entirely egocentric and not environment dependent.

Domain Adaptation

Learning with Stochastic Guidance for Navigation

1 code implementation27 Nov 2018 Linhai Xie, Yishu Miao, Sen Wang, Phil Blunsom, Zhihua Wang, Changhao Chen, Andrew Markham, Niki Trigoni

Due to the sparse rewards and high degree of environment variation, reinforcement learning approaches such as Deep Deterministic Policy Gradient (DDPG) are plagued by issues of high variance when applied in complex real world environments.

Robotics

Selective Sensor Fusion for Neural Visual-Inertial Odometry

no code implementations CVPR 2019 Changhao Chen, Stefano Rosa, Yishu Miao, Chris Xiaoxuan Lu, Wei Wu, Andrew Markham, Niki Trigoni

Deep learning approaches for Visual-Inertial Odometry (VIO) have proven successful, but they rarely focus on incorporating robust fusion strategies for dealing with imperfect input sensory data.

Autonomous Driving Sensor Fusion

Learning Monocular Visual Odometry through Geometry-Aware Curriculum Learning

no code implementations25 Mar 2019 Muhamad Risqi U. Saputra, Pedro P. B. de Gusmao, Sen Wang, Andrew Markham, Niki Trigoni

Inspired by the cognitive process of humans and animals, Curriculum Learning (CL) trains a model by gradually increasing the difficulty of the training data.

Monocular Visual Odometry Optical Flow Estimation

Learning Object Bounding Boxes for 3D Instance Segmentation on Point Clouds

1 code implementation NeurIPS 2019 Bo Yang, Jianan Wang, Ronald Clark, Qingyong Hu, Sen Wang, Andrew Markham, Niki Trigoni

The framework directly regresses 3D bounding boxes for all instances in a point cloud, while simultaneously predicting a point-level mask for each instance.

Ranked #13 on 3D Instance Segmentation on S3DIS (mPrec metric)

3D Instance Segmentation Clustering +2

DynaNet: Neural Kalman Dynamical Model for Motion Estimation and Prediction

no code implementations11 Aug 2019 Changhao Chen, Chris Xiaoxuan Lu, Bing Wang, Niki Trigoni, Andrew Markham

In addition we show how DynaNet can indicate failures through investigation of properties such as the rate of innovation (Kalman Gain).

Motion Estimation Sensor Fusion +1

Autonomous Learning for Face Recognition in the Wild via Ambient Wireless Cues

1 code implementation14 Aug 2019 Chris Xiaoxuan Lu, Xuan Kan, Bowen Du, Changhao Chen, Hongkai Wen, Andrew Markham, Niki Trigoni, John Stankovic

Inspired by the fact that most people carry smart wireless devices with them, e. g. smartphones, we propose to use this wireless identifier as a supervisory label.

Face Recognition

AtLoc: Attention Guided Camera Localization

1 code implementation8 Sep 2019 Bing Wang, Changhao Chen, Chris Xiaoxuan Lu, Peijun Zhao, Niki Trigoni, Andrew Markham

Deep learning has achieved impressive results in camera localization, but current single-image techniques typically suffer from a lack of robustness, leading to large outliers.

Camera Localization Visual Localization

Milli-RIO: Ego-Motion Estimation with Millimetre-Wave Radar and Inertial Measurement Unit Sensor

no code implementations12 Sep 2019 Yasin Almalioglu, Mehmet Turan, Chris Xiaoxuan Lu, Niki Trigoni, Andrew Markham

With the fast-growing demand of location-based services in various indoor environments, robust indoor ego-motion estimation has attracted significant interest in the last decades.

Indoor Localization Motion Estimation +1

DeepPCO: End-to-End Point Cloud Odometry through Deep Parallel Neural Network

no code implementations13 Oct 2019 Wei Wang, Muhamad Risqi U. Saputra, Peijun Zhao, Pedro Gusmao, Bo Yang, Changhao Chen, Andrew Markham, Niki Trigoni

There is considerable work in the area of visual odometry (VO), and recent advances in deep learning have brought novel approaches to VO, which directly learn salient features from raw images.

Translation Visual Odometry

Introducing an Explicit Symplectic Integration Scheme for Riemannian Manifold Hamiltonian Monte Carlo

1 code implementation14 Oct 2019 Adam D. Cobb, Atılım Güneş Baydin, Andrew Markham, Stephen J. Roberts

We introduce a recent symplectic integration scheme derived for solving physically motivated systems with non-separable Hamiltonians.

Bayesian Inference

See Through Smoke: Robust Indoor Mapping with Low-cost mmWave Radar

1 code implementation1 Nov 2019 Chris Xiaoxuan Lu, Stefano Rosa, Peijun Zhao, Bing Wang, Changhao Chen, John A. Stankovic, Niki Trigoni, Andrew Markham

This paper presents the design, implementation and evaluation of milliMap, a single-chip millimetre wave (mmWave) radar based indoor mapping system targetted towards low-visibility environments to assist in emergency response.

SelfVIO: Self-Supervised Deep Monocular Visual-Inertial Odometry and Depth Estimation

no code implementations22 Nov 2019 Yasin Almalioglu, Mehmet Turan, Alp Eren Sari, Muhamad Risqi U. Saputra, Pedro P. B. de Gusmão, Andrew Markham, Niki Trigoni

In the last decade, numerous supervised deep learning approaches requiring large amounts of labeled data have been proposed for visual-inertial odometry (VIO) and depth map estimation.

Depth Estimation Pose Estimation +3

Snoopy: Sniffing Your Smartwatch Passwords via Deep Sequence Learning

1 code implementation10 Dec 2019 Chris Xiaoxuan Lu, Bowen Du, Hongkai Wen, Sen Wang, Andrew Markham, Ivan Martinovic, Yiran Shen, Niki Trigoni

Demand for smartwatches has taken off in recent years with new models which can run independently from smartphones and provide more useful features, becoming first-class mobile platforms.

Learning Selective Sensor Fusion for States Estimation

no code implementations30 Dec 2019 Changhao Chen, Stefano Rosa, Chris Xiaoxuan Lu, Bing Wang, Niki Trigoni, Andrew Markham

By integrating the observations from different sensors, these mobile agents are able to perceive the environment and estimate system states, e. g. locations and orientations.

Autonomous Vehicles Sensor Fusion

Deep Learning based Pedestrian Inertial Navigation: Methods, Dataset and On-Device Inference

no code implementations13 Jan 2020 Changhao Chen, Peijun Zhao, Chris Xiaoxuan Lu, Wei Wang, Andrew Markham, Niki Trigoni

Modern inertial measurements units (IMUs) are small, cheap, energy efficient, and widely employed in smart devices and mobile robots.

PointLoc: Deep Pose Regressor for LiDAR Point Cloud Localization

2 code implementations5 Mar 2020 Wei Wang, Bing Wang, Peijun Zhao, Changhao Chen, Ronald Clark, Bo Yang, Andrew Markham, Niki Trigoni

In this paper, we present a novel end-to-end learning-based LiDAR relocalization framework, termed PointLoc, which infers 6-DoF poses directly using only a single point cloud as input, without requiring a pre-built map.

Robotics

VMLoc: Variational Fusion For Learning-Based Multimodal Camera Localization

1 code implementation12 Mar 2020 Kaichen Zhou, Changhao Chen, Bing Wang, Muhamad Risqi U. Saputra, Niki Trigoni, Andrew Markham

We conjecture that this is because of the naive approaches to feature space fusion through summation or concatenation which do not take into account the different strengths of each modality.

Camera Relocalization Visual Localization

Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset, Benchmarks and Challenges

2 code implementations CVPR 2021 Qingyong Hu, Bo Yang, Sheikh Khalid, Wen Xiao, Niki Trigoni, Andrew Markham

An essential prerequisite for unleashing the potential of supervised deep learning algorithms in the area of 3D scene understanding is the availability of large-scale and richly annotated datasets.

Scene Understanding Semantic Segmentation

RadarLoc: Learning to Relocalize in FMCW Radar

no code implementations22 Mar 2021 Wei Wang, Pedro P. B. de Gusmo, Bo Yang, Andrew Markham, Niki Trigoni

There is considerable work in the field of deep camera relocalization, which directly estimates poses from raw images.

Camera Relocalization

SoundDet: Polyphonic Moving Sound Event Detection and Localization from Raw Waveform

no code implementations13 Jun 2021 Yuhang He, Niki Trigoni, Andrew Markham

Specifically, SoundDet consists of a backbone neural network and two parallel heads for temporal detection and spatial localization, respectively.

Event Detection Sound Event Detection

Finite Basis Physics-Informed Neural Networks (FBPINNs): a scalable domain decomposition approach for solving differential equations

1 code implementation16 Jul 2021 Ben Moseley, Andrew Markham, Tarje Nissen-Meyer

FBINNs are designed to address the spectral bias of neural networks by using separate input normalisation over each subdomain, and reduce the complexity of the underlying optimisation problem by using many smaller neural networks in a parallel divide-and-conquer approach.

Scaling physics-informed neural networks to large domains by using domain decomposition

no code implementations NeurIPS Workshop DLDE 2021 Ben Moseley, Andrew Markham, Tarje Nissen-Meyer

Recently, physics-informed neural networks (PINNs) have offered a powerful new paradigm for solving forward and inverse problems relating to differential equations.

CubeLearn: End-to-end Learning for Human Motion Recognition from Raw mmWave Radar Signals

1 code implementation7 Nov 2021 Peijun Zhao, Chris Xiaoxuan Lu, Bing Wang, Niki Trigoni, Andrew Markham

To avoid the drawbacks of conventional DFT pre-processing, we propose a learnable pre-processing module, named CubeLearn, to directly extract features from raw radar signal and build an end-to-end deep neural network for mmWave FMCW radar motion recognition applications.

Activity Recognition

DeepAoANet: Learning Angle of Arrival from Software Defined Radios with Deep Neural Networks

1 code implementation1 Dec 2021 Zhuangzhuang Dai, Yuhang He, Tran Vu, Niki Trigoni, Andrew Markham

To demonstrate the utility of our approach we have collected IQ (In-phase and Quadrature components) samples from a four-element Universal Linear Array (ULA) in various Light-of-Sight (LOS) and Non-Line-of-Sight (NLOS) environments, and published the dataset.

SensatUrban: Learning Semantics from Urban-Scale Photogrammetric Point Clouds

no code implementations12 Jan 2022 Qingyong Hu, Bo Yang, Sheikh Khalid, Wen Xiao, Niki Trigoni, Andrew Markham

Each point in the dataset has been labelled with fine-grained semantic annotations, resulting in a dataset that is three times the size of the previous existing largest photogrammetric point cloud dataset.

Meta-Sampler: Almost-Universal yet Task-Oriented Sampling for Point Clouds

1 code implementation30 Mar 2022 Ta-Ying Cheng, Qingyong Hu, Qian Xie, Niki Trigoni, Andrew Markham

In this work, we propose an almost-universal sampler, in our quest for a sampler that can learn to preserve the most useful points for a particular task, yet be inexpensive to adapt to different tasks, models, or datasets.

Computational Efficiency

RangeUDF: Semantic Surface Reconstruction from 3D Point Clouds

2 code implementations19 Apr 2022 Bing Wang, Zhengdi Yu, Bo Yang, Jie Qin, Toby Breckon, Ling Shao, Niki Trigoni, Andrew Markham

We present RangeUDF, a new implicit representation based framework to recover the geometry and semantics of continuous 3D scene surfaces from point clouds.

Semantic Segmentation Surface Reconstruction

When the Sun Goes Down: Repairing Photometric Losses for All-Day Depth Estimation

no code implementations28 Jun 2022 Madhu Vankadari, Stuart Golodetz, Sourav Garg, Sangyun Shin, Andrew Markham, Niki Trigoni

In this paper, we show how to use a combination of three techniques to allow the existing photometric losses to work for both day and nighttime images.

Depth Estimation Motion Estimation

Sample, Crop, Track: Self-Supervised Mobile 3D Object Detection for Urban Driving LiDAR

no code implementations21 Sep 2022 Sangyun Shin, Stuart Golodetz, Madhu Vankadari, Kaichen Zhou, Andrew Markham, Niki Trigoni

Supervised approaches typically require the annotation of large training sets; there has thus been great interest in leveraging weakly, semi- or self-supervised methods to avoid this, with much success.

3D Object Detection Object +2

Tracking People in Highly Dynamic Industrial Environments

no code implementations1 Feb 2023 Savvas Papaioannou, Andrew Markham, Niki Trigoni

We have conducted extensive real-world experiments in a construction site showing significant accuracy improvement via cross-modality training and the use of social forces.

valid

Fusion of Radio and Camera Sensor Data for Accurate Indoor Positioning

no code implementations1 Feb 2023 Savvas Papaioannou, Hongkai Wen, Andrew Markham, Niki Trigoni

In this paper, we propose a novel positioning system, RAVEL (Radio And Vision Enhanced Localization), which fuses anonymous visual detections captured by widely available camera infrastructure, with radio readings (e. g. WiFi radio data).

Decoupling Skill Learning from Robotic Control for Generalizable Object Manipulation

no code implementations7 Mar 2023 Kai Lu, Bo Yang, Bing Wang, Andrew Markham

Our experiments on manipulating complex articulated objects show that the proposed approach is more generalizable to unseen objects with large intra-class variations, outperforming previous approaches.

Imitation Learning Reinforcement Learning (RL)

Fast model inference and training on-board of Satellites

2 code implementations17 Jul 2023 Vít Růžička, Gonzalo Mateo-García, Chris Bridges, Chris Brunskill, Cormac Purcell, Nicolas Longépé, Andrew Markham

In this work we demonstrate the reliable use of RaVAEn onboard a satellite, achieving an encoding time of 0. 110s for tiles of a 4. 8x4. 8 km$^2$ area.

Decision Making

Deep Learning for Visual Localization and Mapping: A Survey

no code implementations27 Aug 2023 Changhao Chen, Bing Wang, Chris Xiaoxuan Lu, Niki Trigoni, Andrew Markham

Deep learning based localization and mapping approaches have recently emerged as a new research direction and receive significant attentions from both industry and academia.

Simultaneous Localization and Mapping Visual Localization +1

DynPoint: Dynamic Neural Point For View Synthesis

1 code implementation NeurIPS 2023 Kaichen Zhou, Jia-Xing Zhong, Sangyun Shin, Kai Lu, Yiyuan Yang, Andrew Markham, Niki Trigoni

The introduction of neural radiance fields has greatly improved the effectiveness of view synthesis for monocular videos.

Spherical Mask: Coarse-to-Fine 3D Point Cloud Instance Segmentation with Spherical Representation

1 code implementation18 Dec 2023 Sangyun Shin, Kaichen Zhou, Madhu Vankadari, Andrew Markham, Niki Trigoni

Coarse-to-fine 3D instance segmentation methods show weak performances compared to recent Grouping-based, Kernel-based and Transformer-based methods.

3D Instance Segmentation Semantic Segmentation

MGDepth: Motion-Guided Cost Volume For Self-Supervised Monocular Depth In Dynamic Scenarios

no code implementations23 Dec 2023 Kaichen Zhou, Jia-Xing Zhong, Jia-Wang Bian, Qian Xie, Jian-Qing Zheng, Niki Trigoni, Andrew Markham

Despite advancements in self-supervised monocular depth estimation, challenges persist in dynamic scenarios due to the dependence on assumptions about a static world.

Computational Efficiency Monocular Depth Estimation +1

Learning Continuous 3D Words for Text-to-Image Generation

no code implementations13 Feb 2024 Ta-Ying Cheng, Matheus Gadelha, Thibault Groueix, Matthew Fisher, Radomir Mech, Andrew Markham, Niki Trigoni

We do this by engineering special sets of input tokens that can be transformed in a continuous manner -- we call them Continuous 3D Words.

Text-to-Image Generation

Gen4Gen: Generative Data Pipeline for Generative Multi-Concept Composition

1 code implementation23 Feb 2024 Chun-Hsiao Yeh, Ta-Ying Cheng, He-Yen Hsieh, Chuan-En Lin, Yi Ma, Andrew Markham, Niki Trigoni, H. T. Kung, Yubei Chen

First, current personalization techniques fail to reliably extend to multiple concepts -- we hypothesize this to be due to the mismatch between complex scenes and simple text descriptions in the pre-training dataset (e. g., LAION).

Image Generation

See, Imagine, Plan: Discovering and Hallucinating Tasks from a Single Image

no code implementations18 Mar 2024 Chenyang Ma, Kai Lu, Ta-Ying Cheng, Niki Trigoni, Andrew Markham

Humans can not only recognize and understand the world in its current state but also envision future scenarios that extend beyond immediate perception.

Hallucination Motion Planning

WSCLoc: Weakly-Supervised Sparse-View Camera Relocalization

no code implementations22 Mar 2024 Jialu Wang, Kaichen Zhou, Andrew Markham, Niki Trigoni

Despite the advancements in deep learning for camera relocalization tasks, obtaining ground truth pose labels required for the training process remains a costly endeavor.

Camera Relocalization Image Reconstruction +1

ZeST: Zero-Shot Material Transfer from a Single Image

no code implementations9 Apr 2024 Ta-Ying Cheng, Prafull Sharma, Andrew Markham, Niki Trigoni, Varun Jampani

We propose ZeST, a method for zero-shot material transfer to an object in the input image given a material exemplar image.

Object

Cannot find the paper you are looking for? You can Submit a new open access paper.