PointNetVLAD: Deep Point Cloud Based Retrieval for Large-Scale Place Recognition

CVPR 2018  ·  Mikaela Angelina Uy, Gim Hee Lee ·

Unlike its image based counterpart, point cloud based retrieval for place recognition has remained as an unexplored and unsolved problem. This is largely due to the difficulty in extracting local feature descriptors from a point cloud that can subsequently be encoded into a global descriptor for the retrieval task. In this paper, we propose the PointNetVLAD where we leverage on the recent success of deep networks to solve point cloud based retrieval for place recognition. Specifically, our PointNetVLAD is a combination/modification of the existing PointNet and NetVLAD, which allows end-to-end training and inference to extract the global descriptor from a given 3D point cloud. Furthermore, we propose the "lazy triplet and quadruplet" loss functions that can achieve more discriminative and generalizable global descriptors to tackle the retrieval task. We create benchmark datasets for point cloud based retrieval for place recognition, and the experimental results on these datasets show the feasibility of our PointNetVLAD. Our code and the link for the benchmark dataset downloads are available in our project website. http://github.com/mikacuy/pointnetvlad/

PDF Abstract CVPR 2018 PDF CVPR 2018 Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
3D Place Recognition CS-Campus3D PointNetVLAD AR@1% 41.46 # 7
AR@1 35.57 # 7
AR@1% cross-source 43.53 # 6
AR@1 cross-source 19.07 # 6
Visual Localization Oxford Radar RobotCar (Full-6) PointNetVLAD Mean Translation Error 28.48 # 10
3D Place Recognition Oxford RobotCar Dataset pointnetvlad AR@1% 80.3 # 10
Point Cloud Retrieval Oxford RobotCar (LiDAR 4096 points) PointNetVLAD (baseline) recall@top1% 80.31 # 23
Point Cloud Retrieval Oxford RobotCar (LiDAR 4096 points) PointNetVLAD (refined) recall@top1% 80.09 # 24

Methods