AmsterTime: A Visual Place Recognition Benchmark Dataset for Severe Domain Shift

30 Mar 2022  ·  Burak Yildiz, Seyran Khademi, Ronald Maria Siebes, Jan van Gemert ·

We introduce AmsterTime: a challenging dataset to benchmark visual place recognition (VPR) in presence of a severe domain shift. AmsterTime offers a collection of 2,500 well-curated images matching the same scene from a street view matched to historical archival image data from Amsterdam city. The image pairs capture the same place with different cameras, viewpoints, and appearances. Unlike existing benchmark datasets, AmsterTime is directly crowdsourced in a GIS navigation platform (Mapillary). We evaluate various baselines, including non-learning, supervised and self-supervised methods, pre-trained on different relevant datasets, for both verification and retrieval tasks. Our result credits the best accuracy to the ResNet-101 model pre-trained on the Landmarks dataset for both verification and retrieval tasks by 84% and 24%, respectively. Additionally, a subset of Amsterdam landmarks is collected for feature evaluation in a classification task. Classification labels are further used to extract the visual explanations using Grad-CAM for inspection of the learned similar visuals in a deep metric learning models.

PDF Abstract

Datasets


Introduced in the Paper:

AmsterTime

Used in the Paper:

ImageNet Google Landmarks Dataset v2

Results from the Paper


 Ranked #1 on Image Classification on AmsterTime (using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
Image Classification AmsterTime AP-GeM (ResNet-101) Accuracy 0.84 # 1
Image Retrieval AmsterTime AP-GeM (ResNet-101) mAP 35 # 5

Methods


No methods listed for this paper. Add relevant methods here