Shift-tolerant Perceptual Similarity Metric

27 Jul 2022  ·  Abhijay Ghildyal, Feng Liu ·

Existing perceptual similarity metrics assume an image and its reference are well aligned. As a result, these metrics are often sensitive to a small alignment error that is imperceptible to the human eyes. This paper studies the effect of small misalignment, specifically a small shift between the input and reference image, on existing metrics, and accordingly develops a shift-tolerant similarity metric. This paper builds upon LPIPS, a widely used learned perceptual similarity metric, and explores architectural design considerations to make it robust against imperceptible misalignment. Specifically, we study a wide spectrum of neural network elements, such as anti-aliasing filtering, pooling, striding, padding, and skip connection, and discuss their roles in making a robust metric. Based on our studies, we develop a new deep neural network-based perceptual similarity metric. Our experiments show that our metric is tolerant to imperceptible shifts while being consistent with the human similarity judgment.

PDF Abstract
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Video Quality Assessment MSU SR-QA Dataset ST-LPIPS (VGG) SROCC 0.57336 # 24
PLCC 0.56431 # 23
KLCC 0.45898 # 25
Type FR # 1
Video Quality Assessment MSU SR-QA Dataset ST-LPIPS (Alex) SROCC 0.53473 # 30
PLCC 0.54740 # 28
KLCC 0.42897 # 29
Type FR # 1

Methods


No methods listed for this paper. Add relevant methods here