Dissecting the impact of different loss functions with gradient surgery

27 Jan 2022  ·  Hong Xuan, Robert Pless ·

Pair-wise loss is an approach to metric learning that learns a semantic embedding by optimizing a loss function that encourages images from the same semantic class to be mapped closer than images from different classes. The literature reports a large and growing set of variations of the pair-wise loss strategies. Here we decompose the gradient of these loss functions into components that relate to how they push the relative feature positions of the anchor-positive and anchor-negative pairs. This decomposition allows the unification of a large collection of current pair-wise loss functions. Additionally, explicitly constructing pair-wise gradient updates to separate out these effects gives insights into which have the biggest impact, and leads to a simple algorithm that beats the state of the art for image retrieval on the CAR, CUB and Stanford Online products datasets.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Metric Learning CARS196 Gradient Surgery R@1 86.5 # 23
Metric Learning CUB-200-2011 Gradient Surgery R@1 63.8 # 23
Metric Learning In-Shop Gradient Surgery R@1 92.21 # 7
Metric Learning Stanford Online Products Gradient Surgery R@1 82.3 # 12

Methods


No methods listed for this paper. Add relevant methods here