MES-Loss: Mutually equidistant separation metric learning loss function

Deep metric learning has attracted much attention in recent years due to its extensive applications, such as clustering and image retrieval. Thanks to the success of deep learning (DL), many deep metric learning (DML) methods have been proposed. Neural networks (NNs) utilize DML loss functions to learn a mapping function that maps samples into a highly discriminative low-dimensional feature space, facilitating measuring similarities between pairs of samples in such a manifold. Most existing methods usually try to boost the discriminatory power of NN by enhancing intra-class compactness in the high-level feature space. However, they do not explicitly imply constraints to improve inter-class separation. We propose in this paper a new composite DML loss function that, in addition to the intra-class compactness, explicitly implies regulations to enforce the best inter-class separation by mutually equidistantly distributing the centers of the classes. The proposed DML loss function achieved state-of-the-art results for clustering and image-retrieval tasks on two real-world data sets.

PDF

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Image Clustering CARS196 MES-Loss NMI 74.82 # 1
Image Retrieval CARS196 MES-Loss R@1 87.89 # 4
R@8 97.62 # 1
Image Clustering CUB-200-2011 MES-Loss NMI 73.35 # 1
Image Retrieval CUB-200-2011 MES-Loss R@1 67.03 # 5
R@8 91.50 # 1
R@4 84.75 # 1
R@2 75.67 # 1

Methods


No methods listed for this paper. Add relevant methods here