MergedNET: A simple approach for one-shot learning in siamese networks based on similarity layers

Neurocomputing 2022  ยท  John Atanbori, Samuel Rose ยท

Classifiers trained on disjointed classes with few labelled data points are used in one-shot learning to identify visual concepts from other classes. Recently, Siamese networks and similarity layers have been used to solve the one-shot learning problem, achieving state-of-the-art performance on visual-character recognition datasets. Various techniques have been developed over the years to improve the performance of these networks on fine-grained image classification datasets. They focused primarily on improving the loss and activation functions, augmenting visual features, employing multiscale metric learning, and pre-training and fine-tuning the backbone network. We investigate similarity layers for one-shot learning tasks and propose two frameworks for combining these layers into a MergedNet network. On all four datasets used in our experiment, MergedNet outperformed the baselines based on classification accuracy, and it generalises to other datasets when trained on miniImageNet.

PDF Neurocomputing 2022 PDF
Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Few-Shot Image Classification Caltech-256 5-way (1-shot) MergedNet-Max Accuracy 65.77 # 3
Few-Shot Image Classification Caltech-256 5-way (5-shot) MergedNet-Concat Accuracy 81.34 # 1
Few-Shot Image Classification CUB 200 5-way 1-shot MergedNet-Max Accuracy 75.34 # 23
Few-Shot Image Classification CUB 200 5-way 5-shot MergedNet-Max Accuracy 83.42 # 25
Few-Shot Image Classification Mini-Imagenet 5-way (1-shot) MergedNet-Max Accuracy 68.05 # 36
Few-Shot Image Classification Mini-Imagenet 5-way (5-shot) MergedNet-Max Accuracy 80.40 # 45

Methods


No methods listed for this paper. Add relevant methods here