( Image credit: Prototypical Networks for Few shot Learning in PyTorch )
|TREND||DATASET||BEST METHOD||PAPER TITLE||PAPER||CODE||COMPARE|
The zero-shot paradigm exploits vector-based word representations extracted from text corpora with unsupervised methods to learn general mapping functions from other feature spaces onto word space, where the words associated to the nearest neighbours of the mapped vectors are used as their linguistic labels.
Given a learned knowledge graph (KG), our approach takes as input semantic embeddings for each node (representing visual category).
Given semantic descriptions of object classes, zero-shot learning aims to accurately recognize objects of the unseen classes, from which no examples are available at the training stage, by associating them to the seen classes, from which labeled examples are provided.
In other cases the semantic embedding space is established by an independent natural language processing task, and then the image transformation into that space is learned in a second stage.
We propose prototypical networks for the problem of few-shot classification, where a classifier must generalize to new classes not seen in the training set, given only a small number of examples of each new class.
Once trained, a RN is able to classify images of new classes by computing relation scores between query images and the few examples of each new class without further updating the network.
#7 best model for Few-Shot Learning on Mini-ImageNet - 1-Shot Learning
State-of-the-art methods for zero-shot visual recognition formulate learning as a joint embedding problem of images and side information.
In addition, we show that a simple margin based loss is sufficient to outperform all other loss functions.
#3 best model for Metric Learning on CUB-200-2011 (using extra training data)
Traditionally, for this problem supervision is expressed in the form of sets of points that follow an ordinal relationship -- an anchor point $x$ is similar to a set of positive points $Y$, and dissimilar to a set of negative points $Z$, and a loss defined over these distances is minimized.
Graph convolutional neural networks have recently shown great potential for the task of zero-shot learning.