no code implementations • 1 Jan 2021 • Mengye Ren, Eleni Triantafillou, Kuan-Chieh Wang, James Lucas, Jake Snell, Xaq Pitkow, Andreas S. Tolias, Richard Zemel
In this work, we consider a realistic setting where the relationship between examples can change from episode to episode depending on the task context, which is not given to the learner.
no code implementations • 10 Dec 2020 • Mengye Ren, Eleni Triantafillou, Kuan-Chieh Wang, James Lucas, Jake Snell, Xaq Pitkow, Andreas S. Tolias, Richard Zemel
Despite impressive progress in deep learning, generalizing far beyond the training distribution is an important open challenge.
2 code implementations • ICLR 2021 • Jake Snell, Richard Zemel
Few-shot classification (FSC), the task of adapting a classifier to unseen classes given a small labeled dataset, is an important step on the path toward human-like machine learning.
no code implementations • ICLR 2019 • Marc T. Law, Jake Snell, Amir-Massoud Farahmand, Raquel Urtasun, Richard S. Zemel
Most deep learning models rely on expressive high-dimensional representations to achieve good performance on tasks such as classification.
1 code implementation • NeurIPS 2018 • Jack Klys, Jake Snell, Richard Zemel
We consider the problem of unsupervised learning of features correlated to specific labels in a dataset.
no code implementations • 27 Sep 2018 • Marc T Law, Jake Snell, Richard S Zemel
This formulation produces node representations close to the centroid of their descendants.
9 code implementations • ICLR 2018 • Mengye Ren, Eleni Triantafillou, Sachin Ravi, Jake Snell, Kevin Swersky, Joshua B. Tenenbaum, Hugo Larochelle, Richard S. Zemel
To address this paradigm, we propose novel extensions of Prototypical Networks (Snell et al., 2017) that are augmented with the ability to use unlabeled examples when producing prototypes.
42 code implementations • NeurIPS 2017 • Jake Snell, Kevin Swersky, Richard S. Zemel
We propose prototypical networks for the problem of few-shot classification, where a classifier must generalize to new classes not seen in the training set, given only a small number of examples of each new class.
1 code implementation • 19 Nov 2015 • Jake Snell, Karl Ridgeway, Renjie Liao, Brett D. Roads, Michael C. Mozer, Richard S. Zemel
We propose instead to use a loss function that is better calibrated to human perceptual judgments of image quality: the multiscale structural-similarity score (MS-SSIM).