no code implementations • ICML 2020 • Seong-Jin Park, Seungju Han, Ji-won Baek, Insoo Kim, Juhwan Song, Hae Beom Lee, Jae-Joon Han, Sung Ju Hwang
Humans have the ability to robustly recognize objects with various factors of variations such as nonrigid transformation, background noise, and change in lighting conditions.
1 code implementation • 3 Oct 2023 • Jean-Pierre Falet, Hae Beom Lee, Nikolay Malkin, Chen Sun, Dragos Secrieru, Thomas Jiralerspong, Dinghuai Zhang, Guillaume Lajoie, Yoshua Bengio
We present a new algorithm for amortized inference in sparse probabilistic graphical models (PGMs), which we call $\Delta$-amortized inference ($\Delta$-AI).
1 code implementation • 21 Aug 2022 • Hae Beom Lee, Dong Bok Lee, Sung Ju Hwang
In this paper, we introduce a novel approach for systematically solving dataset condensation problem in an efficient manner by exploiting the regularity in a given dataset.
no code implementations • 5 Mar 2022 • Boyan Gao, Henry Gouk, Hae Beom Lee, Timothy M. Hospedales
The resulting framework, termed Meta Mirror Descent (MetaMD), learns to accelerate optimisation speed.
no code implementations • 12 Oct 2021 • Jeffrey Willette, Hae Beom Lee, Juho Lee, Sung Ju Hwang
Numerous recent works utilize bi-Lipschitz regularization of neural network layers to preserve relative distances between data instances in the feature spaces of each layer.
no code implementations • ICLR 2022 • Hae Beom Lee, Hayeon Lee, Jaewoong Shin, Eunho Yang, Timothy Hospedales, Sung Ju Hwang
Many gradient-based meta-learning methods assume a set of parameters that do not participate in inner-optimization, which can be considered as hyperparameters.
no code implementations • ICLR 2022 • Seanie Lee, Hae Beom Lee, Juho Lee, Sung Ju Hwang
Thanks to the gradients aligned between tasks by our method, the model becomes less vulnerable to negative transfer and catastrophic forgetting.
no code implementations • ICLR 2022 • Jeffrey Ryan Willette, Hae Beom Lee, Juho Lee, Sung Ju Hwang
Numerous recent works utilize bi-Lipschitz regularization of neural network layers to preserve relative distances between data instances in the feature spaces of each layer.
no code implementations • 14 Feb 2021 • Jaewoong Shin, Hae Beom Lee, Boqing Gong, Sung Ju Hwang
Meta-learning of shared initialization parameters has shown to be highly effective in solving few-shot learning tasks.
no code implementations • 1 Jan 2021 • Seong Min Kye, Hae Beom Lee, Hoirin Kim, Sung Ju Hwang
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples, or confidence-weighted average of all the query samples.
1 code implementation • NeurIPS 2020 • Jeongun Ryu, Jaewoong Shin, Hae Beom Lee, Sung Ju Hwang
As MetaPerturb is a set-function trained over diverse distributions across layers and tasks, it can generalize to heterogeneous tasks and architectures.
2 code implementations • ICLR 2020 • Hae Beom Lee, Taewook Nam, Eunho Yang, Sung Ju Hwang
Specifically, we meta-learn a noise generator which outputs a multiplicative noise distribution for latent features, to obtain low errors on the test instances in an input-dependent manner.
Ranked #1 on Meta-Learning on OMNIGLOT - 1-Shot, 20-way
1 code implementation • 6 Apr 2020 • Seong Min Kye, Youngmoon Jung, Hae Beom Lee, Sung Ju Hwang, Hoirin Kim
By combining these two learning schemes, our model outperforms existing state-of-the-art speaker verification models learned with a standard supervised learning framework on short utterance (1-2 seconds) on the VoxCeleb datasets.
1 code implementation • 27 Feb 2020 • Seong Min Kye, Hae Beom Lee, Hoirin Kim, Sung Ju Hwang
To tackle this issue, we propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries such that they improve the model's transductive inference performance on unseen tasks.
no code implementations • 5 Aug 2019 • Hayeon Lee, Donghyun Na, Hae Beom Lee, Sung Ju Hwang
To tackle this issue, we propose a simple yet effective meta-learning framework for metricbased approaches, which we refer to as learning to generalize (L2G), that explicitly constrains the learning on a sampled classification task to reduce the classification error on a randomly sampled unseen classification task with a bilevel optimization scheme.
1 code implementation • ICLR 2020 • Hae Beom Lee, Hayeon Lee, Donghyun Na, Saehoon Kim, Minseop Park, Eunho Yang, Sung Ju Hwang
While tasks could come with varying the number of instances and classes in realistic settings, the existing meta-learning approaches for few-shot classification assume that the number of instances per task and class is fixed.
1 code implementation • 30 May 2019 • Hae Beom Lee, Taewook Nam, Eunho Yang, Sung Ju Hwang
Specifically, we meta-learn a noise generator which outputs a multiplicative noise distribution for latent features, to obtain low errors on the test instances in an input-dependent manner.
no code implementations • 27 Sep 2018 • Juho Lee, Saehoon Kim, Jaehong Yoon, Hae Beom Lee, Eunho Yang, Sung Ju Hwang
With such input-independent dropout, each neuron is evolved to be generic across inputs, which makes it difficult to sparsify networks without accuracy loss.
1 code implementation • 28 May 2018 • Juho Lee, Saehoon Kim, Jaehong Yoon, Hae Beom Lee, Eunho Yang, Sung Ju Hwang
With such input-independent dropout, each neuron is evolved to be generic across inputs, which makes it difficult to sparsify networks without accuracy loss.
2 code implementations • NeurIPS 2018 • Jay Heo, Hae Beom Lee, Saehoon Kim, Juho Lee, Kwang Joon Kim, Eunho Yang, Sung Ju Hwang
Attention mechanism is effective in both focusing the deep learning models on relevant features and interpreting them.
no code implementations • ICLR 2018 • Hae Beom Lee, Juho Lee, Eunho Yang, Sung Ju Hwang
Moreover, the learning of dropout probabilities for non-target classes on each instance allows the classifier to focus more on classification against the most confusing classes.
4 code implementations • NeurIPS 2018 • Hae Beom Lee, Juho Lee, Saehoon Kim, Eunho Yang, Sung Ju Hwang
Moreover, the learning of dropout rates for non-target classes on each instance allows the classifier to focus more on classification against the most confusing classes.
1 code implementation • ICML 2018 • Hae Beom Lee, Eunho Yang, Sung Ju Hwang
We propose Deep Asymmetric Multitask Feature Learning (Deep-AMTFL) which can learn deep representations shared across multiple tasks while effectively preventing negative transfer that may happen in the feature sharing process.