no code implementations • 4 Jul 2023 • Siwon Kim, Sangdoo Yun, Hwaran Lee, Martin Gubri, Sungroh Yoon, Seong Joon Oh
The rapid advancement and widespread use of large language models (LLMs) have raised significant concerns regarding the potential leakage of personally identifiable information (PII).
1 code implementation • 20 Jun 2023 • Byeongho Heo, Taekyung Kim, Sangdoo Yun, Dongyoon Han
While the main model employs conventional training recipes, the sub-model leverages the benefit of additional regularization.
1 code implementation • 27 May 2023 • Jaewoo Ahn, Yeda Song, Sangdoo Yun, Gunhee Kim
In order to build self-consistent personalized dialogue agents, previous research has mostly focused on textual persona that delivers personal facts or personalities.
no code implementations • 24 May 2023 • Taehyun Lee, Seokhee Hong, Jaewoo Ahn, Ilgee Hong, Hwaran Lee, Sangdoo Yun, Jamin Shin, Gunhee Kim
Large language models for code have recently shown remarkable performance in generating executable code.
1 code implementation • 24 May 2023 • Geewook Kim, Hodong Lee, Daehee Kim, Haeji Jung, SangHee Park, Yoonsik Kim, Sangdoo Yun, Taeho Kil, Bado Lee, Seunghyun Park
Advances in Large Language Models (LLMs) have inspired a surge of research exploring their expansion into the visual domain.
1 code implementation • 1 May 2023 • Namuk Park, Wonjae Kim, Byeongho Heo, Taekyung Kim, Sangdoo Yun
We present a comparative study on how and why contrastive learning (CL) and masked image modeling (MIM) differ in their representations and in their performance of downstream tasks.
1 code implementation • 21 Apr 2023 • Seulki Park, Daeho Um, Hajung Yoon, Sanghyuk Chun, Sangdoo Yun, Jin Young Choi
In this paper, we propose a robustness benchmark for image-text matching models to assess their vulnerabilities.
1 code implementation • 10 Apr 2023 • Gyeongsik Moon, Hongsuk Choi, Sanghyuk Chun, Jiyoung Lee, Sangdoo Yun
Recovering 3D human mesh in the wild is greatly challenging as in-the-wild (ITW) datasets provide only 2D pose ground truths (GTs).
Ranked #5 on
3D Multi-Person Pose Estimation
on MuPoTS-3D
2 code implementations • 30 Mar 2023 • Dongyoon Han, Junsuk Choe, Seonghyeok Chun, John Joon Young Chung, Minsuk Chang, Sangdoo Yun, Jean Y. Song, Seong Joon Oh
We refer to the new paradigm of training models with annotation byproducts as learning using annotation byproducts (LUAB).
1 code implementation • 21 Mar 2023 • Geonmo Gu, Sanghyuk Chun, Wonjae Kim, HeeJae Jun, Yoohoon Kang, Sangdoo Yun
This paper proposes a novel diffusion-based model, CompoDiff, for solving Composed Image Retrieval (CIR) with latent diffusion and presents a newly created dataset of 18 million reference images, conditions, and corresponding target image triplets to train the model.
1 code implementation • 20 Mar 2023 • Song Park, Sanghyuk Chun, Byeongho Heo, Wonjae Kim, Sangdoo Yun
We need billion-scale images to achieve more generalizable and ground-breaking vision models, as well as massive dataset storage to ship the images (e. g., the LAION-4B dataset needs 240TB storage space).
no code implementations • 29 Jan 2023 • Jang-Hyun Kim, Sangdoo Yun, Hyun Oh Song
In this paper, we propose a unified approach for identifying the problematic data by utilizing a largely ignored source of information: a relational structure of data in the feature-embedded space.
Out-of-Distribution Detection
Out of Distribution (OOD) Detection
no code implementations • 8 Dec 2022 • Byungsoo Ko, Han-Gyu Kim, Byeongho Heo, Sangdoo Yun, Sanghyuk Chun, Geonmo Gu, Wonjae Kim
As ViT groups the channels via a multi-head attention mechanism, grouping the channels by GGeM leads to lower head-wise dependence while amplifying important channels on the activation maps.
1 code implementation • 21 Aug 2022 • Chanwoo Park, Sangdoo Yun, Sanghyuk Chun
Our theoretical results show that regardless of the choice of the mixing strategy, MSDA behaves as a pixel-level regularization of the underlying training loss and a regularization of the first layer parameters.
no code implementations • 30 Jun 2022 • Taeoh Kim, Jinhyung Kim, Minho Shim, Sangdoo Yun, Myunggu Kang, Dongyoon Wee, Sangyoun Lee
The magnitude of augmentation operations on each frame is changed by an effective mechanism, Fourier Sampling that parameterizes diverse, smooth, and realistic temporal variations.
2 code implementations • 30 May 2022 • Jang-Hyun Kim, Jinuk Kim, Seong Joon Oh, Sangdoo Yun, Hwanjun Song, JoonHyun Jeong, Jung-Woo Ha, Hyun Oh Song
The great success of machine learning with massive amounts of data comes at a price of huge computation costs and storage for training and tuning.
1 code implementation • CVPR 2022 • Jungbeom Lee, Seong Joon Oh, Sangdoo Yun, Junsuk Choe, Eunji Kim, Sungroh Yoon
However, training on class labels only, classifiers suffer from the spurious correlation between foreground and background cues (e. g. train and rail), fundamentally bounding the performance of WSSS.
Weakly supervised Semantic Segmentation
Weakly-Supervised Semantic Segmentation
2 code implementations • 7 Feb 2022 • Saehyung Lee, Sanghyuk Chun, Sangwon Jung, Sangdoo Yun, Sungroh Yoon
However, in this study, we prove that the existing DC methods can perform worse than the random selection method when task-irrelevant information forms a significant part of the training dataset.
1 code implementation • CVPR 2022 • Jongin Lim, Sangdoo Yun, Seulki Park, Jin Young Choi
In this paper, we propose Hypergraph-Induced Semantic Tuplet (HIST) loss for deep metric learning that leverages the multilateral semantic relations of multiple samples to multiple classes via hypergraph modeling.
1 code implementation • CVPR 2022 • Seulki Park, Youngkyu Hong, Byeongho Heo, Sangdoo Yun, Jin Young Choi
The problem of class imbalanced data is that the generalization performance of the classifier deteriorates due to the lack of data from minority classes.
Ranked #18 on
Long-tail Learning
on ImageNet-LT
3 code implementations • 30 Nov 2021 • Geewook Kim, Teakgyu Hong, Moonbin Yim, Jeongyeon Nam, Jinyoung Park, Jinyeong Yim, Wonseok Hwang, Sangdoo Yun, Dongyoon Han, Seunghyun Park
Current Visual Document Understanding (VDU) methods outsource the task of reading text to off-the-shelf Optical Character Recognition (OCR) engines and focus on the understanding task with the OCR outputs.
Ranked #8 on
Document Image Classification
on RVL-CDIP
Document Image Classification
Optical Character Recognition
+1
no code implementations • 8 Oct 2021 • JoonHyun Jeong, Sungmin Cha, Youngjoon Yoo, Sangdoo Yun, Taesup Moon, Jongwon Choi
Image-mixing augmentations (e. g., Mixup and CutMix), which typically involve mixing two images, have become the de-facto training techniques for image classification.
no code implementations • ICLR 2022 • Luca Scimeca, Seong Joon Oh, Sanghyuk Chun, Michael Poli, Sangdoo Yun
This phenomenon, also known as shortcut learning, is emerging as a key limitation of the current generation of machine learning models.
1 code implementation • 3 Oct 2021 • Youngsoo Lee, Sangdoo Yun, Yeonghun Kim, Sunghee Choi
On the server-side, a deep learning model is divided and progressively transmitted to the user devices.
1 code implementation • ICCV 2021 • Jeesoo Kim, Junsuk Choe, Sangdoo Yun, Nojun Kwak
Weakly-supervised object localization (WSOL) enables finding an object using a dataset without any localization information.
no code implementations • 14 Jun 2021 • Seulki Park, Hwanjun Song, Daeho Um, Dae Ung Jo, Sangdoo Yun, Jin Young Choi
Deep neural network can easily overfit to even noisy labels due to its high capacity, which degrades the generalization performance of a model.
10 code implementations • ICCV 2021 • Byeongho Heo, Sangdoo Yun, Dongyoon Han, Sanghyuk Chun, Junsuk Choe, Seong Joon Oh
We empirically show that such a spatial dimension reduction is beneficial to a transformer architecture as well, and propose a novel Pooling-based Vision Transformer (PiT) upon the original ViT model.
Ranked #310 on
Image Classification
on ImageNet
2 code implementations • CVPR 2021 • Sangdoo Yun, Seong Joon Oh, Byeongho Heo, Dongyoon Han, Junsuk Choe, Sanghyuk Chun
However, they have not fixed the training set, presumably because of a formidable annotation cost.
Ranked #21 on
Image Classification
on OmniBenchmark
2 code implementations • 7 Dec 2020 • Sangdoo Yun, Seong Joon Oh, Byeongho Heo, Dongyoon Han, Jinhyung Kim
Recent data augmentation strategies have been reported to address the overfitting problems in static image classifiers.
10 code implementations • CVPR 2021 • Dongyoon Han, Sangdoo Yun, Byeongho Heo, Youngjoon Yoo
We then investigate the channel configuration of a model by searching network architectures concerning the channel configuration under the computational cost restriction.
Ranked #419 on
Image Classification
on ImageNet
4 code implementations • ICLR 2021 • Byeongho Heo, Sanghyuk Chun, Seong Joon Oh, Dongyoon Han, Sangdoo Yun, Gyuwan Kim, Youngjung Uh, Jung-Woo Ha
Because of the scale invariance, this modification only alters the effective step sizes without changing the effective update directions, thus enjoying the original convergence properties of GD optimizers.
no code implementations • 9 Mar 2020 • Sanghyuk Chun, Seong Joon Oh, Sangdoo Yun, Dongyoon Han, Junsuk Choe, Youngjoon Yoo
Despite apparent human-level performances of deep neural networks (DNN), they behave fundamentally differently from humans.
no code implementations • 15 Oct 2019 • YoungJoon Yoo, Sanghyuk Chun, Sangdoo Yun, Jung-Woo Ha, Jaejun Yoo
We first assume that the priors of future samples can be generated in an independently and identically distributed (i. i. d.)
3 code implementations • ICML 2020 • Hyojin Bahng, Sanghyuk Chun, Sangdoo Yun, Jaegul Choo, Seong Joon Oh
This tactic is feasible in many scenarios where it is much easier to define a set of biased representations than to define and quantify bias.
no code implementations • CVPR 2017 • YoungJoon Yoo, Sangdoo Yun, Hyung Jin Chang, Yiannis Demiris, Jin Young Choi
(iii) The proposed regression is embedded into a generative model, and the whole procedure is developed by the variational autoencoder framework.
2 code implementations • 15 Jun 2019 • YoungJoon Yoo, Dongyoon Han, Sangdoo Yun
In this paper, we propose a new multi-scale face detector having an extremely tiny number of parameters (EXTD), less than 0. 1 million, as well as achieving comparable performance to deep heavy detectors.
Ranked #22 on
Face Detection
on WIDER Face (Hard)
29 code implementations • ICCV 2019 • Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, Youngjoon Yoo
Regional dropout strategies have been proposed to enhance the performance of convolutional neural network classifiers.
Ranked #1 on
Out-of-Distribution Generalization
on ImageNet-W
11 code implementations • ICCV 2019 • Jeonghun Baek, Geewook Kim, Junyeop Lee, Sungrae Park, Dongyoon Han, Sangdoo Yun, Seong Joon Oh, Hwalsuk Lee
Many new proposals for scene text recognition (STR) models have been introduced in recent years.
Ranked #6 on
Scene Text Recognition
on ICDAR 2003
2 code implementations • ICCV 2019 • Byeongho Heo, Jeesoo Kim, Sangdoo Yun, Hyojin Park, Nojun Kwak, Jin Young Choi
We investigate the design aspects of feature distillation methods achieving network compression and propose a novel feature distillation method in which the distillation loss is designed to make a synergy among various aspects: teacher transform, student transform, distillation feature position and distance function.
Ranked #32 on
Knowledge Distillation
on ImageNet
18 code implementations • CVPR 2019 • Youngmin Baek, Bado Lee, Dongyoon Han, Sangdoo Yun, Hwalsuk Lee
Scene text detection methods based on neural networks have emerged recently and have shown promising results.
Ranked #1 on
Scene Text Detection
on ICDAR 2013
(Precision metric)
2 code implementations • 12 Dec 2018 • Hyojin Park, Youngjoon Yoo, Geonseok Seo, Dongyoon Han, Sangdoo Yun, Nojun Kwak
To resolve this problem, we propose a new block called Concentrated-Comprehensive Convolution (C3) which applies the asymmetric convolutions before the depth-wise separable dilated convolution to compensate for the information loss due to dilated convolution.
2 code implementations • 8 Nov 2018 • Byeongho Heo, Minsik Lee, Sangdoo Yun, Jin Young Choi
In this paper, we propose a knowledge transfer method via distillation of activation boundaries formed by hidden neurons.
1 code implementation • 15 May 2018 • Byeongho Heo, Minsik Lee, Sangdoo Yun, Jin Young Choi
In this paper, we provide a new perspective based on a decision boundary, which is one of the most important component of a classifier.
1 code implementation • CVPR 2018 • Jongwon Choi, Hyung Jin Chang, Tobias Fischer, Sangdoo Yun, Kyuewang Lee, Jiyeoup Jeong, Yiannis Demiris, Jin Young Choi
We propose a new context-aware correlation filter based tracking framework to achieve both high computational speed and state-of-the-art performance among real-time trackers.
Ranked #14 on
Visual Object Tracking
on VOT2017/18
no code implementations • 27 Nov 2017 • YoungJoon Yoo, SeongUk Park, Junyoung Choi, Sangdoo Yun, Nojun Kwak
In addition to this performance enhancement problem, we show that the proposed PGN can be adopted to solve the classical adversarial problem without utilizing the information on the target classifier.
1 code implementation • IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017 • Jongwon Choi, Hyung Jin Chang, Sangdoo Yun, Tobias Fischer, Yiannis Demiris, Jin Young Choi
We propose a new tracking framework with an attentional mechanism that chooses a subset of the associated correlation filters for increased robustness and computational efficiency.
1 code implementation • The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops 2017 • Junho Cho, Sangdoo Yun, Kyoung Mu Lee, Jin Young Choi
PaletteNet is then designed to change the color concept of a source image so that the palette of the output image is close to the target palette.
1 code implementation • CVPR 2017 • Sangdoo Yun, Jongwon Choi, Youngjoon Yoo, Kimin Yun, Jin Young Choi
In contrast to the existing trackers using deep networks, the proposed tracker is designed to achieve a light computation as well as satisfactory tracking accuracy in both location and scale.
1 code implementation • ECCV 2018 • Donghoon Lee, Sangdoo Yun, Sungjoon Choi, Hwiyeon Yoo, Ming-Hsuan Yang, Songhwai Oh
We introduce a new problem of generating an image based on a small number of key local patches without any geometric prior.
no code implementations • CVPR 2016 • YoungJoon Yoo, Kimin Yun, Sangdoo Yun, JongHee Hong, Hawook Jeong, Jin Young Choi
In this paper, we consider moving dynamics of co-occurring objects for path prediction in a scene that includes crowded moving objects.