no code implementations • ECCV 2020 • Yunjae Jung, Donghyeon Cho, Sanghyun Woo, In So Kweon
In order to summarize a content video properly, it is important to grasp the sequential structure of video as well as the long-term dependency between frames.
no code implementations • 27 Jun 2024 • Kibaek Park, Francois Rameau, Jaesik Park, In So Kweon
The large abundance of perspective camera datasets facilitated the emergence of novel learning-based strategies for various tasks, such as camera localization, single image depth estimation, or view synthesis.
1 code implementation • 13 Jun 2024 • Youngtaek Oh, Pyunghwan Ahn, Jinhyung Kim, Gwangmo Song, Soonyoung Lee, In So Kweon, Junmo Kim
Vision and language models (VLMs) such as CLIP have showcased remarkable zero-shot recognition abilities yet face challenges in visio-linguistic compositionality, particularly in linguistic comprehension and fine-grained image-text alignment.
no code implementations • 4 Jun 2024 • Inkyu Shin, Qihang Yu, Xiaohui Shen, In So Kweon, Kuk-Jin Yoon, Liang-Chieh Chen
In the second stage, we leverage the reconstruction ability developed in the first stage to impose the temporal constraints on the video diffusion model.
no code implementations • 29 Mar 2024 • Byeongin Joung, Byeong-Uk Lee, Jaesung Choe, Ukcheol Shin, Minjun Kang, Taeyeop Lee, In So Kweon, Kuk-Jin Yoon
This paper proposes an algorithm for synthesizing novel views under few-shot setup.
no code implementations • CVPR 2024 • Sanghyun Woo, KwanYong Park, Inkyu Shin, Myungchul Kim, In So Kweon
Multi-target multi-camera tracking is a crucial task that involves identifying and tracking individuals over time using video streams from multiple cameras.
no code implementations • 28 Mar 2024 • Chenshuang Zhang, Chaoning Zhang, Kang Zhang, Axi Niu, Junmo Kim, In So Kweon
There is a growing concern about applying batch normalization (BN) in adversarial training (AT), especially when the model is trained on both adversarial samples and clean samples (termed Hybrid-AT).
1 code implementation • CVPR 2024 • Chenshuang Zhang, Fei Pan, Junmo Kim, In So Kweon, Chengzhi Mao
In this work, we introduce generative model as a data source for synthesizing hard images that benchmark deep models' robustness.
no code implementations • 30 Nov 2023 • Axi Niu, Kang Zhang, Joshua Tian Jin Tee, Trung X. Pham, Jinqiu Sun, Chang D. Yoo, In So Kweon, Yanning Zhang
It is well known the adversarial optimization of GAN-based image super-resolution (SR) methods makes the preceding SR model generate unpleasant and undesirable artifacts, leading to large distortion.
no code implementations • 8 Nov 2023 • Dawit Mureja Argaw, Junsik Kim, In So Kweon
Existing video compression (VC) methods primarily aim to reduce the spatial and temporal redundancies between consecutive frames in a video while preserving its quality.
1 code implementation • 21 Sep 2023 • Fei Pan, Xu Yin, Seokju Lee, Axi Niu, SungEui Yoon, In So Kweon
Then, we propose a semantic mining module that takes the object masks to refine the pseudo labels in the target domain.
no code implementations • 5 Sep 2023 • TaeHoon Kim, Pyunghwan Ahn, Sangyun Kim, Sihaeng Lee, Mark Marsden, Alessandra Sala, Seung Hwan Kim, Bohyung Han, Kyoung Mu Lee, Honglak Lee, Kyounghoon Bae, Xiangyu Wu, Yi Gao, Hailiang Zhang, Yang Yang, Weili Guo, Jianfeng Lu, Youngtaek Oh, Jae Won Cho, Dong-Jin Kim, In So Kweon, Junmo Kim, Wooyoung Kang, Won Young Jhoo, Byungseok Roh, Jonghwan Mun, Solgil Oh, Kenan Emir Ak, Gwang-Gook Lee, Yan Xu, Mingwei Shen, Kyomin Hwang, Wonsik Shin, Kamin Lee, Wonhark Park, Dongkwan Lee, Nojun Kwak, Yujin Wang, Yimu Wang, Tiancheng Gu, Xingchang Lv, Mingmao Sun
In this report, we introduce NICE (New frontiers for zero-shot Image Captioning Evaluation) project and share the results and outcomes of 2023 challenge.
no code implementations • ICCV 2023 • Dawit Mureja Argaw, Joon-Young Lee, Markus Woodson, In So Kweon, Fabian Caba Heilbron
While great progress has been attained, there is still a need for a pretrained multimodal model that can perform well in the ever-growing set of movie understanding tasks the community has been establishing.
no code implementations • 3 Jul 2023 • Axi Niu, Pham Xuan Trung, Kang Zhang, Jinqiu Sun, Yu Zhu, In So Kweon, Yanning Zhang
To speed up inference and further enhance the performance, our research revisits diffusion models in image super-resolution and proposes a straightforward yet significant diffusion model-based super-resolution method called ACDMSR (accelerated conditional diffusion model for image super-resolution).
no code implementations • 26 May 2023 • Axi Niu, Kang Zhang, Trung X. Pham, Pei Wang, Jinqiu Sun, In So Kweon, Yanning Zhang
Currently, there are two popular approaches for addressing real-world image super-resolution problems: degradation-estimation-based and blind-based methods.
no code implementations • 1 May 2023 • Chenshuang Zhang, Chaoning Zhang, Taegoo Kang, Donghun Kim, Sung-Ho Bae, In So Kweon
Beyond the basic goal of mask removal, we further investigate and find that it is possible to generate any desired mask by the adversarial attack.
no code implementations • 10 Apr 2023 • Inkyu Shin, Dahun Kim, Qihang Yu, Jun Xie, Hong-Seok Kim, Bradley Green, In So Kweon, Kuk-Jin Yoon, Liang-Chieh Chen
The meta architecture of the proposed Video-kMaX consists of two components: within clip segmenter (for clip-level segmentation) and cross-clip associater (for association beyond clips).
no code implementations • 4 Apr 2023 • Chaoning Zhang, Chenshuang Zhang, Chenghao Li, Yu Qiao, Sheng Zheng, Sumit Kumar Dam, Mengchun Zhang, Jung Uk Kim, Seong Tae Kim, Jinwoo Choi, Gyeong-Moon Park, Sung-Ho Bae, Lik-Hang Lee, Pan Hui, In So Kweon, Choong Seon Hong
Overall, this work is the first to survey ChatGPT with a comprehensive review of its underlying technology, applications, and challenges.
1 code implementation • 30 Mar 2023 • Ukcheol Shin, Kyunghyun Lee, In So Kweon, Jean Oh
Also, the proposed self-distillation loss encourages the network to extract complementary and meaningful representations from a single modality or complementary masked modalities.
Ranked #2 on Thermal Image Segmentation on PST900
no code implementations • 30 Mar 2023 • Hyeonggon Ryu, Arda Senocak, In So Kweon, Joon Son Chung
The objective of this work is to explore the learning of visually grounded speech models (VGS) from multilingual perspective.
no code implementations • CVPR 2023 • Taeyeop Lee, Jonathan Tremblay, Valts Blukis, Bowen Wen, Byeong-Uk Lee, Inkyu Shin, Stan Birchfield, In So Kweon, Kuk-Jin Yoon
Unlike previous unsupervised domain adaptation methods for category-level object pose estimation, our approach processes the test data in a sequential, online manner, and it does not require access to the source domain at runtime.
no code implementations • 23 Mar 2023 • Chenshuang Zhang, Chaoning Zhang, Sheng Zheng, Mengchun Zhang, Maryam Qamar, Sung-Ho Bae, In So Kweon
This work conducts a survey on audio diffusion model, which is complementary to existing surveys that either lack the recent progress of diffusion-based speech synthesis or highlight an overall picture of applying diffusion model in multiple fields.
no code implementations • 21 Mar 2023 • Chaoning Zhang, Chenshuang Zhang, Sheng Zheng, Yu Qiao, Chenghao Li, Mengchun Zhang, Sumit Kumar Dam, Chu Myaet Thwal, Ye Lin Tun, Le Luang Huy, Donguk Kim, Sung-Ho Bae, Lik-Hang Lee, Yang Yang, Heng Tao Shen, In So Kweon, Choong Seon Hong
As ChatGPT goes viral, generative AI (AIGC, a. k. a AI-generated content) has made headlines everywhere because of its ability to analyze and create text, images, and beyond.
no code implementations • 21 Mar 2023 • Youngjoon Jang, Youngtaek Oh, Jae Won Cho, Myungchul Kim, Dong-Jin Kim, In So Kweon, Joon Son Chung
The goal of this work is to develop self-sufficient framework for Continuous Sign Language Recognition (CSLR) that addresses key issues of sign language recognition.
no code implementations • 14 Mar 2023 • Chenshuang Zhang, Chaoning Zhang, Mengchun Zhang, In So Kweon
This survey reviews text-to-image diffusion models in the context that diffusion models have emerged to be popular for a wide range of generative tasks.
1 code implementation • CVPR 2023 • Junha Song, Jungsoo Lee, In So Kweon, Sungha Choi
Second, our novel self-distilled regularization controls the output of the meta networks not to deviate significantly from the output of the frozen original networks, thereby preserving well-trained knowledge from the source domain.
no code implementations • 14 Feb 2023 • Axi Niu, Kang Zhang, Trung X. Pham, Jinqiu Sun, Yu Zhu, In So Kweon, Yanning Zhang
Diffusion probabilistic models (DPM) have been widely adopted in image-to-image translation to generate high-quality images.
no code implementations • 26 Jan 2023 • Dong-Jin Kim, Tae-Hyun Oh, Jinsoo Choi, In So Kweon
We present a novel data-efficient semi-supervised framework to improve the generalization of image captioning models.
13 code implementations • CVPR 2023 • Sanghyun Woo, Shoubhik Debnath, Ronghang Hu, Xinlei Chen, Zhuang Liu, In So Kweon, Saining Xie
This co-design of self-supervised learning techniques and architectural improvement results in a new model family called ConvNeXt V2, which significantly improves the performance of pure ConvNets on various recognition benchmarks, including ImageNet classification, COCO detection, and ADE20K segmentation.
Ranked #45 on Semantic Segmentation on ADE20K
no code implementations • CVPR 2023 • KwanYong Park, Sanghyun Woo, Seoung Wug Oh, In So Kweon, Joon-Young Lee
Mask-guided matting has shown great practicality compared to traditional trimap-based methods.
no code implementations • CVPR 2023 • Byeong-Uk Lee, Jianming Zhang, Yannick Hold-Geoffroy, In So Kweon
In this paper, we propose a single image scale estimation method based on a novel scale field representation.
no code implementations • ICCV 2023 • Jaesung Choe, Christopher Choy, Jaesik Park, In So Kweon, Anima Anandkumar
We propose an algorithm, 4DRegSDF, for the spacetime surface regularization to improve the fidelity of neural rendering and reconstruction in dynamic scenes.
1 code implementation • CVPR 2023 • Ukcheol Shin, Jinsun Park, In So Kweon
Secondly, we conduct an exhaustive validation process of monocular and stereo depth estimation algorithms designed on visible spectrum bands to benchmark their performance in the thermal image domain.
no code implementations • 20 Dec 2022 • Sanghyun Woo, KwanYong Park, Seoung Wug Oh, In So Kweon, Joon-Young Lee
First, no tracking supervisions are in LVIS, which leads to inconsistent learning of detection (with LVIS and TAO) and tracking (only with TAO).
no code implementations • 20 Dec 2022 • Sanghyun Woo, KwanYong Park, Seoung Wug Oh, In So Kweon, Joon-Young Lee
The tracking-by-detection paradigm today has become the dominant method for multi-object tracking and works by detecting objects in each frame and then performing data association across frames.
no code implementations • 16 Dec 2022 • Junha Song, KwanYong Park, Inkyu Shin, Sanghyun Woo, Chaoning Zhang, In So Kweon
In addition, to prevent overfitting of the TTA model, we devise novel regularization which modulates the adaptation rates using domain-similarity between the source and the current target domain.
no code implementations • 16 Dec 2022 • Sungsu Hur, Inkyu Shin, KwanYong Park, Sanghyun Woo, In So Kweon
To successfully train our framework, we collect the partial, confident target samples that are classified as known or unknown through on our proposed multi-criteria selection.
1 code implementation • ICCV 2023 • M. Jehanzeb Mirza, Inkyu Shin, Wei Lin, Andreas Schriebl, Kunyang Sun, Jaesung Choe, Horst Possegger, Mateusz Kozinski, In So Kweon, Kun-Jin Yoon, Horst Bischof
Our MATE is the first Test-Time-Training (TTT) method designed for 3D data, which makes deep networks trained for point cloud classification robust to distribution shifts occurring in test data.
1 code implementation • 1 Nov 2022 • Youngjoon Jang, Youngtaek Oh, Jae Won Cho, Dong-Jin Kim, Joon Son Chung, In So Kweon
Most existing Continuous Sign Language Recognition (CSLR) benchmarks have fixed backgrounds and are filmed in studios with a static monochromatic background.
no code implementations • 21 Oct 2022 • Valts Blukis, Taeyeop Lee, Jonathan Tremblay, Bowen Wen, In So Kweon, Kuk-Jin Yoon, Dieter Fox, Stan Birchfield
At test-time, we build the representation from a single RGB input image observing the scene from only one viewpoint.
1 code implementation • 13 Sep 2022 • Joohyung Lee, Jieun Oh, Inkyu Shin, You-sung Kim, Dae Kyung Sohn, Tae-sung Kim, In So Kweon
In this study, we present a volumetric convolutional neural network to accurately discriminate T2 from T3 stage rectal cancer with rectal MR volumes.
1 code implementation • CVPR 2022 • KwanYong Park, Sanghyun Woo, Seoung Wug Oh, In So Kweon, Joon-Young Lee
In this per-clip inference scheme, we update the memory with an interval and simultaneously process a set of consecutive frames (i. e. clip) between the memory updates.
1 code implementation • CVPR 2023 • Jae Won Cho, Dong-Jin Kim, Hyeonggon Ryu, In So Kweon
In this work, in order to better learn the bias a target VQA model suffers from, we propose a generative method to train the bias model directly from the target model, called GenB.
no code implementations • 30 Jul 2022 • Chaoning Zhang, Chenshuang Zhang, Junha Song, John Seon Keun Yi, Kang Zhang, In So Kweon
Masked autoencoders are scalable vision learners, as the title of MAE \cite{he2022masked}, which suggests that self-supervised learning (SSL) in vision might undertake a similar trajectory as in NLP.
2 code implementations • 22 Jul 2022 • Chaoning Zhang, Kang Zhang, Chenshuang Zhang, Axi Niu, Jiu Feng, Chang D. Yoo, In So Kweon
Adversarial training (AT) for robust representation learning and self-supervised learning (SSL) for unsupervised representation learning are two active research fields.
1 code implementation • 20 Jul 2022 • Dawit Mureja Argaw, Fabian Caba Heilbron, Joon-Young Lee, Markus Woodson, In So Kweon
Machine learning is transforming the video editing industry.
no code implementations • 19 Jul 2022 • Fei Pan, Sungsu Hur, Seokju Lee, Junsik Kim, In So Kweon
Open compound domain adaptation (OCDA) considers the target domain as the compound of multiple unknown homogeneous subdomains.
no code implementations • 7 Jul 2022 • Ukcheol Shin, Kyunghyun Lee, In So Kweon
In this paper, we propose a multi-objective camera ISP framework that utilizes Deep Reinforcement Learning (DRL) and camera ISP toolbox that consist of network-based and conventional ISP tools.
no code implementations • 1 Jun 2022 • Fei Pan, Francois Rameau, Junsik Kim, In So Kweon
In this work, we propose a new domain adaptation framework for semantic segmentation with annotated points via active selection.
no code implementations • CVPR 2022 • Dahun Kim, Jun Xie, Huiyu Wang, Siyuan Qiao, Qihang Yu, Hong-Seok Kim, Hartwig Adam, In So Kweon, Liang-Chieh Chen
We present TubeFormer-DeepLab, the first attempt to tackle multiple core video segmentation tasks in a unified manner.
no code implementations • CVPR 2022 • Inkyu Shin, Yi-Hsuan Tsai, Bingbing Zhuang, Samuel Schulter, Buyu Liu, Sparsh Garg, In So Kweon, Kuk-Jin Yoon
In this paper, we propose and explore a new multi-modal extension of test-time adaptation for 3D semantic segmentation.
2 code implementations • CVPR 2022 • Chaoning Zhang, Kang Zhang, Trung X. Pham, Axi Niu, Zhinan Qiao, Chang D. Yoo, In So Kweon
Contrastive learning (CL) is widely known to require many negative samples, 65536 in MoCo for instance, for which the performance of a dictionary-free framework is often inferior because the negative sample size (NSS) is limited by its mini-batch size (MBS).
no code implementations • 30 Mar 2022 • Chaoning Zhang, Philipp Benz, Adil Karjauv, Jae Won Cho, Kang Zhang, In So Kweon
It is widely reported that stronger I-FGSM transfers worse than simple FGSM, leading to a popular belief that transferability is at odds with the white-box attack strength.
no code implementations • 30 Mar 2022 • Chaoning Zhang, Kang Zhang, Chenshuang Zhang, Trung X. Pham, Chang D. Yoo, In So Kweon
This yields a unified perspective on how negative samples and SimSiam alleviate collapse.
no code implementations • CVPR 2022 • Dawit Mureja Argaw, In So Kweon
We argue that when there is a large gap between inputs, instead of estimating imprecise motion that will eventually lead to inaccurate interpolation, we can safely propagate from one side of the input up to a reliable time frame using the other input as a reference.
no code implementations • 12 Feb 2022 • Arda Senocak, Junsik Kim, Tae-Hyun Oh, Hyeonggon Ryu, DIngzeyu Li, In So Kweon
Human brain is continuously inundated with the multisensory information and their complex interactions coming from the outside world at any given moment.
no code implementations • 11 Feb 2022 • Axi Niu, Kang Zhang, Chaoning Zhang, Chenshuang Zhang, In So Kweon, Chang D. Yoo, Yanning Zhang
The former works only for a relatively small perturbation 8/255 with the l_\infty constraint, and GradAlign improves it by extending the perturbation size to 16/255 (with the l_\infty constraint) but at the cost of being 3 to 4 times slower.
no code implementations • 7 Feb 2022 • Arda Senocak, Hyeonggon Ryu, Junsik Kim, In So Kweon
Thus, these semantically correlated pairs, "hard positives", are mistakenly grouped as negatives.
1 code implementation • 12 Jan 2022 • Ukcheol Shin, Kyunghyun Lee, Byeong-Uk Lee, In So Kweon
Based on the analysis, we propose an effective thermal image mapping method that significantly increases image information, such as overall structure, contrast, and details, while preserving temporal consistency.
1 code implementation • Computer Vision and Image Understanding 2022 • Francois Rameau, Jinsun Park, Oleksandr Bailo, In So Kweon
In this paper, we present MC-Calib, a novel and robust toolbox dedicated to the calibration of complex synchronized multi-camera systems using an arbitrary number of fiducial marker-based patterns.
no code implementations • CVPR 2022 • Chaoning Zhang, Philipp Benz, Adil Karjauv, Jae Won Cho, Kang Zhang, In So Kweon
It is widely reported that stronger I-FGSM transfers worse than simple FGSM, leading to a popular belief that transferability is at odds with the white-box attack strength.
no code implementations • 25 Nov 2021 • Minjun Kang, Jaesung Choe, Hyowon Ha, Hae-Gon Jeon, Sunghoon Im, In So Kweon, Kuk-Jin Yoon
Many mobile manufacturers recently have adopted Dual-Pixel (DP) sensors in their flagship models for faster auto-focus and aesthetic image captures.
no code implementations • CVPR 2022 • Taeyeop Lee, Byeong-Uk Lee, Inkyu Shin, Jaesung Choe, Ukcheol Shin, In So Kweon, Kuk-Jin Yoon
Inspired by recent multi-modal UDA techniques, the proposed method exploits a teacher-student self-supervised learning scheme to train a pose estimation network without using target domain pose labels.
Ranked #5 on 6D Pose Estimation using RGBD on REAL275
no code implementations • ICLR 2022 • Jaesung Choe, Byeongin Joung, Francois Rameau, Jaesik Park, In So Kweon
In particular, we further improve the performance of transformer by a newly proposed module called amplified positional encoding.
3 code implementations • 22 Nov 2021 • Jaesung Choe, Chunghyun Park, Francois Rameau, Jaesik Park, In So Kweon
MLP-Mixer has newly appeared as a new challenger against the realm of CNNs and transformer.
Ranked #20 on Semantic Segmentation on S3DIS Area5
no code implementations • 10 Nov 2021 • Jinsoo Choi, Jaesik Park, In So Kweon
Videos are a popular media form, where online video streaming has recently gathered much popularity.
no code implementations • 21 Oct 2021 • Dong-Jin Kim, Jae Won Cho, Jinsoo Choi, Yunjae Jung, In So Kweon
In this work, we address Active Learning in the multi-modal setting of Visual Question Answering (VQA).
no code implementations • ICCV 2021 • Seokju Lee, Francois Rameau, Fei Pan, In So Kweon
Experiments on KITTI, Cityscapes, and Waymo Open Dataset demonstrate the relevance of our approach and show that our method outperforms state-of-the-art algorithms for the tasks of self-supervised monocular depth estimation, object motion segmentation, monocular scene flow estimation, and visual odometry.
no code implementations • NeurIPS 2020 • KwanYong Park, Sanghyun Woo, Inkyu Shin, In So Kweon
The scheme first clusters compound target data based on style, discovering multiple latent domains (discover).
1 code implementation • 6 Oct 2021 • Philipp Benz, Soomin Ham, Chaoning Zhang, Adil Karjauv, In So Kweon
Thus, it is critical for the community to know whether the newly proposed ViT and MLP-Mixer are also vulnerable to adversarial attacks.
no code implementations • ICLR 2022 • Chaoning Zhang, Kang Zhang, Chenshuang Zhang, Trung X. Pham, Chang D. Yoo, In So Kweon
Towards avoiding collapse in self-supervised learning (SSL), contrastive loss is widely used but often requires a large number of negative samples.
no code implementations • 29 Sep 2021 • Chaoning Zhang, Gyusang Cho, Philipp Benz, Kang Zhang, Chenshuang Zhang, Chan-Hyun Youn, In So Kweon
The transferability of adversarial examples (AE); known as adversarial transferability, has attracted significant attention because it can be exploited for TransferableBlack-box Attacks (TBA).
1 code implementation • 9 Sep 2021 • Dong-Jin Kim, Xiao Sun, Jinsoo Choi, Stephen Lin, In So Kweon
A common problem in the task of human-object interaction (HOI) detection is that numerous HOI classes have only a small number of labeled examples, resulting in training sets with a long-tailed distribution.
Ranked #42 on Human-Object Interaction Detection on HICO-DET
no code implementations • 1 Sep 2021 • Taeyeop Lee, Byeong-Uk Lee, Myungchul Kim, In So Kweon
Our framework has two branches: the Metric Scale Object Shape branch (MSOS) and the Normalized Object Coordinate Space branch (NOCS).
no code implementations • ICCV 2021 • Jaesung Choe, Sunghoon Im, Francois Rameau, Minjun Kang, In So Kweon
To reconstruct a 3D scene from a set of calibrated views, traditional multi-view stereo techniques rely on two distinct stages: local depth maps computation and global depth maps fusion.
4 code implementations • 15 Aug 2021 • Dahun Kim, Tsung-Yi Lin, Anelia Angelova, In So Kweon, Weicheng Kuo
In this paper, we identify that the problem is that the binary classifiers in existing proposal methods tend to overfit to the training categories.
Ranked #2 on Open World Object Detection on COCO VOC to non-VOC
no code implementations • ICCV 2021 • Inkyu Shin, Dong-Jin Kim, Jae Won Cho, Sanghyun Woo, KwanYong Park, In So Kweon
In order to find the uncertain points, we generate an inconsistency mask using the proposed adaptive pixel selector and we label these segment-based regions to achieve near supervised performance with only a small fraction (about 2. 2%) ground truth points, which we call "Segment based Pixel-Labeling (SPL)".
1 code implementation • 12 Aug 2021 • Antyanta Bangunharcana, Jae Won Cho, Seokju Lee, In So Kweon, Kyung-Soo Kim, Soohyun Kim
Volumetric deep learning approach towards stereo matching aggregates a cost volume computed from input left and right images using 3D convolutions.
no code implementations • 23 Jul 2021 • Inkyu Shin, KwanYong Park, Sanghyun Woo, In So Kweon
In this work, we present a new video extension of this task, namely Unsupervised Domain Adaptation for Video Semantic Segmentation.
1 code implementation • 23 Jul 2021 • Jae Won Cho, Dong-Jin Kim, Yunjae Jung, In So Kweon
Recent state-of-the-art active learning methods have mostly leveraged Generative Adversarial Networks (GAN) for sample acquisition; however, GAN is usually known to suffer from instability and sensitivity to hyper-parameters.
no code implementations • CVPR 2021 • Sanghyun Woo, Dahun Kim, Joon-Young Lee, In So Kweon
Temporal correspondence - linking pixels or objects across frames - is a fundamental supervisory signal for the video models.
Ranked #6 on Video Panoptic Segmentation on Cityscapes-VPS (using extra training data)
1 code implementation • CVPR 2022 • Youngtaek Oh, Dong-Jin Kim, In So Kweon
The capability of the traditional semi-supervised learning (SSL) methods is far from real-world application due to severely biased pseudo-labels caused by (1) class imbalance and (2) class distribution mismatch between labeled and unlabeled data.
no code implementations • 19 Apr 2021 • Dawit Mureja Argaw, Junsik Kim, Francois Rameau, Chaoning Zhang, In So Kweon
We formulate video restoration from a single blurred image as an inverse problem by setting clean image sequence and their respective motion as latent factors, and the blurred image as an observation.
no code implementations • CVPR 2021 • Byeong-Uk Lee, Kyunghyun Lee, In So Kweon
The basic framework of depth completion is to predict a pixel-wise dense depth map using very sparse input data.
no code implementations • 13 Apr 2021 • Jae Won Cho, Dong-Jin Kim, Jinsoo Choi, Yunjae Jung, In So Kweon
In this work, we address the issues of missing modalities that have arisen from the Visual Question Answer-Difference prediction task and find a novel method to solve the task at hand.
no code implementations • 7 Apr 2021 • Philipp Benz, Chaoning Zhang, Adil Karjauv, In So Kweon
The SOTA universal adversarial training (UAT) method optimizes a single perturbation for all training samples in the mini-batch.
no code implementations • 24 Mar 2021 • Jaesung Choe, Kyungdon Joo, Tooba Imtiaz, In So Kweon
The key idea of our network is to exploit sparse and accurate point clouds as a cue for guiding correspondences of stereo images in a unified 3D volume space.
no code implementations • 23 Mar 2021 • Jaesung Choe, Kyungdon Joo, Francois Rameau, In So Kweon
This paper presents a stereo object matching method that exploits both 2D contextual information from images as well as 3D object-level information.
no code implementations • 4 Mar 2021 • Dawit Mureja Argaw, Junsik Kim, Francois Rameau, Jae Won Cho, In So Kweon
A flow estimator network is then used to estimate optical flow from the decoded features in a coarse-to-fine manner.
no code implementations • 4 Mar 2021 • Dawit Mureja Argaw, Junsik Kim, Francois Rameau, In So Kweon
Abrupt motion of camera or objects in a scene result in a blurry video, and therefore recovering high quality video requires two types of enhancements: visual enhancement and temporal upsampling.
1 code implementation • 2 Mar 2021 • Chaoning Zhang, Philipp Benz, Chenguo Lin, Adil Karjauv, Jing Wu, In So Kweon
The intriguing phenomenon of adversarial examples has attracted significant attention in machine learning and what might be more surprising to the community is the existence of universal adversarial perturbations (UAPs), i. e. a single perturbation to fool the target DNN for most images.
no code implementations • 2 Mar 2021 • Chaoning Zhang, Chenguo Lin, Philipp Benz, Kejiang Chen, Weiming Zhang, In So Kweon
Data hiding is the art of concealing messages with limited perceptual changes.
2 code implementations • 1 Mar 2021 • Ukcheol Shin, Kyunghyun Lee, Seokju Lee, In So Kweon
Based on the proposed module, the photometric consistency loss can provide complementary self-supervision to networks.
no code implementations • 12 Feb 2021 • Chaoning Zhang, Philipp Benz, Adil Karjauv, In So Kweon
We perform task-specific and joint analysis and reveal that (a) frequency is a key factor that influences their performance based on the proposed entropy metric for quantifying the frequency distribution; (b) their success can be attributed to a DNN being highly sensitive to high-frequency content.
1 code implementation • 4 Feb 2021 • Seokju Lee, Sunghoon Im, Stephen Lin, In So Kweon
We present an end-to-end joint training framework that explicitly models 6-DoF motion of multiple dynamic objects, ego-motion and depth in a monocular camera setup without supervision.
Ranked #2 on Monocular Depth Estimation on Cityscapes
no code implementations • 1 Jan 2021 • Junsoo Lee, Hojoon Lee, Inkyu Shin, Jaekyoung Bae, In So Kweon, Jaegul Choo
Learning visual representations using large-scale unlabelled images is a holy grail for most of computer vision tasks.
no code implementations • ICCV 2021 • Chaoning Zhang, Philipp Benz, Adil Karjauv, In So Kweon
For a more practical universal attack, our investigation of untargeted UAP focuses on alleviating the dependence on the original training samples, from removing the need for sample labels to limiting the sample size.
1 code implementation • 30 Dec 2020 • Chaoning Zhang, Adil Karjauv, Philipp Benz, In So Kweon
Recently, deep learning has shown large success in data hiding, while non-differentiability of JPEG makes it challenging to train a deep pipeline for improving robustness against lossy compression.
1 code implementation • NeurIPS 2020 • Kyunghyun Lee, Byeong-Uk Lee, Ukcheol Shin, In So Kweon
Deep reinforcement learning (DRL) algorithms and evolution strategies (ES) have been applied to various tasks, showing excellent performances.
no code implementations • 26 Nov 2020 • Myungchul Kim, Sanghyun Woo, Dahun Kim, In So Kweon
In this work, we propose Boundary Basis based Instance Segmentation(B2Inst) to learn a global boundary representation that can complement existing global-mask-based methods that are often lacking high-frequency details.
no code implementations • 26 Oct 2020 • Philipp Benz, Chaoning Zhang, Adil Karjauv, In So Kweon
Adversarial training is the most widely used technique for improving adversarial robustness to strong white-box attacks.
1 code implementation • 23 Oct 2020 • Chaoning Zhang, Philipp Benz, Dawit Mureja Argaw, Seokju Lee, Junsik Kim, Francois Rameau, Jean-Charles Bazin, In So Kweon
ResNet or DenseNet?
1 code implementation • 8 Oct 2020 • Dong-Jin Kim, Tae-Hyun Oh, Jinsoo Choi, In So Kweon
To this end, we propose the multi-task triple-stream network (MTTSNet) which consists of three recurrent units responsible for each POS which is trained by jointly predicting the correct captions and POS for each word.
1 code implementation • 7 Oct 2020 • Philipp Benz, Chaoning Zhang, Tooba Imtiaz, In So Kweon
This universal perturbation attacks one targeted source class to sink class, while having a limited adversarial effect on other non-targeted source classes, for avoiding raising suspicions.
no code implementations • 7 Oct 2020 • Chaoning Zhang, Philipp Benz, Tooba Imtiaz, In So Kweon
Since the proposed attack generates a universal adversarial perturbation that is discriminative to targeted and non-targeted classes, we term it class discriminative universal adversarial perturbation (CD-UAP).
1 code implementation • ICCV 2021 • Philipp Benz, Chaoning Zhang, In So Kweon
This work attempts to understand the impact of BN on DNNs from a non-robust feature perspective.
no code implementations • 7 Oct 2020 • Philipp Benz, Chaoning Zhang, Adil Karjauv, In So Kweon
We find that simply estimating and adapting the BN statistics on a few (32 for instance) representation samples, without retraining the model, improves the corruption robustness by a large margin on several benchmark datasets with a wide range of model architectures.
1 code implementation • ECCV 2020 • Jinsun Park, Kyungdon Joo, Zhe Hu, Chi-Kuei Liu, In So Kweon
In this paper, we propose a robust and efficient end-to-end non-local spatial propagation network for depth completion.
Ranked #1 on Depth Completion on NYU-Depth V2
1 code implementation • 17 Jul 2020 • Dong-Jin Kim, Xiao Sun, Jinsoo Choi, Stephen Lin, In So Kweon
A common problem in human-object interaction (HOI) detection task is that numerous HOI classes have only a small number of labeled examples, resulting in training sets with a long-tailed distribution.
1 code implementation • CVPR 2020 • Dahun Kim, Sanghyun Woo, Joon-Young Lee, In So Kweon
In this paper, we propose and explore a new video extension of this task, called video panoptic segmentation.
Ranked #7 on Video Panoptic Segmentation on Cityscapes-VPS (using extra training data)
1 code implementation • CVPR 2020 • Fei Pan, Inkyu Shin, Francois Rameau, Seokju Lee, In So Kweon
Finally, to decrease the intra-domain gap, we propose to employ a self-supervised adaptation technique from the easy to the hard split.
Ranked #2 on Domain Adaptation on Synscapes-to-Cityscapes
no code implementations • 3 Feb 2020 • Yunjae Jung, Dahun Kim, Sanghyun Woo, Kyung-Su Kim, Sungjin Kim, In So Kweon
In this paper, we propose to explicitly learn to imagine a storyline that bridges the visual gap.
Ranked #7 on Visual Storytelling on VIST
1 code implementation • 19 Dec 2019 • Seokju Lee, Sunghoon Im, Stephen Lin, In So Kweon
We present an end-to-end joint training framework that explicitly models 6-DoF motion of multiple dynamic objects, ego-motion and depth in a monocular camera setup without supervision.
1 code implementation • 20 Nov 2019 • Arda Senocak, Tae-Hyun Oh, Junsik Kim, Ming-Hsuan Yang, In So Kweon
Visual events are usually accompanied by sounds in our daily lives.
no code implementations • 16 Sep 2019 • Seokju Lee, Junsik Kim, Tae-Hyun Oh, Yongseop Jeong, Donggeun Yoo, Stephen Lin, In So Kweon
We postulate that success on this task requires the network to learn semantic and geometric knowledge in the ego-centric view.
no code implementations • 16 Sep 2019 • Seokju Lee, Sunghoon Im, Stephen Lin, In So Kweon
Based on rigid projective geometry, the estimated stereo depth is used to guide the camera motion estimation, and the depth and camera motion are used to guide the residual flow estimation.
no code implementations • IJCNLP 2019 • Dong-Jin Kim, Jinsoo Choi, Tae-Hyun Oh, In So Kweon
To this end, our proposed semi-supervised learning method assigns pseudo-labels to unpaired samples via Generative Adversarial Networks to learn the joint distribution of image and caption.
2 code implementations • 5 Sep 2019 • Jinsoo Choi, In So Kweon
We present a novel deep approach to video stabilization which can generate video frames without cropping and low distortion.
no code implementations • 21 Aug 2019 • Kwanyong Park, Sanghyun Woo, Dahun Kim, Donghyeon Cho, In So Kweon
In this paper, we investigate the problem of unpaired video-to-video translation.
no code implementations • 30 Jul 2019 • Ho-Deok Jang, Sanghyun Woo, Philipp Benz, Jinsun Park, In So Kweon
We present a simple yet effective prediction module for a one-stage detector.
1 code implementation • 11 Jul 2019 • Ukcheol Shin, Jinsun Park, Gyumin Shim, Francois Rameau, In So Kweon
In this paper, we propose a noise-aware exposure control algorithm for robust robot vision.
no code implementations • 30 May 2019 • Sanghyun Woo, Dahun Kim, KwanYong Park, Joon-Young Lee, In So Kweon
Our video inpainting network consists of two stages.
7 code implementations • CVPR 2019 • Donggeun Yoo, In So Kweon
In this paper, we propose a novel active learning method that is simple but task-agnostic, and works efficiently with the deep networks.
Ranked #3 on Active Learning on CIFAR10 (10,000)
1 code implementation • CVPR 2019 • Dahun Kim, Sanghyun Woo, Joon-Young Lee, In So Kweon
Blind video decaptioning is a problem of automatically removing text overlays and inpainting the occluded parts in videos without any input masks.
2 code implementations • CVPR 2019 • Dahun Kim, Sanghyun Woo, Joon-Young Lee, In So Kweon
Video inpainting aims to fill spatio-temporal holes with plausible content in a video.
Ranked #7 on Video Inpainting on DAVIS
1 code implementation • ICLR 2019 • Sunghoon Im, Hae-Gon Jeon, Stephen Lin, In So Kweon
The cost volume is constructed using a differentiable warping process that allows for end-to-end training of the network.
1 code implementation • CVPR 2019 • Junsik Kim, Tae-Hyun Oh, Seokju Lee, Fei Pan, In So Kweon
We take an approach to learn a generalizable embedding space for novel tasks.
1 code implementation • CVPR 2019 • Dong-Jin Kim, Jinsoo Choi, Tae-Hyun Oh, In So Kweon
Our goal in this work is to train an image captioning model that generates more dense and informative captions.
1 code implementation • 24 Nov 2018 • Yunjae Jung, Donghyeon Cho, Dahun Kim, Sanghyun Woo, In So Kweon
The proposed variance loss allows a network to predict output scores for each frame with high discrepancy which enables effective feature learning and significantly improves model performance.
Ranked #3 on Unsupervised Video Summarization on SumMe
Supervised Video Summarization Unsupervised Video Summarization
no code implementations • 24 Nov 2018 • Dahun Kim, Donghyeon Cho, In So Kweon
Self-supervised tasks such as colorization, inpainting and zigsaw puzzle have been utilized for visual representation learning for still images, when the number of labeled images is limited or absent at all.
Ranked #42 on Self-Supervised Action Recognition on HMDB51
3 code implementations • NeurIPS 2018 • Sanghyun Woo, Dahun Kim, Donghyeon Cho, In So Kweon
In this paper, we present a method that improves scene graph generation by explicitly modeling inter-dependency among the entire object instances.
10 code implementations • 17 Jul 2018 • Jongchan Park, Sanghyun Woo, Joon-Young Lee, In So Kweon
In this work, we focus on the effect of attention in general deep neural networks.
31 code implementations • ECCV 2018 • Sanghyun Woo, Jongchan Park, Joon-Young Lee, In So Kweon
We propose Convolutional Block Attention Module (CBAM), a simple yet effective attention module for feed-forward convolutional neural networks.
Ranked #7 on Object Detection on PKU-DDD17-Car
no code implementations • CVPR 2018 • Kyungdon Joo, Tae-Hyun Oh, In So Kweon, Jean-Charles Bazin
In this work, we describe man-made structures via an appropriate structure assumption, called Atlanta world, which contains a vertical direction (typically the gravity direction) and a set of horizontal directions orthogonal to the vertical direction.
1 code implementation • Pattern Recognition Letters 2018 • Oleksandr Bailo, Francois Rameau, Kyungdon Joo, Jinsun Park, Oleksandr Bogdan, In So Kweon
Keypoint detection usually results in a large number of keypoints which are mostly clustered, redundant, and noisy.
no code implementations • CVPR 2018 • Jongchan Park, Joon-Young Lee, Donggeun Yoo, In So Kweon
In addition, we present a 'distort-and-recover' training scheme which only requires high-quality reference images for training instead of input and retouched image pairs.
2 code implementations • CVPR 2018 • Changha Shin, Hae-Gon Jeon, Youngjin Yoon, In So Kweon, Seon Joo Kim
Light field cameras capture both the spatial and the angular properties of light rays in space.
no code implementations • CVPR 2018 • Sunghoon Im, Hae-Gon Jeon, In So Kweon
As demand for advanced photographic applications on hand-held devices grows, these electronics require the capture of high quality depth.
no code implementations • CVPR 2018 • Arda Senocak, Tae-Hyun Oh, Junsik Kim, Ming-Hsuan Yang, In So Kweon
We show that even with a few supervision, false conclusion is able to be corrected and the source of sound in a visual scene can be localized effectively.
no code implementations • 14 Feb 2018 • Dong-Jin Kim, Jinsoo Choi, Tae-Hyun Oh, Youngjin Yoon, In So Kweon
Human behavior understanding is arguably one of the most important mid-level components in artificial intelligence.
no code implementations • 6 Feb 2018 • Dahun Kim, Donghyeon Cho, Donggeun Yoo, In So Kweon
The recovery of the aforementioned damage pushes the network to obtain robust and general-purpose representations.
no code implementations • 5 Dec 2017 • Junsik Kim, Seokju Lee, Tae-Hyun Oh, In So Kweon
Recent advances in visual recognition show overarching success by virtue of large amounts of supervised data.
1 code implementation • PSIVT 2017 • Oleksandr Bogdan, Oleg Yurchenko, Oleksandr Bailo, Francois Rameau, Donggeun Yoo, In So Kweon
This paper proposes a wearable system for visually impaired people that can be utilized to obtain an extensive feedback about their surrounding environment.
1 code implementation • 20 Oct 2017 • Oleksandr Bailo, Francois Rameau, In So Kweon
In this paper, we propose an efficient algorithm for robust place recognition and loop detection using camera information only.
3 code implementations • ICCV 2017 • Seokju Lee, Junsik Kim, Jae Shin Yoon, Seunghak Shin, Oleksandr Bailo, Namil Kim, Tae-Hee Lee, Hyun Seok Hong, Seung-Hoon Han, In So Kweon
In this paper, we propose a unified end-to-end trainable multi-task network that jointly handles lane and road marking detection and recognition that is guided by a vanishing point under adverse weather conditions.
Ranked #1 on Lane Detection on Caltech Lanes Washington
no code implementations • ICCV 2017 • Hyowon Ha, Michal Perdoch, Hatem Alismail, In So Kweon, Yaser Sheikh
The recent proliferation of high resolution cameras presents an opportunity to achieve unprecedented levels of precision in visual 3D reconstruction.
no code implementations • 18 Sep 2017 • Sanghyun Woo, Soonmin Hwang, In So Kweon
One-stage object detectors such as SSD or YOLO already have shown promising accuracy with small memory footprint and fast speed.
no code implementations • 24 Aug 2017 • Inwook Shim, Tae-Hyun Oh, Joon-Young Lee, Jinwook Choi, Dong-Geol Choi, In So Kweon
We introduce a novel method to automatically adjust camera exposure for image processing and computer vision applications on mobile robot platforms.
no code implementations • ICCV 2017 • Jae Shin Yoon, Francois Rameau, Junsik Kim, Seokju Lee, Seunghak Shin, In So Kweon
We propose a novel video object segmentation algorithm based on pixel-level matching using Convolutional Neural Networks (CNN).
Ranked #73 on Semi-Supervised Video Object Segmentation on DAVIS 2016
2 code implementations • ICCV 2017 • Donghyeon Cho, Jinsun Park, Tae-Hyun Oh, Yu-Wing Tai, In So Kweon
Our method implicitly learns an attention map, which leads to a content-aware shift map for image retargeting.
no code implementations • ICCV 2017 • Tae-Hyun Oh, Kyungdon Joo, Neel Joshi, Baoyuan Wang, In So Kweon, Sing Bing Kang
Cinemagraphs are a compelling way to convey dynamic aspects of a scene.
no code implementations • ICCV 2017 • Dahun Kim, Donghyeon Cho, Donggeun Yoo, In So Kweon
Weakly supervised semantic segmentation and localiza- tion have a problem of focusing only on the most important parts of an image since they use only image-level annota- tions.
no code implementations • CVPR 2017 • Jaeheung Surh, Hae-Gon Jeon, Yunwon Park, Sunghoon Im, Hyowon Ha, In So Kweon
With the result from the FM, the role of a DfF pipeline is to determine and recalculate unreliable measurements while enhancing those that are reliable.
1 code implementation • CVPR 2017 • Jinsun Park, Yu-Wing Tai, Donghyeon Cho, In So Kweon
In this paper, we introduce robust and synergetic hand-crafted features and a simple but efficient deep feature from a convolutional neural network (CNN) architecture for defocus estimation.
Ranked #2 on Defocus Estimation on CUHK - Blur Detection Dataset
no code implementations • 6 Feb 2017 • Jinsoo Choi, Tae-Hyun Oh, In So Kweon
Despite the challenging baselines, our method still manages to show comparable or even exceeding performance.
no code implementations • 20 Dec 2016 • Donggeun Yoo, Sunggyun Park, Kyunghyun Paeng, Joon-Young Lee, In So Kweon
In this paper, we present an "action-driven" detection mechanism using our "top-down" visual attention model.
no code implementations • 18 Aug 2016 • Gyeongmin Choe, Jaesik Park, Yu-Wing Tai, In So Kweon
To resolve the ambiguity in our model between the normals and distances, we utilize an initial 3D mesh from the Kinect fusion and multi-view information to reliably estimate surface details that were not captured and reconstructed by the Kinect fusion.
no code implementations • 23 Jun 2016 • Hyoseok Hwang, Hyun Sung Chang, Dongkyung Nam, In So Kweon
Experimental results demonstrate that our method is quite accurate, about a half order of magnitude higher than prior work; is efficient, spending less than 2 s for computation; and is robust to noise, working well in the SNR regime as low as 6 dB.
no code implementations • CVPR 2016 • Kyungdon Joo, Tae-Hyun Oh, Junsik Kim, In So Kweon
Given a set of surface normals, we pose a Manhattan Frame (MF) estimation problem as a consensus set maximization that maximizes the number of inliers over the rotation search space.
no code implementations • CVPR 2016 • Jaesik Park, Yu-Wing Tai, Sudipta N. Sinha, In So Kweon
We present a robust low-rank matrix factorization method to estimate the unknown parameters of this model.
no code implementations • CVPR 2016 • Hae-Gon Jeon, Joon-Young Lee, Sunghoon Im, Hyowon Ha, In So Kweon
Consumer devices with stereo cameras have become popular because of their low-cost depth sensing capability.
1 code implementation • CVPR 2016 • Hyowon Ha, Sunghoon Im, Jaesik Park, Hae-Gon Jeon, In So Kweon
We propose a novel approach that generates a high-quality depth map from a set of images captured with a small viewpoint variation, namely small motion clip.
no code implementations • CVPR 2016 • Jinsoo Choi, Tae-Hyun Oh, In So Kweon
Inspired by plot analysis of written stories, our method generates a sequence of video clips ordered in such a way that it reflects plot dynamics and content coherency.
no code implementations • CVPR 2016 • Gyeongmin Choe, Srinivasa G. Narasimhan, In So Kweon
Near-Infrared (NIR) images of most materials exhibit less texture or albedo variations making them beneficial for vision tasks such as intrinsic image decomposition and structured light depth estimation.
no code implementations • 12 May 2016 • Kyungdon Joo, Tae-Hyun Oh, Junsik Kim, In So Kweon
Most man-made environments, such as urban and indoor scenes, consist of a set of parallel and orthogonal planar structures.
1 code implementation • 24 Mar 2016 • Donggeun Yoo, Namil Kim, Sunggyun Park, Anthony S. Paek, In So Kweon
We present an image-conditional image generation model.