no code implementations • ECCV 2020 • Sangryul Jeon, Dongbo Min, Seungryong Kim, Jihwan Choe, Kwanghoon Sohn
Establishing dense semantic correspondences requires dealing with large geometric variations caused by the unconstrained setting of images.
no code implementations • 12 Dec 2024 • Honggyu An, Jinhyeon Kim, Seonghoon Park, Jaewoo Jung, Jisang Han, Sunghwan Hong, Seungryong Kim
In this work, we explore new perspectives on cross-view completion learning by drawing an analogy to self-supervised correspondence learning.
no code implementations • 5 Dec 2024 • Donghoon Ahn, Jiwon Kang, SangHyun Lee, Jaewon Min, Minjae Kim, Wooseok Jang, Hyoungwon Cho, Sayak Paul, SeonHwa Kim, Eunju Cha, Kyong Hwan Jin, Seungryong Kim
Observing that noise obtained via diffusion inversion can reconstruct high-quality images without guidance, we focus on the initial noise of the denoising pipeline.
no code implementations • 4 Dec 2024 • Siyoon Jin, Jisu Nam, Jiyoung Kim, Dahyun Chung, Yeong-Seok Kim, Joonhyung Park, Heonjeong Chu, Seungryong Kim
Recent tuning-free approaches address this limitation by transferring local appearance from the exemplar image to the synthesized image through implicit cross-image matching in the augmented self-attention mechanism of pre-trained diffusion models.
no code implementations • 2 Dec 2024 • Sangbeom Lim, Seongchan Kim, Seungjun An, Seokju Cho, Paul Hongsuck Seo, Seungryong Kim
Thus, developing a new video segmentation dataset aimed at tracking multi-granularity segmentation target in the video scene is necessary.
no code implementations • 2 Dec 2024 • Seongchan Kim, Woojeong Jin, Sangbeom Lim, Heeji Yoon, Hyunwook Choi, Seungryong Kim
In this work, we present Selection by Object Language Alignment (SOLA), a novel framework that reformulates RVOS into two sub-problems, track generation and track selection.
no code implementations • 2 Dec 2024 • Wooseok Jang, Youngjun Hong, Geonho Cha, Seungryong Kim
Manipulation of facial images to meet specific controls such as pose, expression, and lighting, also known as face rigging, is a complex task in computer vision.
1 code implementation • 29 Oct 2024 • Sunghwan Hong, Jaewoo Jung, Heeseong Shin, Jisang Han, Jiaolong Yang, Chong Luo, Seungryong Kim
We then introduce lightweight, learnable modules to refine depth and pose estimates from the coarse alignments, improving the quality of 3D reconstruction and novel view synthesis.
no code implementations • 30 Sep 2024 • Heeseong Shin, Chaehyun Kim, Sunghwan Hong, Seokju Cho, Anurag Arnab, Paul Hongsuck Seo, Seungryong Kim
Large-scale vision-language models like CLIP have demonstrated impressive open-vocabulary capabilities for image-level tasks, excelling in recognizing what objects are present.
1 code implementation • 22 Jul 2024 • Seokju Cho, Jiahui Huang, Jisu Nam, Honggyu An, Seungryong Kim, Joon-Young Lee
We introduce LocoTrack, a highly accurate and efficient model designed for the task of tracking any point (TAP) across video sequences.
Ranked #1 on Point Tracking on TAP-Vid-DAVIS-First
no code implementations • 24 Jun 2024 • Min-Seop Kwak, Donghoon Ahn, Ines Hyeonsu Kim, Jin-Hwa Kim, Seungryong Kim
We demonstrate that our method significantly improves performance, successfully addressing the geometric inconsistency problems in text-to-3D generation task with minimal computation cost and being compatible with existing score distillation-based models.
no code implementations • 23 Jun 2024 • Inès Hyeonsu Kim, Joungbin Lee, Woojeong Jin, Soowon Son, Kyusun Cho, Junyoung Seo, Min-Seop Kwak, Seokju Cho, JeongYeol Baek, Byeongwon Lee, Seungryong Kim
Person re-identification (Re-ID) often faces challenges due to variations in human poses and camera viewpoints, which significantly affect the appearance of individuals across images.
1 code implementation • 27 May 2024 • Junyoung Seo, Kazumi Fukuda, Takashi Shibuya, Takuya Narihira, Naoki Murata, Shoukang Hu, Chieh-Hsin Lai, Seungryong Kim, Yuki Mitsufuji
In these methods, an input view is geometrically warped to novel views with estimated depth maps, then the warped image is inpainted by T2I models.
1 code implementation • 24 Apr 2024 • Kyusun Cho, Joungbin Lee, Heeji Yoon, Yeobin Hong, Jaehoon Ko, Sangjun Ahn, Seungryong Kim
A key insight is to encode the 3D Gaussian attributes into a shared implicit feature representation, where it is merged with audio features to manipulate each Gaussian attribute.
1 code implementation • 29 Mar 2024 • Jaehoon Ko, Kyusun Cho, Joungbin Lee, Heeji Yoon, Sangmin Lee, Sangjun Ahn, Seungryong Kim
Given the personalized 3D generative model, we present a novel audio-guided attention U-Net architecture that predicts the dynamic face variations in the NeRF space driven by audio.
1 code implementation • 28 Mar 2024 • Seyeon Kim, Siyoon Jin, JiHye Park, Kihong Kim, Jiyoung Kim, Jisu Nam, Seungryong Kim
AToM excels in capturing subtle lip movements by leveraging an audio attention mechanism.
3 code implementations • 26 Mar 2024 • Donghoon Ahn, Hyoungwon Cho, Jaewon Min, Wooseok Jang, Jungwoo Kim, SeonHwa Kim, Hyun Hee Park, Kyong Hwan Jin, Seungryong Kim
These techniques are often not applicable in unconditional generation or in various downstream tasks such as image restoration.
no code implementations • 17 Mar 2024 • Sunghwan Hong, Seokju Cho, Seungryong Kim, Stephen Lin
In this work, we first show that feature aggregation and cost aggregation exhibit distinct characteristics and reveal the potential for substantial benefits stemming from the judicious use of both aggregation processes.
1 code implementation • 14 Mar 2024 • Jaewoo Jung, Jisang Han, Honggyu An, Jiwon Kang, Seonghoon Park, Seungryong Kim
3D Gaussian splatting (3DGS) has recently demonstrated impressive capabilities in real-time novel view synthesis and 3D reconstruction.
1 code implementation • 7 Mar 2024 • Gyudong Kim, Mehdi Ghasemi, Soroush Heidari, Seungryong Kim, Young Geun Kim, Sarma Vrudhula, Carole-Jean Wu
Such fragmentation introduces a new type of data heterogeneity in FL, namely \textit{system-induced data heterogeneity}, as each device generates distinct data depending on its hardware and software configurations.
no code implementations • 28 Feb 2024 • Changho Choi, Minho Kim, Junhyeok Lee, Hyoung-Kyu Song, Younggeun Kim, Seungryong Kim
We show that our framework is applicable to other generators such as StyleNeRF, paving a way to 3D-aware face swapping and is also compatible with other downstream StyleGAN2 generator tasks.
no code implementations • 21 Feb 2024 • Kihong Kim, Haneol Lee, JiHye Park, Seyeon Kim, Kwanghee Lee, Seungryong Kim, Jaejun Yoo
Generating high-quality videos that synthesize desired realistic content is a challenging task due to their intricate high-dimensionality and complexity of videos.
1 code implementation • CVPR 2024 • Jisu Nam, Heesu Kim, Dongjae Lee, Siyoon Jin, Seungryong Kim, Seunggyu Chang
The objective of text-to-image (T2I) personalization is to customize a diffusion model to a user-provided reference concept, generating diverse images of the concept aligned with the target prompts.
1 code implementation • 5 Feb 2024 • Junyoung Seo, Susung Hong, Wooseok Jang, Inès Hyeonsu Kim, Minseop Kwak, Doyup Lee, Seungryong Kim
We leverage the retrieved asset to incorporate its geometric prior in the variational objective and adapt the diffusion model's 2D prior toward view consistency, achieving drastic improvements in both geometry and fidelity of generated scenes.
no code implementations • CVPR 2024 • Sunghwan Hong, Jaewoo Jung, Heeseong Shin, Jiaolong Yang, Seungryong Kim, Chong Luo
This work delves into the task of pose-free novel view synthesis from stereo pairs a challenging and pioneering task in 3D vision.
no code implementations • CVPR 2024 • Seokju Cho, Jiahui Huang, Seungryong Kim, Joon-Young Lee
In the domain of video tracking existing methods often grapple with a trade-off between spatial density and temporal range.
no code implementations • 22 Dec 2023 • Seungjun An, Seonghoon Park, Gyeongnyeon Kim, JeongYeol Baek, Byeongwon Lee, Seungryong Kim
With the increasing importance of video data in real-world applications, there is a rising need for efficient object detection methods that utilize temporal information.
1 code implementation • 21 Dec 2023 • Kwangrok Ryoo, Yeonsik Jo, Seungjun Lee, Mira Kim, Ahra Jo, Seung Hwan Kim, Seungryong Kim, Soonyoung Lee
For object detection task with noisy labels, it is important to consider not only categorization noise, as in image classification, but also localization noise, missing annotations, and bogus bounding boxes.
1 code implementation • 12 Dec 2023 • Sunghwan Hong, Jaewoo Jung, Heeseong Shin, Jiaolong Yang, Seungryong Kim, Chong Luo
This work delves into the task of pose-free novel view synthesis from stereo pairs, a challenging and pioneering task in 3D vision.
1 code implementation • 2 Dec 2023 • Jaewoo Jung, Jisang Han, Jiwon Kang, Seongchan Kim, Min-Seop Kwak, Seungryong Kim
We formulate few-shot NeRF into a teacher-student framework to guide the network to learn a more robust representation of the scene by training the student with additional pseudo labels generated from the teacher.
1 code implementation • 30 Nov 2023 • Jiwon Kim, Byeongho Heo, Sangdoo Yun, Seungryong Kim, Dongyoon Han
This paper builds on the hypothesis that there is an inherent data-hungry matter in learning semantic correspondences and uncovers the models can be more trained by employing densified training pairs.
1 code implementation • 17 Oct 2023 • Gyuseong Lee, Wooseok Jang, Jinhyeon Kim, Jaewoo Jung, Seungryong Kim
Our focus in this study is on leveraging the knowledge of large pretrained models to improve handling of OOD scenarios and tackle domain generalization problems.
Ranked #1 on Domain Generalization on Office-Home
no code implementations • 5 Jun 2023 • Sunwoo Kim, Wooseok Jang, Hyunsu Kim, Junho Kim, Yunjey Choi, Seungryong Kim, Gayeong Lee
From the users' standpoint, prompt engineering is a labor-intensive process, and users prefer to provide a target word for editing instead of a full sentence.
1 code implementation • 30 May 2023 • Jisu Nam, Gyuseong Lee, Sunwoo Kim, Hyeonsu Kim, Hyoungwon Cho, Seyeon Kim, Seungryong Kim
The objective for establishing dense correspondence between paired images consists of two terms: a data term and a prior term.
1 code implementation • 30 May 2023 • Jiuhn Song, Seonghoon Park, Honggyu An, Seokju Cho, Min-Seop Kwak, SungJin Cho, Seungryong Kim
Employing monocular depth estimation (MDE) networks, pretrained on large-scale RGB-D datasets, with powerful generalization capability would be a key to solving this problem: however, using MDE in conjunction with NeRF comes with a new set of challenges due to various ambiguity problems exhibited by monocular depths.
1 code implementation • 23 May 2023 • Susung Hong, Junyoung Seo, Heeseong Shin, Sunghwan Hong, Seungryong Kim
In the paradigm of AI-generated content (AIGC), there has been increasing attention to transferring knowledge from pre-trained text-to-image (T2I) models to text-to-video (T2V) generation.
no code implementations • 11 Apr 2023 • Soohyun Kim, Junho Kim, Taekyung Kim, Hwan Heo, Seungryong Kim, Jiyoung Lee, Jin-Hwa Kim
This task is difficult due to the geometric distortion of panoramic images and the lack of a panoramic image dataset with diverse conditions, like weather or time.
no code implementations • CVPR 2023 • Minsu Kim, Seungryong Kim, Jungin Park, Seongheon Park, Kwanghoon Sohn
Modern data augmentation using a mixture-based technique can regularize the models from overfitting to the training data in various computer vision applications, but a proper data augmentation technique tailored for the part-based Visible-Infrared person Re-IDentification (VI-ReID) models remains unexplored.
1 code implementation • NeurIPS 2023 • Susung Hong, Donghoon Ahn, Seungryong Kim
In this work, we explore existing frameworks for score-distilling text-to-3D generation and identify the main causes of the view inconsistency problem -- the embedded bias of 2D diffusion models.
no code implementations • 21 Mar 2023 • SeokYeong Lee, Junyong Choi, Seungryong Kim, Ig-Jae Kim, Junghyun Cho
In this paper, we introduce a new challenge for synthesizing novel view images in practical environments with limited input multi-view images and varying lighting conditions.
3 code implementations • CVPR 2024 • Seokju Cho, Heeseong Shin, Sunghwan Hong, Anurag Arnab, Paul Hongsuck Seo, Seungryong Kim
Open-vocabulary semantic segmentation presents the challenge of labeling each pixel within an image based on a wide range of text descriptions.
Ranked #1 on Open Vocabulary Semantic Segmentation on ADE20K-150
1 code implementation • 14 Mar 2023 • Junyoung Seo, Wooseok Jang, Min-Seop Kwak, Hyeonsu Kim, Jaehoon Ko, Junho Kim, Jin-Hwa Kim, Jiyoung Lee, Seungryong Kim
Text-to-3D generation has shown rapid progress in recent days with the advent of score distillation, a methodology of using pretrained text-to-2D diffusion models to optimize neural radiance field (NeRF) in the zero-shot setting.
1 code implementation • 26 Jan 2023 • Min-Seop Kwak, Jiuhn Song, Seungryong Kim
We present a novel framework to regularize Neural Radiance Field (NeRF) in a few-shot setting with a geometry-aware consistency regularization.
1 code implementation • 6 Jan 2023 • Jungwoo Lim, Myunghoon Kang, Yuna Hur, SeungWon Jung, Jinsung Kim, Yoonna Jang, Dongyub Lee, Hyesung Ji, Donghoon Shin, Seungryong Kim, Heuiseok Lim
The agent selects the proper knowledge and persona to use for generating the answers with our candidate scoring implemented with a poly-encoder.
1 code implementation • 27 Dec 2022 • Kihong Kim, Yunho Kim, Seokju Cho, Junyoung Seo, Jisu Nam, Kychul Lee, Seungryong Kim, Kwanghee Lee
In this paper, we propose a diffusion-based face swapping framework for the first time, called DiffFace, composed of training ID conditional DDPM, sampling with facial guidance, and a target-preserving blending.
1 code implementation • 21 Dec 2022 • Jongbeom Baek, Gyeongnyeon Kim, Seonghoon Park, Honggyu An, Matteo Poggi, Seungryong Kim
We propose MaskingDepth, a novel semi-supervised learning framework for monocular depth estimation to mitigate the reliance on large ground-truth depth quantities.
1 code implementation • 17 Dec 2022 • Gyeongnyeon Kim, Wooseok Jang, Gyuseong Lee, Susung Hong, Junyoung Seo, Seungryong Kim
Generative models have recently undergone significant advancement due to the diffusion models.
no code implementations • 20 Nov 2022 • Daehwan Kim, Kwangrok Ryoo, Hansang Cho, Seungryong Kim
To address this, some methods were proposed to automatically split clean and noisy labels, and learn a semi-supervised learner in a Learning with Noisy Labels (LNL) framework.
1 code implementation • 14 Oct 2022 • Sunwoo Kim, Youngjo Min, Younghun Jung, Seungryong Kim
We propose a controllable style transfer framework based on Implicit Neural Representation that pixel-wisely controls the stylized output via test-time training.
1 code implementation • 13 Oct 2022 • Jaehoon Ko, Kyusun Cho, Daewon Choi, Kwangrok Ryoo, Seungryong Kim
With the recent advances in NeRF-based 3D aware GANs quality, projecting an image into the latent space of these 3D-aware GANs has a natural advantage over 2D GAN inversion: not only does it allow multi-view consistent editing of the projected image, but it also enables 3D reconstruction and novel view synthesis when given only a single image.
1 code implementation • 6 Oct 2022 • Sunghwan Hong, Jisu Nam, Seokju Cho, Susung Hong, Sangryul Jeon, Dongbo Min, Seungryong Kim
Existing pipelines of semantic correspondence commonly include extracting high-level semantic features for the invariance against intra-class variations and background clutters.
no code implementations • 4 Oct 2022 • Yunsung Lee, Gyuseong Lee, Kwangrok Ryoo, Hyojun Go, JiHye Park, Seungryong Kim
In addition, through Fourier analysis of feature maps, the model's response patterns according to signal frequency changes, we observe which inductive bias is advantageous for each data scale.
5 code implementations • ICCV 2023 • Susung Hong, Gyuseong Lee, Wooseok Jang, Seungryong Kim
Denoising diffusion models (DDMs) have attracted attention for their exceptional generation quality and diversity.
1 code implementation • 22 Sep 2022 • Junyoung Seo, Gyuseong Lee, Seokju Cho, Jiyoung Lee, Seungryong Kim
Specifically, we formulate a diffusion-based matching-and-generation framework that interleaves cross-domain matching and diffusion steps in the latent space by iteratively feeding the intermediate warp into the noising process and denoising it to generate a translated image.
no code implementations • 19 Sep 2022 • Sunghwan Hong, Seokju Cho, Seungryong Kim, Stephen Lin
The current state-of-the-art are Transformer-based approaches that focus on either feature descriptors or cost volume aggregation.
Ranked #1 on Geometric Matching on HPatches
1 code implementation • CVPR 2023 • JiHye Park, Sunwoo Kim, Soohyun Kim, Seokju Cho, Jaejun Yoo, Youngjung Uh, Seungryong Kim
Existing techniques for image-to-image translation commonly have suffered from two critical problems: heavy reliance on per-sample domain annotation and/or inability of handling multiple attributes per image.
1 code implementation • 18 Aug 2022 • Jiwon Kim, Youngjo Min, Daehwan Kim, Gyuseong Lee, Junyoung Seo, Kwangrok Ryoo, Seungryong Kim
We present a novel semi-supervised learning framework that intelligently leverages the consistency regularization between the model's predictions from two strongly-augmented views of an image, weighted by a confidence of pseudo-label, dubbed ConMatch.
1 code implementation • 22 Jul 2022 • Sunghwan Hong, Seokju Cho, Jisu Nam, Stephen Lin, Seungryong Kim
However, the tokenization of a correlation map for transformer processing can be detrimental, because the discontinuity at token boundaries reduces the local context available near the token edges and decreases inductive bias.
Ranked #2 on Semantic correspondence on PF-WILLOW
no code implementations • 28 Apr 2022 • Mira Kim, Jaehoon Ko, Kyusun Cho, Junmyeong Choi, Daewon Choi, Seungryong Kim
We propose a novel framework for 3D-aware object manipulation, called Auto-Encoding Neural Radiance Fields (AE-NeRF).
no code implementations • 5 Apr 2022 • Jiwon Kim, Youngjo Min, Mira Kim, Seungryong Kim
In this paper, we propose a novel framework for jointly learning feature extraction and cost aggregation for semantic correspondence.
1 code implementation • CVPR 2022 • Soohyun Kim, Jongbeom Baek, JiHye Park, Gyeongnyeon Kim, Seungryong Kim
By augmenting such tokens with an instance-level feature extracted from the content feature with respect to bounding box information, our framework is capable of learning an interaction between object instances and the global image, thus boosting the instance-awareness.
no code implementations • CVPR 2022 • Jiwon Kim, Kwangrok Ryoo, Junyoung Seo, Gyuseong Lee, Daehwan Kim, Hansang Cho, Seungryong Kim
In this paper, we present a simple, but effective solution for semantic correspondence that learns the networks in a semi-supervised manner by supplementing few ground-truth correspondences via utilization of a large amount of confident correspondences as pseudo-labels, called SemiMatch.
no code implementations • 18 Mar 2022 • Jongbeom Baek, Gyeongnyeon Kim, Seungryong Kim
We propose a semi-supervised learning framework for monocular depth estimation.
1 code implementation • 14 Feb 2022 • Seokju Cho, Sunghwan Hong, Seungryong Kim
Cost aggregation is a highly important process in image matching tasks, which aims to disambiguate the noisy matching scores.
Ranked #3 on Semantic correspondence on PF-PASCAL
no code implementations • 25 Jan 2022 • Jiwon Kim, Kwangrok Ryoo, Gyuseong Lee, Seokju Cho, Junyoung Seo, Daehwan Kim, Hansang Cho, Seungryong Kim
In this paper, we address this limitation with a novel SSL framework for aggregating pseudo labels, called AggMatch, which refines initial pseudo labels by using different confident instances.
no code implementations • 6 Jan 2022 • Jaehoon Cho, Seungryong Kim, Kwanghoon Sohn
To address this problem, we propose a novel network architecture based on a memory network that explicitly helps to capture long-term rain streak information in the time-lapse data.
2 code implementations • 22 Dec 2021 • Sunghwan Hong, Seokju Cho, Jisu Nam, Seungryong Kim
We introduce a novel cost aggregation network, dubbed Volumetric Aggregation with Transformers (VAT), to tackle the few-shot segmentation task by using both convolutions and transformers to efficiently handle high dimensional correlation maps between query and support.
Ranked #3 on Semantic correspondence on PF-WILLOW
2 code implementations • 16 Dec 2021 • Yoonna Jang, Jungwoo Lim, Yuna Hur, Dongsuk Oh, Suhyune Son, Yeonsoo Lee, Donghoon Shin, Seungryong Kim, Heuiseok Lim
Humans usually have conversations by making use of prior knowledge about a topic and background information of the people whom they are talking to.
1 code implementation • 12 Dec 2021 • Sunwoo Kim, Soohyun Kim, Seungryong Kim
Recent techniques to solve photorealistic style transfer within deep convolutional neural networks (CNNs) generally require intensive training from large-scale datasets, thus having limited applicability and poor generalization ability to unseen images or styles.
no code implementations • 21 Nov 2021 • Yunsung Lee, Teakgyu Hong, Han-Cheol Cho, Junbum Cha, Seungryong Kim
Compared to previous works, our method shows better or comparable performance on dense prediction fine-tuning tasks.
no code implementations • 17 Aug 2021 • Hojoon Lee, Dongyoon Hwang, Sunghwan Hong, Changyeon Kim, Seungryong Kim, Jaegul Choo
Successful sequential recommendation systems rely on accurately capturing the user's short-term and long-term interest.
no code implementations • ICCV 2021 • Sunghun Joung, Seungryong Kim, Minsu Kim, Ig-Jae Kim, Kwanghoon Sohn
By incorporating 3D shape and appearance jointly in a deep representation, our method learns the discriminative representation of the object and achieves competitive performance on fine-grained image recognition and vehicle re-identification.
no code implementations • CVPR 2021 • Sangryul Jeon, Dongbo Min, Seungryong Kim, Kwanghoon Sohn
We present a novel framework for contrastive learning of pixel-level representation using only unlabeled video.
1 code implementation • ICCV 2021 • Sunghwan Hong, Seungryong Kim
Conventional techniques to establish dense correspondences across visually or semantically similar images focused on designing a task-specific matching prior, which is difficult to model.
Ranked #1 on Dense Pixel Correspondence Estimation on HPatches (using extra training data)
1 code implementation • NeurIPS 2021 • Seokju Cho, Sunghwan Hong, Sangryul Jeon, Yunsung Lee, Kwanghoon Sohn, Seungryong Kim
We propose a novel cost aggregation network, called Cost Aggregation Transformers (CATs), to find dense correspondences between semantically similar images with additional challenges posed by large intra-class appearance and geometric variations.
Ranked #5 on Semantic correspondence on PF-WILLOW
1 code implementation • 8 Apr 2021 • Bahar Aydemir, Deblina Bhattacharjee, Tong Zhang, Seungryong Kim, Mathieu Salzmann, Sabine Süsstrunk
Saliency prediction has made great strides over the past two decades, with current techniques modeling low-level information, such as color, intensity and size contrasts, and high-level ones, such as attention and gaze direction for entire objects.
2 code implementations • CVPR 2021 • Sungha Choi, Sanghun Jung, Huiwon Yun, Joanne Kim, Seungryong Kim, Jaegul Choo
Enhancing the generalization capability of deep neural networks to unseen domains is crucial for safety-critical applications in the real world such as autonomous driving.
Ranked #5 on Robust Object Detection on DWD
1 code implementation • 2 Jan 2021 • Matteo Poggi, Seungryong Kim, Fabio Tosi, Sunok Kim, Filippo Aleotti, Dongbo Min, Kwanghoon Sohn, Stefano Mattoccia
Stereo matching is one of the most popular techniques to estimate dense depth maps by finding the disparity between matching pixels on two, synchronized and rectified images.
1 code implementation • 15 Dec 2020 • Minsu Kim, Sunghun Joung, Seungryong Kim, Jungin Park, Ig-Jae Kim, Kwanghoon Sohn
Existing techniques to adapt semantic segmentation networks across the source and target domains within deep convolutional neural networks (CNNs) deal with all the samples from the two domains in a global or category-aware manner.
no code implementations • 18 Nov 2020 • Taewon Kang, Soohyun Kim, Sunwoo Kim, Seungryong Kim
Existing techniques to solve exemplar-based image-to-image translation within deep convolutional neural networks (CNNs) generally require a training phase to optimize the network parameters on domain-specific and task-specific benchmarks, thus having limited applicability and generalization ability.
1 code implementation • ICCV 2021 • Hyesong Choi, Hunsang Lee, Sunkyung Kim, Sunok Kim, Seungryong Kim, Kwanghoon Sohn, Dongbo Min
To cope with the prediction error of the confidence map itself, we also leverage the threshold network that learns the threshold dynamically conditioned on the pseudo depth maps.
no code implementations • ECCV 2020 • Seungryong Kim, Sabine Süsstrunk, Mathieu Salzmann
We design our VTN as an encoder-decoder network, with modules dedicated to letting the information flow across the feature channels, to account for the dependencies between the semantic parts.
no code implementations • CVPR 2020 • Sunghun Joung, Seungryong Kim, Hanjae Kim, Minsu Kim, Ig-Jae Kim, Junghyun Cho, Kwanghoon Sohn
To overcome this limitation, we introduce a learnable module, cylindrical convolutional networks (CCNs), that exploit cylindrical representation of a convolutional kernel defined in the 3D space.
no code implementations • ICCV 2019 • Sangryul Jeon, Dongbo Min, Seungryong Kim, Kwanghoon Sohn
Based on the key insight that the two tasks can mutually provide supervisions to each other, our networks accomplish this through a joint loss function that alternatively imposes a consistency constraint between the two tasks, thereby boosting the performance and addressing the lack of training data in a principled manner.
1 code implementation • ICCV 2019 • Jiyoung Lee, Seungryong Kim, Sunok Kim, Jungin Park, Kwanghoon Sohn
We present deep networks for context-aware emotion recognition, called CAER-Net, that exploit not only human facial expression but also context information in a joint and boosting manner.
Ranked #1 on Emotion Recognition in Context on CAER-Dynamic
no code implementations • CVPR 2019 • Seungryong Kim, Dongbo Min, Somi Jeong, Sunok Kim, Sangryul Jeon, Kwanghoon Sohn
SAM-Net accomplishes this through an iterative process of establishing reliable correspondences by reducing the attribute discrepancy between the images and synthesizing attribute transferred images using the learned correspondences.
1 code implementation • NeurIPS 2018 • Seungryong Kim, Stephen Lin, Sangryul Jeon, Dongbo Min, Kwanghoon Sohn
Our networks accomplish this through an iterative process of estimating spatial transformations between the input images and using these transformations to generate aligned convolutional activations.
no code implementations • ECCV 2018 • Sangryul Jeon, Seungryong Kim, Dongbo Min, Kwanghoon Sohn
To the best of our knowledge, it is the first work that attempts to estimate dense affine transformation fields in a coarse-to-fine manner within deep networks.
no code implementations • ICCV 2017 • Seungryong Kim, Dongbo Min, Stephen Lin, Kwanghoon Sohn
In this way, our approach draws solutions from the continuous space of affine transformations in a manner that can be computed efficiently through constant-time edge-aware filtering and a proposed affine-varying CNN-based descriptor.
1 code implementation • CVPR 2017 • Seungryong Kim, Dongbo Min, Bumsub Ham, Sangryul Jeon, Stephen Lin, Kwanghoon Sohn
The sampling patterns of local structure and the self-similarity measure are jointly learned within the proposed network in an end-to-end and multi-scale manner.
no code implementations • 27 Apr 2016 • Seungryong Kim, Dongbo Min, Bumsub Ham, Minh N. Do, Kwanghoon Sohn
In this paper, we propose a novel dense descriptor, called dense adaptive self-correlation (DASC), to estimate multi-modal and multi-spectral dense correspondences.
no code implementations • 21 Mar 2016 • Seungryong Kim, Dongbo Min, Stephen Lin, Kwanghoon Sohn
We present a novel descriptor, called deep self-convolutional activations (DeSCA), designed for establishing dense correspondences between images taken under different imaging modalities, such as different spectral ranges or lighting conditions.
1 code implementation • 21 Mar 2016 • Seungryong Kim, Kihong Park, Kwanghoon Sohn, Stephen Lin
We present a method for jointly predicting a depth map and intrinsic images from single-image input.
no code implementations • CVPR 2015 • Seungryong Kim, Dongbo Min, Bumsub Ham, Seungchul Ryu, Minh N. Do, Kwanghoon Sohn
To further improve the matching quality and runtime efficiency, we propose a patch-wise receptive field pooling, in which a sampling pattern is optimized with a discriminative learning.