1 code implementation • ECCV 2020 • Seonguk Seo, Joon-Young Lee, Bohyung Han
We propose a unified referring video object segmentation network (URVOS).
Ranked #7 on Referring Video Object Segmentation on MeViS
1 code implementation • 22 Jul 2024 • Seokju Cho, Jiahui Huang, Jisu Nam, Honggyu An, Seungryong Kim, Joon-Young Lee
We introduce LocoTrack, a highly accurate and efficient model designed for the task of tracking any point (TAP) across video sequences.
Ranked #1 on Point Tracking on TAP-Vid-DAVIS-First
1 code implementation • 9 Jul 2024 • Jeongseok Hyun, Su Ho Han, Hyolim Kang, Joon-Young Lee, Seon Joo Kim
First, a class-agnostic action localizer is trained on a human-labeled TAL dataset and used to generate pseudo-labels for unlabeled videos.
no code implementations • CVPR 2024 • Chuong Huynh, Seoung Wug Oh, Abhinav Shrivastava, Joon-Young Lee
Human matting is a foundation task in image and video processing, where human foreground pixels are extracted from the input.
no code implementations • 17 Apr 2024 • Adrit Rao, Andrea Fisher, Ken Chang, John Christopher Panagides, Katherine McNamara, Joon-Young Lee, Oliver Aalami
We propose the Interactive Medical Image Learning (IMIL) framework, a novel approach for improving the training of medical image analysis algorithms that enables clinician-guided intermediate training data augmentations on misprediction outliers, focusing the algorithm on relevant visual information.
no code implementations • CVPR 2024 • Gihyun Kwon, Simon Jenni, DIngzeyu Li, Joon-Young Lee, Jong Chul Ye, Fabian Caba Heilbron
While there has been significant progress in customizing text-to-image generation models, generating images that combine multiple personalized concepts remains challenging.
no code implementations • CVPR 2024 • Seokju Cho, Jiahui Huang, Seungryong Kim, Joon-Young Lee
In the domain of video tracking existing methods often grapple with a trade-off between spatial density and temporal range.
1 code implementation • CVPR 2024 • Ho Kei Cheng, Seoung Wug Oh, Brian Price, Joon-Young Lee, Alexander Schwing
We present Cutie, a video object segmentation (VOS) network with object-level memory reading, which puts the object representation from memory back into the video object segmentation result.
Ranked #1 on Semi-Supervised Video Object Segmentation on MOSE
1 code implementation • ICCV 2023 • Ho Kei Cheng, Seoung Wug Oh, Brian Price, Alexander Schwing, Joon-Young Lee
To 'track anything' without training on video data for every individual task, we develop a decoupled video segmentation approach (DEVA), composed of task-specific image-level segmentation and class/task-agnostic bi-directional temporal propagation.
Ranked #1 on Unsupervised Video Object Segmentation on DAVIS 2016 val (using extra training data)
Open-Vocabulary Video Segmentation Open-World Video Segmentation +7
no code implementations • 23 Aug 2023 • Adrit Rao, Joon-Young Lee, Oliver Aalami
In this paper, we evaluate the effects of three modern augmentation techniques, CutMix, MixUp, and CutOut on the calibration and performance of CNNs for medical tasks.
no code implementations • ICCV 2023 • Dawit Mureja Argaw, Joon-Young Lee, Markus Woodson, In So Kweon, Fabian Caba Heilbron
While great progress has been attained, there is still a need for a pretrained multimodal model that can perform well in the ever-growing set of movie understanding tasks the community has been establishing.
2 code implementations • ICCV 2023 • Maksym Bekuzarov, Ariana Bermudez, Joon-Young Lee, Hao Li
Despite advancements in user-guided video segmentation, extracting complex objects consistently for highly complex scenes is still a labor-intensive task, especially for production.
no code implementations • 15 Jul 2023 • Jiahui Huang, Leonid Sigal, Kwang Moo Yi, Oliver Wang, Joon-Young Lee
We present Interactive Neural Video Editing (INVE), a real-time video editing solution, which can assist the video editing process by consistently propagating sparse frame edits to the entire video clip.
no code implementations • CVPR 2023 • KwanYong Park, Sanghyun Woo, Seoung Wug Oh, In So Kweon, Joon-Young Lee
Mask-guided matting has shown great practicality compared to traditional trimap-based methods.
no code implementations • 20 Dec 2022 • Sanghyun Woo, KwanYong Park, Seoung Wug Oh, In So Kweon, Joon-Young Lee
First, no tracking supervisions are in LVIS, which leads to inconsistent learning of detection (with LVIS and TAO) and tracking (only with TAO).
no code implementations • 20 Dec 2022 • Sanghyun Woo, KwanYong Park, Seoung Wug Oh, In So Kweon, Joon-Young Lee
The tracking-by-detection paradigm today has become the dominant method for multi-object tracking and works by detecting objects in each frame and then performing data association across frames.
no code implementations • 22 Nov 2022 • David Chuan-En Lin, Fabian Caba Heilbron, Joon-Young Lee, Oliver Wang, Nikolas Martelaro
This paper investigates the challenge of extracting highlight moments from videos.
no code implementations • 22 Nov 2022 • David Chuan-En Lin, Fabian Caba Heilbron, Joon-Young Lee, Oliver Wang, Nikolas Martelaro
Video editing is a creative and complex endeavor and we believe that there is potential for reimagining a new video editing interface to better support the creative and exploratory nature of video editing.
1 code implementation • CVPR 2023 • Miran Heo, Sukjun Hwang, Jeongseok Hyun, Hanjung Kim, Seoung Wug Oh, Joon-Young Lee, Seon Joo Kim
Notably, we greatly outperform the state-of-the-art on the long VIS benchmark (OVIS), improving 5. 6 AP with ResNet-50 backbone.
Ranked #8 on Video Instance Segmentation on YouTube-VIS 2021 (using extra training data)
1 code implementation • CVPR 2022 • KwanYong Park, Sanghyun Woo, Seoung Wug Oh, In So Kweon, Joon-Young Lee
In this per-clip inference scheme, we update the memory with an interval and simultaneously process a set of consecutive frames (i. e. clip) between the memory updates.
1 code implementation • 27 Jul 2022 • Hongje Seong, Seoung Wug Oh, Brian Price, Euntai Kim, Joon-Young Lee
A key of OTVM is the joint modeling of trimap propagation and alpha prediction.
1 code implementation • 20 Jul 2022 • Dawit Mureja Argaw, Fabian Caba Heilbron, Joon-Young Lee, Markus Woodson, In So Kweon
Machine learning is transforming the video editing industry.
1 code implementation • 9 Jun 2022 • Miran Heo, Sukjun Hwang, Seoung Wug Oh, Joon-Young Lee, Seon Joo Kim
Specifically, we use an image object detector as a means of distilling object-specific contexts into object tokens.
Ranked #13 on Video Instance Segmentation on YouTube-VIS 2021 (using extra training data)
no code implementations • 10 Jan 2022 • Seonguk Seo, Joon-Young Lee, Bohyung Han
We propose an information-theoretic bias measurement technique through a causal interpretation of spurious correlation, which is effective to identify the feature-level algorithmic bias by taking advantage of conditional mutual information.
1 code implementation • ICCV 2021 • Hongje Seong, Seoung Wug Oh, Joon-Young Lee, Seongwon Lee, Suhyeon Lee, Euntai Kim
Based on a recent memory-based method [33], we propose two advanced memory read modules that enable us to perform memory reading in multiple scales while exploiting temporal smoothness.
no code implementations • 2 Sep 2021 • Adrit Rao, Jongchan Park, Sanghyun Woo, Joon-Young Lee, Oliver Aalami
The use of computer vision to automate the classification of medical images is widely studied.
1 code implementation • CVPR 2022 • Seonguk Seo, Joon-Young Lee, Bohyung Han
Dataset bias is a critical challenge in machine learning since it often leads to a negative impact on a model due to the unintended decision rules captured by spurious correlations.
no code implementations • CVPR 2021 • Sanghyun Woo, Dahun Kim, Joon-Young Lee, In So Kweon
Temporal correspondence - linking pixels or objects across frames - is a fundamental supervisory signal for the video models.
Ranked #6 on Video Panoptic Segmentation on Cityscapes-VPS (using extra training data)
1 code implementation • 3 Jun 2021 • Juan Leon Alcazar, Long Mai, Federico Perazzi, Joon-Young Lee, Pablo Arbelaez, Bernard Ghanem, Fabian Caba Heilbron
To showcase the potential of our new dataset, we propose an audiovisual baseline and benchmark for person retrieval.
1 code implementation • CVPR 2021 • Gunhee Nam, Miran Heo, Seoung Wug Oh, Joon-Young Lee, Seon Joo Kim
Since the existing datasets are not suitable to validate our method, we build a new polygonal point set tracking dataset and demonstrate the superior performance of our method over the baselines and existing contour-based VOS methods.
3 code implementations • CVPR 2021 • Jaedong Hwang, Seoung Wug Oh, Joon-Young Lee, Bohyung Han
We extend panoptic segmentation to the open-world and introduce an open-set panoptic segmentation (OPS) task.
1 code implementation • CVPR 2020 • Dahun Kim, Sanghyun Woo, Joon-Young Lee, In So Kweon
In this paper, we propose and explore a new video extension of this task, called video panoptic segmentation.
Ranked #7 on Video Panoptic Segmentation on Cityscapes-VPS (using extra training data)
1 code implementation • CVPR 2020 • Juan Leon Alcazar, Fabian Caba Heilbron, Long Mai, Federico Perazzi, Joon-Young Lee, Pablo Arbelaez, Bernard Ghanem
Current methods for active speak er detection focus on modeling short-term audiovisual information from a single speaker.
Active Speaker Detection Audio-Visual Active Speaker Detection
2 code implementations • ACL 2020 • Shubham Agarwal, Trung Bui, Joon-Young Lee, Ioannis Konstas, Verena Rieser
Visual Dialog involves "understanding" the dialog history (what has been discussed previously) and the current question (what is asked), in addition to grounding information in the image, to generate the correct response.
no code implementations • 20 Mar 2020 • Gunhee Nam, Seoung Wug Oh, Joon-Young Lee, Seon Joo Kim
We propose a novel memory-based tracker via part-level dense memory and voting-based retrieval, called DMV.
1 code implementation • ICCV 2019 • Seoung Wug Oh, Sungho Lee, Joon-Young Lee, Seon Joo Kim
Given a set of reference images and a target image with holes, our network fills the hole by referring the contents in the reference images.
no code implementations • 30 May 2019 • Sanghyun Woo, Dahun Kim, KwanYong Park, Joon-Young Lee, In So Kweon
Our video inpainting network consists of two stages.
2 code implementations • CVPR 2019 • Xingyu Liu, Joon-Young Lee, Hailin Jin
In particular, it can effectively learn representations for videos by mixing appearance and long-range motion with an RGB-only input.
1 code implementation • CVPR 2019 • Dahun Kim, Sanghyun Woo, Joon-Young Lee, In So Kweon
Blind video decaptioning is a problem of automatically removing text overlays and inpainting the occluded parts in videos without any input masks.
2 code implementations • CVPR 2019 • Dahun Kim, Sanghyun Woo, Joon-Young Lee, In So Kweon
Video inpainting aims to fill spatio-temporal holes with plausible content in a video.
Ranked #7 on Video Inpainting on DAVIS
1 code implementation • CVPR 2019 • Seoung Wug Oh, Joon-Young Lee, Ning Xu, Seon Joo Kim
We propose a new multi-round training scheme for the interactive video object segmentation so that the networks can learn how to understand the user's intention and update incorrect estimations during the training.
Ranked #6 on Interactive Video Object Segmentation on DAVIS 2017 (AUC-J metric)
3 code implementations • ICCV 2019 • Seoung Wug Oh, Joon-Young Lee, Ning Xu, Seon Joo Kim
In our framework, the past frames with object masks form an external memory, and the current frame as the query is segmented using the mask information in the memory.
Ranked #4 on Interactive Video Object Segmentation on DAVIS 2017 (using extra training data)
no code implementations • 21 Sep 2018 • Xin Ye, Zhe Lin, Joon-Young Lee, Jianming Zhang, Shibin Zheng, Yezhou Yang
We study the problem of learning a generalizable action policy for an intelligent agent to actively approach an object of interest in an indoor environment solely from its visual inputs.
1 code implementation • ECCV 2018 • Wei-Chih Hung, Jianming Zhang, Xiaohui Shen, Zhe Lin, Joon-Young Lee, Ming-Hsuan Yang
Specifically, given a foreground image and a background image, our proposed method automatically generates a set of blending photos with scores that indicate the aesthetics quality with the proposed quality network and policy network.
no code implementations • ECCV 2018 • Fabian Caba Heilbron, Joon-Young Lee, Hailin Jin, Bernard Ghanem
In this paper, we introduce a novel active learning framework for temporal localization that aims to mitigate this data dependency issue.
no code implementations • ECCV 2018 • Rameswar Panda, Jianming Zhang, Haoxiang Li, Joon-Young Lee, Xin Lu, Amit K. Roy-Chowdhury
While machine learning approaches to visual emotion recognition offer great promise, current methods consider training and testing models on small scale datasets covering limited visual emotion concepts.
10 code implementations • 17 Jul 2018 • Jongchan Park, Sanghyun Woo, Joon-Young Lee, In So Kweon
In this work, we focus on the effect of attention in general deep neural networks.
31 code implementations • ECCV 2018 • Sanghyun Woo, Jongchan Park, Joon-Young Lee, In So Kweon
We propose Convolutional Block Attention Module (CBAM), a simple yet effective attention module for feed-forward convolutional neural networks.
Ranked #7 on Object Detection on PKU-DDD17-Car
2 code implementations • CVPR 2018 • Seoung Wug Oh, Joon-Young Lee, Kalyan Sunkavalli, Seon Joo Kim
We validate our method on four benchmark sets that cover single and multiple object segmentation.
no code implementations • CVPR 2018 • Jongchan Park, Joon-Young Lee, Donggeun Yoo, In So Kweon
In addition, we present a 'distort-and-recover' training scheme which only requires high-quality reference images for training instead of input and retouched image pairs.
1 code implementation • 24 Feb 2018 • Kaichun Mo, Haoxiang Li, Zhe Lin, Joon-Young Lee
Synthetic data suffers from domain gap to the real-world scenes while visual inputs rendered from 3D reconstructed scenes have undesired holes and artifacts.
Robotics
no code implementations • ICCV 2017 • Zhuo Hui, Kalyan Sunkavalli, Joon-Young Lee, Sunil Hadap, Jian Wang, Aswin C. Sankaranarayanan
A collocated setup provides only a 1-D "univariate" sampling of the 4-D BRDF.
no code implementations • 24 Aug 2017 • Inwook Shim, Tae-Hyun Oh, Joon-Young Lee, Jinwook Choi, Dong-Geol Choi, In So Kweon
We introduce a novel method to automatically adjust camera exposure for image processing and computer vision applications on mobile robot platforms.
no code implementations • CVPR 2017 • Yinda Zhang, Shuran Song, Ersin Yumer, Manolis Savva, Joon-Young Lee, Hailin Jin, Thomas Funkhouser
One of the bottlenecks in training for better representations is the amount of available per-pixel ground truth data that is required for core scene understanding tasks such as semantic segmentation, normal prediction, and object edge detection.
no code implementations • 20 Dec 2016 • Donggeun Yoo, Sunggyun Park, Kyunghyun Paeng, Joon-Young Lee, In So Kweon
In this paper, we present an "action-driven" detection mechanism using our "top-down" visual attention model.
no code implementations • CVPR 2016 • Hae-Gon Jeon, Joon-Young Lee, Sunghoon Im, Hyowon Ha, In So Kweon
Consumer devices with stereo cameras have become popular because of their low-cost depth sensing capability.
no code implementations • 24 Mar 2016 • Youngjin Yoon, Gyeongmin Choe, Namil Kim, Joon-Young Lee, In So Kweon
We present surface normal estimation using a single near infrared (NIR) image.
no code implementations • ICCV 2015 • Hae-Gon Jeon, Joon-Young Lee, Yudeog Han, Seon Joo Kim, In So Kweon
In this paper, we present a novel multi-image motion deblurring method utilizing the coded exposure technique.
no code implementations • CVPR 2016 • Joon-Young Lee, Kalyan Sunkavalli, Zhe Lin, Xiaohui Shen, In So Kweon
We introduce a new technique that automatically generates diverse, visually compelling stylizations for a photograph in an unsupervised manner.
no code implementations • 21 Sep 2015 • Inwook Shim, Seunghak Shin, Yunsu Bok, Kyungdon Joo, Dong-Geol Choi, Joon-Young Lee, Jaesik Park, Jun-Ho Oh, In So Kweon
This paper presents a vision system and a depth processing algorithm for DRC-HUBO+, the winner of the DRC finals 2015.
no code implementations • ICCV 2015 • Donggeun Yoo, Sunggyun Park, Joon-Young Lee, Anthony S. Paek, In So Kweon
We present a novel detection method using a deep convolutional neural network (CNN), named AttentionNet.
no code implementations • CVPR 2015 • Jiyoung Jung, Joon-Young Lee, In So Kweon
We present an outdoor photometric stereo method using images captured in a single day.
no code implementations • 4 Dec 2014 • Donggeun Yoo, Sunggyun Park, Joon-Young Lee, In So Kweon
In this paper, we present a straightforward framework for better image representation by combining the two approaches.
no code implementations • CVPR 2014 • Youngbae Hwang, Joon-Young Lee, In So Kweon, Seon Joo Kim
This paper introduces a new color transfer method which is a process of transferring color of an image to match the color of another image of the same scene.