no code implementations • 13 Mar 2024 • Yi Zhou, HUI ZHANG, Jiaqian Yu, Yifan Yang, Sangil Jung, Seung-In Park, ByungIn Yoo
Concretely, we introduce a hybrid representation called HIQuery to represent all map elements, and propose a point-element interactor to interactively extract and encode the hybrid information of elements, e. g. point position and element shape, into the HIQuery.
no code implementations • 21 Aug 2023 • Dongwook Lee, Wonjun Choi, Seohyung Lee, ByungIn Yoo, Eunho Yang, Seongju Hwang
An effective method to tackle these challenges is template update, which updates the template to reflect the change of appearance in the target object during tracking.
no code implementations • 13 Mar 2023 • Hyeongseok Son, Sangil Jung, Solae Lee, Seongeun Kim, Seung-In Park, ByungIn Yoo
Human is one of the most essential classes in visual recognition tasks such as detection, segmentation, and pose estimation.
no code implementations • CVPR 2022 • Yi Zhou, HUI ZHANG, Hana Lee, Shuyang Sun, Pingjun Li, Yangguang Zhu, ByungIn Yoo, Xiaojuan Qi, Jae-Joon Han
We encode all panoptic entities in a video, including both foreground instances and background semantics, with a unified representation called panoptic slots.
no code implementations • CVPR 2021 • Kinam Kwon, Eunhee Kang, Sangwon Lee, Su-Jin Lee, Hyong-Euk Lee, ByungIn Yoo, Jae-Joon Han
However, this causes inevitable image degradation in the form of spatially variant blur and noise because of the opaque display in front of the camera.
no code implementations • CVPR 2021 • Jaehyoung Yoo, Dongwook Lee, Changyong Son, Sangil Jung, ByungIn Yoo, Changkyu Choi, Jae-Joon Han, Bohyung Han
RaScaNet reads only a few rows of pixels at a time using a convolutional neural network and then sequentially learns the representation of the whole image using a recurrent neural network.
1 code implementation • ICCV 2021 • Dongyoung Kim, Jinwoo Kim, Seonghyeon Nam, Dongwoo Lee, Yeonkyung Lee, Nahyup Kang, Hyong-Euk Lee, ByungIn Yoo, Jae-Joon Han, Seon Joo Kim
Images in our dataset are mostly captured with illuminants existing in the scene, and the ground truth illumination is computed by taking the difference between the images with different illumination combination.
no code implementations • 15 Dec 2019 • ByungIn Yoo, Tristan Sylvain, Yoshua Bengio, Junmo Kim
In this paper, we propose a Generative Translation Classification Network (GTCN) for improving visual classification accuracy in settings where classes are visually similar and data is scarce.
no code implementations • 21 Mar 2017 • Youngsung Kim, ByungIn Yoo, Youngjun Kwak, Changkyu Choi, Junmo Kim
In this paper, we propose to utilize contrastive representation that embeds a distinctive expressive factor for a discriminative purpose.
Facial Expression Recognition Facial Expression Recognition (FER)
no code implementations • CVPR 2015 • Junho Yim, Heechul Jung, ByungIn Yoo, Changkyu Choi, Dusik Park, Junmo Kim
This paper proposes a new deep architecture based on a novel type of multitask learning, which can achieve superior performance in rotating to a target-pose face image from an arbitrary pose and illumination image while preserving identity.