no code implementations • ECCV 2020 • Woobin Im, Tae-Kyun Kim, Sung-Eui Yoon
Deep unsupervised learning for optical flow has been proposed, where the loss measures image similarity with the warping function parameterized by estimated flow.
no code implementations • ECCV 2020 • Xuepeng Shi, Zhixiang Chen, Tae-Kyun Kim
Monocular 3D object detection plays an important role in autonomous driving and still remains challenging.
no code implementations • CVPR 2024 • Rumeysa Bodur, Binod Bhattarai, Tae-Kyun Kim
Text-guided image editing finds applications in various creative and practical fields.
no code implementations • 10 Oct 2024 • Sohwi Kim, Tae-Kyun Kim
Super-resolution methods are increasingly becoming popular for both real-world and face-specific tasks.
1 code implementation • 27 Sep 2024 • Donghwan Kim, Tae-Kyun Kim
Compared to previous implicit-function-based methods, the point cloud diffusion model can capture the global consistent features to generate the occluded regions, and the denoising process corrects the misaligned SMPL meshes.
no code implementations • 26 Sep 2024 • Luiz Felipe Vecchietti, Minji Lee, Begench Hangeldiyev, Hyunkyu Jung, Hahnbeom Park, Tae-Kyun Kim, Meeyoung Cha, Ho Min Kim
Recent advancements in machine learning (ML) are transforming the field of structural biology.
no code implementations • 26 Sep 2024 • Taeyun Woo, Tae-Kyun Kim, Jinah Park
Estimating the poses of both a hand and an object has become an important area of research due to the growing need for advanced vision computing.
no code implementations • 10 Sep 2024 • Minju Kang, Taehun Kong, Tae-Kyun Kim
We explore a novel teacher-student framework employing channel augmentation for 3D semi-supervised object detection.
no code implementations • 6 Sep 2024 • Woojin Cho, Jihyun Lee, Minjae Yi, Minje Kim, Taeyun Woo, Donghwan Kim, Taewook Ha, Hyokeun Lee, Je-Hwan Ryu, Woontack Woo, Tae-Kyun Kim
Accurate hand and object 3D meshes are obtained by fitting the hand parametric model (MANO) and the hand implicit function (HALO) to multi-view RGBD frames, with the MoCap system only for objects.
no code implementations • CVPR 2024 • Jihyun Lee, Shunsuke Saito, Giljoo Nam, Minhyuk Sung, Tae-Kyun Kim
Sampling from our model yields plausible and diverse two-hand shapes in close interaction with or without an object.
no code implementations • CVPR 2024 • Jinseok Kim, Tae-Kyun Kim
The method consists of a pretrained auto-encoder, a latent diffusion model, and an implicit neural decoder, and their learning strategies.
1 code implementation • CVPR 2024 • Minje Kim, Tae-Kyun Kim
Creating personalized hand avatars is important to offer a realistic experience to users on AR / VR platforms.
no code implementations • 6 Feb 2024 • Jinjing Zhu, Zhedong Hu, Tae-Kyun Kim, Lin Wang
Our framework incorporates two novel components: energy-based feature fusion (EB2F) and energy-based reliable fusion Assessment (RFA) modules.
1 code implementation • 4 Feb 2024 • Tao Wang, Wanglong Lu, Kaihao Zhang, Wenhan Luo, Tae-Kyun Kim, Tong Lu, Hongdong Li, Ming-Hsuan Yang
For the prompt generation, we first propose a prompt pre-training strategy to train a frequency prompt encoder that encodes the ground-truth image into LF and HF prompts.
no code implementations • 19 Jan 2024 • Hao Ai, Zidong Cao, Haonan Lu, Chen Chen, Jian Ma, Pengyuan Zhou, Tae-Kyun Kim, Pan Hui, Lin Wang
To this end, we propose a transformer-based 360 image outpainting framework called Dream360, which can generate diverse, high-fidelity, and high-resolution panoramas from user-selected viewports, considering the spherical properties of 360 images.
no code implementations • 9 Sep 2023 • Xuanxi Chen, Tao Wang, Ziqian Shao, Kaihao Zhang, Wenhan Luo, Tong Lu, Zikun Liu, Tae-Kyun Kim, Hongdong Li
With the pipeline, we build the first large-scale UDC video restoration dataset called PexelsUDC, which includes two subsets named PexelsUDC-T and PexelsUDC-P corresponding to different displays for UDC.
1 code implementation • 27 Jul 2023 • Tao Wang, Kaihao Zhang, Ziqian Shao, Wenhan Luo, Bjorn Stenger, Tae-Kyun Kim, Wei Liu, Hongdong Li
In this paper, we address this limitation by proposing a degradation-aware learning scheme for LLIE using diffusion models, which effectively integrates degradation and image priors into the diffusion process, resulting in improved image enhancement.
1 code implementation • 29 May 2023 • Tao Wang, Kaihao Zhang, Ziqian Shao, Wenhan Luo, Bjorn Stenger, Tong Lu, Tae-Kyun Kim, Wei Liu, Hongdong Li
Second, we introduce a residual dense transformer block (RDTB) as the final GridFormer layer.
no code implementations • 10 May 2023 • Rumeysa Bodur, Erhan Gundogdu, Binod Bhattarai, Tae-Kyun Kim, Michael Donoser, Loris Bazzani
We propose a novel learning method for text-guided image editing, namely \texttt{iEdit}, that generates images conditioned on a source image and a textual edit prompt.
1 code implementation • ICCV 2023 • Jiaze Sun, Zhixiang Chen, Tae-Kyun Kim
Unsupervised methods have been proposed for graph convolutional models but they require ground truth correspondence between the source and target inputs.
no code implementations • 3 Apr 2023 • Pedro Castro, Tae-Kyun Kim
However, these methods are often inefficient and limited by their reliance on pre-trained models that have not be designed specifically for pose estimation.
1 code implementation • CVPR 2023 • Junbong Jang, Kwonmoo Lee, Tae-Kyun Kim
For quantitative evaluation, we labeled sparse tracking points along the contour of live cells from two live cell datasets taken with phase contrast and confocal fluorescence microscopes.
1 code implementation • CVPR 2023 • Jihyun Lee, Minhyuk Sung, Honggyu Choi, Tae-Kyun Kim
To handle the shape complexity and interaction context between two hands, Im2Hands models the occupancy volume of two hands - conditioned on an RGB image and coarse 3D keypoints - by two novel attention-based modules responsible for (1) initial occupancy estimation and (2) context-aware occupancy refinement, respectively.
1 code implementation • 4 Jan 2023 • Razvan Caramalau, Binod Bhattarai, Danail Stoyanov, Tae-Kyun Kim
We present MoBYv2AL, a novel self-supervised active learning framework for image classification.
no code implementations • ICCV 2023 • Xuepeng Shi, Georgi Dikov, Gerhard Reitmayr, Tae-Kyun Kim, Mohsen Ghafoorian
Self-supervised monocular depth estimation (SSMDE) aims at predicting the dense depth maps of monocular images, by learning to minimize a photometric loss using spatially neighboring image pairs during training.
no code implementations • 6 Dec 2022 • Honggyu Choi, Zhixiang Chen, Xuepeng Shi, Tae-Kyun Kim
Unlike existing suboptimal methods, we propose a two-step pseudo-label filtering for the classification and regression heads in a teacher-student framework.
Ranked #14 on Semi-Supervised Object Detection on COCO 1% labeled data
1 code implementation • 21 Oct 2022 • Pedro Castro, Tae-Kyun Kim
Learning based 6D object pose estimation methods rely on computing large intermediate pose representations and/or iteratively refining an initial estimation with a slow render-compare pipeline.
no code implementations • CVPR 2022 • Jihyun Lee, Minhyuk Sung, HyunJin Kim, Tae-Kyun Kim
We propose a framework that can deform an object in a 2D image as it exists in 3D space.
1 code implementation • 28 Mar 2022 • Dafni Antotsiou, Carlo Ciliberto, Tae-Kyun Kim
This is done by using proto-policies as modules to divide the tasks into simple sub-behaviours that can be shared.
no code implementations • 8 Dec 2021 • Rumeysa Bodur, Binod Bhattarai, Tae-Kyun Kim
Recently, hierarchical networks that consist of both a global network dealing with the whole image and multiple local networks focusing on local parts are showing success.
1 code implementation • CVPR 2021 • Lin Wang, Yujeong Chae, Sung-Hoon Yoon, Tae-Kyun Kim, Kuk-Jin Yoon
To enable KD across the unpaired modalities, we first propose a bidirectional modality reconstruction (BMR) module to bridge both modalities and simultaneously exploit them to distill knowledge via the crafted pairs, causing no extra computation in the inference.
Ranked #7 on Event-based Object Segmentation on MVSEC-SEG
no code implementations • 17 Nov 2021 • Jiaze Sun, Binod Bhattarai, Zhixiang Chen, Tae-Kyun Kim
Whilst both branches are required during training, the RGB branch is our primary network and the semantic branch is not needed for inference.
1 code implementation • 7 Jun 2021 • Razvan Caramalau, Binod Bhattarai, Tae-Kyun Kim
In this paper, we present a novel pipeline for pool-based Active Learning.
1 code implementation • 31 May 2021 • Jianhong Wang, Yuan Zhang, Yunjie Gu, Tae-Kyun Kim
This paper studies a theoretical framework for value factorisation with interpretability via Shapley value theory.
no code implementations • 12 May 2021 • Suman Sapkota, Bidur Khanal, Binod Bhattarai, Bishesh Khanal, Tae-Kyun Kim
Multi-domain image-to-image translation with conditional Generative Adversarial Networks (GANs) can generate highly photo realistic images with desired target classes, yet these synthetic images have not always been helpful to improve downstream supervised tasks such as image classification.
1 code implementation • CVPR 2021 • Zhixiang Chen, Tae-Kyun Kim
3D morphable models are widely used for the shape representation of an object class in computer vision and graphics applications.
1 code implementation • ICCV 2021 • Xuepeng Shi, Qi Ye, Xiaozhi Chen, Chuangrong Chen, Zhixiang Chen, Tae-Kyun Kim
The experimental results show that our method achieves the state-of-the-art performance on the monocular 3D Object Detection and Birds Eye View tasks of the KITTI dataset, and can generalize to images with different camera intrinsics.
Ranked #16 on Monocular 3D Object Detection on KITTI Cars Moderate
1 code implementation • 25 Mar 2021 • Dafni Antotsiou, Carlo Ciliberto, Tae-Kyun Kim
Deep Imitation Learning requires a large number of expert demonstrations, which are not always easy to obtain, especially for complex tasks.
1 code implementation • 1 Oct 2020 • Razvan Caramalau, Binod Bhattarai, Tae-Kyun Kim
We propose a Bayesian approximation to a deep learning architecture for 3D hand pose estimation.
no code implementations • 30 Sep 2020 • Rumeysa Bodur, Binod Bhattarai, Tae-Kyun Kim
We utilise this dataset to minimise the novel depth consistency loss via adversarial learning (note we do not have ground truth depth maps for generated face images) and the depth categorical loss of synthetic data on the discriminator.
no code implementations • 7 Aug 2020 • Guillermo Garcia-Hernando, Edward Johns, Tae-Kyun Kim
Dexterous manipulation of objects in virtual environments with our bare hands, by using only a depth sensor and a state-of-the-art 3D hand pose estimator (HPE), is challenging.
1 code implementation • CVPR 2021 • Razvan Caramalau, Binod Bhattarai, Tae-Kyun Kim
We flip the label of newly queried nodes from unlabelled to labelled, re-train the learner to optimise the downstream task and the graph to minimise its modified objective.
Ranked #4 on Active Learning on CIFAR10 (10,000)
no code implementations • 11 Jun 2020 • Jiaze Sun, Binod Bhattarai, Tae-Kyun Kim
We perform augmentation by randomly sampling sensible labels from the label space of the few labelled examples available and assigning them as target labels to the abundant unlabelled examples from the same distribution as that of the labelled ones.
2 code implementations • ICLR 2021 • Jianhong Wang, Yuan Zhang, Tae-Kyun Kim, Yunjie Gu
We test HDNO on MultiWoz 2. 0 and MultiWoz 2. 1, the datasets on multi-domain dialogues, in comparison with word-level E2E model trained with RL, LaRL and HDSA, showing improvements on the performance evaluated by automatic evaluation metrics and human evaluation.
Hierarchical Reinforcement Learning reinforcement-learning +2
no code implementations • ECCV 2020 • Anil Armagan, Guillermo Garcia-Hernando, Seungryul Baek, Shreyas Hampali, Mahdi Rad, Zhaohui Zhang, Shipeng Xie, Mingxiu Chen, Boshen Zhang, Fu Xiong, Yang Xiao, Zhiguo Cao, Junsong Yuan, Pengfei Ren, Weiting Huang, Haifeng Sun, Marek Hrúz, Jakub Kanis, Zdeněk Krňoul, Qingfu Wan, Shile Li, Linlin Yang, Dongheui Lee, Angela Yao, Weiguo Zhou, Sijia Mei, Yun-hui Liu, Adrian Spurr, Umar Iqbal, Pavlo Molchanov, Philippe Weinzaepfel, Romain Brégier, Grégory Rogez, Vincent Lepetit, Tae-Kyun Kim
To address these issues, we designed a public challenge (HANDS'19) to evaluate the abilities of current 3D hand pose estimators (HPEs) to interpolate and extrapolate the poses of a training set.
no code implementations • 27 Mar 2020 • Juil Sock, Guillermo Garcia-Hernando, Anil Armagan, Tae-Kyun Kim
Most successful approaches to estimate the 6D pose of an object typically train a neural network by supervising the learning with annotated poses in real world images.
no code implementations • 23 Mar 2020 • Sharib Ali, Binod Bhattarai, Tae-Kyun Kim, Jens Rittscher
In this work, we propose to use a few-shot learning approach that requires less training data and can be used to predict label classes of test samples from an unseen dataset.
1 code implementation • CVPR 2020 • Lin Wang, Tae-Kyun Kim, Kuk-Jin Yoon
While each phase is mainly for one of the three tasks, the networks in earlier phases are fine-tuned by respective loss functions in an end-to-end manner.
no code implementations • ECCV 2020 • Binod Bhattarai, Tae-Kyun Kim
Existing conditional GANs commonly encode target domain label information as hard-coded categorical vectors in the form of 0s and 1s.
no code implementations • 28 Jan 2020 • Caner Sahin, Guillermo Garcia-Hernando, Juil Sock, Tae-Kyun Kim
In this paper, we present the first comprehensive and most recent review of the methods on object pose recovery, from 3D bounding box detectors to full 6D pose estimators.
no code implementations • 23 Oct 2019 • Pedro Castro, Anil Armagan, Tae-Kyun Kim
Current 6D object pose methods consist of deep CNN models fully optimized for a single object but with its architecture standardized among objects with different shapes.
no code implementations • 19 Oct 2019 • Juil Sock, Guillermo Garcia-Hernando, Tae-Kyun Kim
In this work, we explore how a strategic selection of camera movements can facilitate the task of 6D multi-object pose estimation in cluttered scenarios while respecting real-world constraints important in robotics and augmented reality applications, such as time and distance traveled.
no code implementations • 10 Sep 2019 • Binod Bhattarai, Seungryul Baek, Rumeysa Bodur, Tae-Kyun Kim
Unlike previous studies of randomly augmenting the synthetic data with real data, we present our simple, effective and easy to implement synthetic data sampling methods to train deep CNNs more efficiently and accurately.
no code implementations • 22 Jul 2019 • Mang Shao, Danhang Tang, Tae-Kyun Kim
In this work, we present a modified fuzzy decision forest for real-time 3D object pose estimation based on typical template representation.
no code implementations • 15 Jul 2019 • Binod Bhattarai, Rumeysa Bodur, Tae-Kyun Kim
Augmenting data in image space (eg.
2 code implementations • 11 Jul 2019 • Jianhong Wang, Yuan Zhang, Tae-Kyun Kim, Yunjie Gu
To deal with this problem, we i) introduce a cooperative-game theoretical framework called extended convex game (ECG) that is a superset of global reward game, and ii) propose a local reward approach called Shapley Q-value.
Multi-agent Reinforcement Learning Policy Gradient Methods +1
no code implementations • CVPR 2019 • Seungryul Baek, Kwang In Kim, Tae-Kyun Kim
Once the model is successfully fitted to input RGB images, its meshes i. e. shapes and articulations, are realistic, and we augment view-points on top of estimated dense hand poses.
no code implementations • 11 Mar 2019 • Caner Sahin, Guillermo Garcia-Hernando, Juil Sock, Tae-Kyun Kim
6D object pose estimation is an important task that determines the 3D position and 3D rotation of an object in camera-centred coordinates.
no code implementations • 18 Nov 2018 • Shanxin Yuan, Bjorn Stenger, Tae-Kyun Kim
We explore different ways of using this privileged information: (1) using depth data to initially train a depth-based network, (2) using the features from the depth-based network of the paired depth images to constrain mid-level RGB network weights, and (3) using the foreground mask, obtained from the depth data, to suppress the responses from the background area.
no code implementations • 25 Oct 2018 • Iason Oikonomidis, Guillermo Garcia-Hernando, Angela Yao, Antonis Argyros, Vincent Lepetit, Tae-Kyun Kim
The fourth instantiation of this workshop attracted significant interest from both academia and the industry.
no code implementations • 9 Oct 2018 • Tomas Hodan, Rigas Kouskouridas, Tae-Kyun Kim, Federico Tombari, Kostas Bekris, Bertram Drost, Thibault Groueix, Krzysztof Walas, Vincent Lepetit, Ales Leonardis, Carsten Steger, Frank Michel, Caner Sahin, Carsten Rother, Jiri Matas
The workshop featured four invited talks, oral and poster presentations of accepted workshop papers, and an introduction of the BOP benchmark for 6D object pose estimation.
1 code implementation • 3 Oct 2018 • Dafni Antotsiou, Guillermo Garcia-Hernando, Tae-Kyun Kim
In this work, we capture the hand information by using a state-of-the-art hand pose estimator.
1 code implementation • ECCV 2018 • Tomas Hodan, Frank Michel, Eric Brachmann, Wadim Kehl, Anders Glent Buch, Dirk Kraft, Bertram Drost, Joel Vidal, Stephan Ihrke, Xenophon Zabulis, Caner Sahin, Fabian Manhardt, Federico Tombari, Tae-Kyun Kim, Jiri Matas, Carsten Rother
We propose a benchmark for 6D pose estimation of a rigid object from a single RGB-D input image.
no code implementations • IEEE Access 2018 • Xinghao Chen, Guijin Wang, Cairong Zhang, Tae-Kyun Kim, Xiangyang Ji
The semantic segmentation network assigns semantic labels for each point in the point set.
Ranked #7 on Hand Pose Estimation on MSRA Hands
no code implementations • 1 Aug 2018 • Caner Sahin, Tae-Kyun Kim
Intra-class variations, distribution shifts among source and target domains are the major challenges of category-level tasks.
no code implementations • 11 Jun 2018 • Juil Sock, Kwang In Kim, Caner Sahin, Tae-Kyun Kim
Our architecture jointly learns multiple sub-tasks: 2D detection, depth, and 3D pose estimation of individual objects; and joint registration of multiple objects.
no code implementations • CVPR 2018 • Seungryul Baek, Kwang In Kim, Tae-Kyun Kim
By training the HPG and HPE in a single unified optimization framework enforcing that 1) the HPE agrees with the paired depth and skeleton entries; and 2) the HPG-HPE combination satisfies the cyclic consistency (both the input and the output of HPG-HPE are skeletons) observed via the newly generated unpaired skeletons, our algorithm constructs a HPE which is robust to variations that go beyond the coverage of the existing database.
1 code implementation • ECCV 2018 • Baris Gecer, Binod Bhattarai, Josef Kittler, Tae-Kyun Kim
We propose a novel end-to-end semi-supervised adversarial framework to generate photorealistic face images of new identities with wide ranges of expressions, poses, and illuminations conditioned by a 3D morphable model.
Ranked #16 on Face Verification on IJB-A
1 code implementation • CVPR 2018 • Shanxin Yuan, Guillermo Garcia-Hernando, Bjorn Stenger, Gyeongsik Moon, Ju Yong Chang, Kyoung Mu Lee, Pavlo Molchanov, Jan Kautz, Sina Honari, Liuhao Ge, Junsong Yuan, Xinghao Chen, Guijin Wang, Fan Yang, Kai Akiyama, Yang Wu, Qingfu Wan, Meysam Madadi, Sergio Escalera, Shile Li, Dongheui Lee, Iason Oikonomidis, Antonis Argyros, Tae-Kyun Kim
Official Torch7 implementation of "V2V-PoseNet: Voxel-to-Voxel Prediction Network for Accurate 3D Hand and Human Pose Estimation from a Single Depth Map", CVPR 2018
Ranked #5 on Hand Pose Estimation on HANDS 2017
no code implementations • ECCV 2018 • Qi Ye, Tae-Kyun Kim
The proposed method leverages the state-of-the-art hand pose estimators based on Convolutional Neural Networks to facilitate feature learning, while it models the multiple modes in a two-level hierarchy to reconcile single-valued and multi-valued mapping in its output.
no code implementations • ICCV 2017 • Vassileios Balntas, Andreas Doumanoglou, Caner Sahin, Juil Sock, Rigas Kouskouridas, Tae-Kyun Kim
In this paper we examine the effects of using object poses as guidance to learning robust features for 3D object pose estimation.
no code implementations • 1 Aug 2017 • Baris Gecer, Vassileios Balntas, Tae-Kyun Kim
In this work, we investigate several methods and strategies to learn deep embeddings for face recognition, using joint sample- and set-based optimization.
no code implementations • 7 Jul 2017 • Shanxin Yuan, Qi Ye, Guillermo Garcia-Hernando, Tae-Kyun Kim
We present the 2017 Hands in the Million Challenge, a public competition designed for the evaluation of the task of 3D hand pose estimation.
no code implementations • 10 Jun 2017 • Caner Sahin, Tae-Kyun Kim
A large number of studies analyse object detection and pose estimation at visual level in 2D, discussing the effects of challenges such as occlusion, clutter, texture, etc., on the performances of the methods, which work in the context of RGB modality.
no code implementations • 6 Jun 2017 • Seungryul Baek, Kwang In Kim, Tae-Kyun Kim
Each response map-or node-in both the convolutional and fully-connected layers selectively respond to class labels s. t.
Ranked #175 on Image Classification on CIFAR-100 (using extra training data)
no code implementations • CVPR 2017 • Shanxin Yuan, Qi Ye, Bjorn Stenger, Siddhant Jain, Tae-Kyun Kim
We also show significant improvements in egocentric hand pose estimation with a CNN trained on the new dataset.
1 code implementation • CVPR 2018 • Guillermo Garcia-Hernando, Shanxin Yuan, Seungryul Baek, Tae-Kyun Kim
Our dataset and experiments can be of interest to communities of 3D hand pose estimation, 6D object pose, and robotics as well as action recognition.
no code implementations • CVPR 2017 • Zhiyuan Shi, Tae-Kyun Kim
Our PI-based classification loss maintains a consistency between latent PI and predicted distribution.
no code implementations • 9 Jan 2017 • Caner Sahin, Rigas Kouskouridas, Tae-Kyun Kim
The iterative refinement is accomplished based on finer (smaller) parts that are represented with more discriminative control point descriptors by using our Iterative Hough Forest.
no code implementations • 28 Oct 2016 • Seungryul Baek, Kwang In Kim, Tae-Kyun Kim
Online action detection (OAD) is challenging since 1) robust yet computationally expensive features cannot be straightforwardly used due to the real-time processing requirements and 2) the localization and classification of actions have to be performed even before they are fully observed.
no code implementations • 23 Jul 2016 • Seungryul Baek, Zhiyuan Shi, Masato Kawade, Tae-Kyun Kim
In this paper, we tackle the problem of 24 hours-monitoring patient actions in a ward such as "stretching an arm out of the bed", "falling out of the bed", where temporal movements are subtle or significant.
no code implementations • CVPR 2017 • Guillermo Garcia-Hernando, Tae-Kyun Kim
A human action can be seen as transitions between one's body poses over time, where the transition depicts a temporal relation between two poses.
no code implementations • 8 Jul 2016 • Andreas Doumanoglou, Vassileios Balntas, Rigas Kouskouridas, Tae-Kyun Kim
Furthermore, we argue that our pose-guided feature learning using our Siamese Regression Network generates more discriminative features that outperform the state of the art.
no code implementations • 31 May 2016 • Yang Liu, Minh Hoai, Mang Shao, Tae-Kyun Kim
LBSVM is based on Structured-Output SVM, but extends it to handle noisy video data and ensure consistency of the output decision throughout time.
no code implementations • 29 Apr 2016 • Chourmouzios Tsiotsios, Maria E. Angelopoulou, Andrew J. Davison, Tae-Kyun Kim
Backscatter corresponds to a complex term with several unknown variables, and makes the problem of normal estimation hard.
1 code implementation • 12 Apr 2016 • Qi Ye, Shanxin Yuan, Tae-Kyun Kim
In this paper, a hybrid hand pose estimation method is proposed by applying the kinematic hierarchy strategy to the input space (as well as the output space) of the discriminative method by a spatial attention mechanism and to the optimization of the generative method by hierarchical Particle Swarm Optimization (PSO).
no code implementations • 8 Mar 2016 • Caner Sahin, Rigas Kouskouridas, Tae-Kyun Kim
State-of-the-art techniques proposed for 6D object pose recovery depend on occlusion-free point clouds to accurately register objects in 3D space.
no code implementations • 3 Feb 2016 • Rigas Kouskouridas, Alykhan Tejani, Andreas Doumanoglou, Danhang Tang, Tae-Kyun Kim
In this paper we present Latent-Class Hough Forests, a method for object detection and 6 DoF pose estimation in heavily cluttered and occluded scenarios.
no code implementations • CVPR 2016 • Andreas Doumanoglou, Rigas Kouskouridas, Sotiris Malassiotis, Tae-Kyun Kim
In this work, we present a complete framework for both single shot-based 6D object pose estimation and next-best-view prediction based on Hough Forests, the state of the art object pose estimator that performs classification and regression jointly.
no code implementations • ICCV 2015 • Chao Xiong, Xiaowei Zhao, Danhang Tang, Karlekar Jayashree, Shuicheng Yan, Tae-Kyun Kim
Faces in the wild are usually captured with various poses, illuminations and occlusions, and thus inherently multimodally distributed in many tasks.
no code implementations • ICCV 2015 • Danhang Tang, Jonathan Taylor, Pushmeet Kohli, Cem Keskin, Tae-Kyun Kim, Jamie Shotton
In this paper, we show that we can significantly improving upon black box optimization by exploiting high-level knowledge of the structure of the parameters and using a local surrogate energy function.
no code implementations • 14 Oct 2014 • Wai Lam Hoo, Tae-Kyun Kim, Yuru Pei, Chee Seng Chan
Image understanding is an important research domain in the computer vision due to its wide real-world applications.
no code implementations • 26 Sep 2014 • Wenhan Luo, Junliang Xing, Anton Milan, Xiaoqin Zhang, Wei Liu, Tae-Kyun Kim
We inspect the recent advances in various aspects and propose some interesting directions for future research.
no code implementations • CVPR 2014 • Wenhan Luo, Tae-Kyun Kim, Bjorn Stenger, Xiaowei Zhao, Roberto Cipolla
In this paper, we propose a label propagation framework to handle the multiple object tracking (MOT) problem for a generic object type (cf.
no code implementations • CVPR 2014 • Danhang Tang, Hyung Jin Chang, Alykhan Tejani, Tae-Kyun Kim
In contrast to prior forest-based methods, which take dense pixels as input, classify them independently and then estimate joint positions afterwards; our method can be considered as a structured coarse-to-fine search, starting from the centre of mass of a point cloud until locating all the skeletal joints.
no code implementations • CVPR 2014 • Xiaowei Zhao, Tae-Kyun Kim, Wenhan Luo
In this paper, we present a unified method for joint face image analysis, i. e., simultaneously estimating head pose, facial expression and landmark positions in real-world face images.
no code implementations • CVPR 2014 • Chourmouzios Tsiotsios, Maria E. Angelopoulou, Tae-Kyun Kim, Andrew J. Davison
We compare our method with previous approaches through extensive experimental results, where a variety of objects are imaged in a big water tank whose turbidity is systematically increased, and show reconstruction quality which degrades little relative to clean water results even with a very significant scattering level.
no code implementations • CVPR 2013 • Tsz-Ho Yu, Tae-Kyun Kim, Roberto Cipolla
This work addresses the challenging problem of unconstrained 3D human pose estimation (HPE) from a novel perspective.
no code implementations • NeurIPS 2008 • Tae-Kyun Kim, Roberto Cipolla
We present a new co-clustering problem of images and visual features.