1 code implementation • 2 Jun 2023 • Lei Ke, Mingqiao Ye, Martin Danelljan, Yifan Liu, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu
HQ-SAM is only trained on the introduced detaset of 44k masks, which takes only 4 hours on 8 GPUs.
no code implementations • 28 May 2023 • Yue Xu, Yong-Lu Li, Kaitong Cui, Ziyu Wang, Cewu Lu, Yu-Wing Tai, Chi-Keung Tang
The new strategy significantly reduces the training cost and extends a variety of existing distillation algorithms to larger and more diversified datasets, e. g. in some cases only 0. 04% training data is sufficient for comparable distillation performance.
no code implementations • 24 May 2023 • Xinhang Liu, Shiu-hong Kao, Jiaben Chen, Yu-Wing Tai, Chi-Keung Tang
This paper introduces Deceptive-NeRF, a new method for enhancing the quality of reconstructed NeRF models using synthetically generated pseudo-observations, capable of handling sparse input and removing floater artifacts.
no code implementations • 22 May 2023 • Han Jiang, Ruoxuan Li, Haosen Sun, Yu-Wing Tai, Chi-Keung Tang
No significant work has been done to directly merge two partially overlapping scenes using NeRF representations.
no code implementations • 10 Apr 2023 • Benran Hu, Junkai Huang, Yichen Liu, Yu-Wing Tai, Chi-Keung Tang
This paper presents one of the first learning-based NeRF 3D instance segmentation pipelines, dubbed as Instance Neural Radiance Field, or Instance NeRF.
1 code implementation • CVPR 2023 • Lei Ke, Martin Danelljan, Henghui Ding, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu
A consistency loss is then enforced on the found matches.
no code implementations • 26 Mar 2023 • Xinhang Liu, Yu-Wing Tai, Chi-Keung Tang
This paper analyzes the NeRF's struggles in such settings and proposes Clean-NeRF for accurate 3D reconstruction and novel view rendering in complex scenes.
no code implementations • CVPR 2023 • Yingwei Wang, Xu Jia, Xin Tao, Takashi Isobe, Huchuan Lu, Yu-Wing Tai
Videos stored on mobile devices or delivered on the Internet are usually in compressed format and are of various unknown compression parameters, but most video super-resolution (VSR) methods often assume ideal inputs resulting in large performance gap between experimental settings and real-world applications.
1 code implementation • CVPR 2023 • Yanan sun, Chi-Keung Tang, Yu-Wing Tai
Instead, our method resorts to spatial and temporal sparsity for solving general UHR matting.
no code implementations • 22 Nov 2022 • Shengnan Liang, Yichen Liu, Shangzhe Wu, Yu-Wing Tai, Chi-Keung Tang
We present ONeRF, a method that automatically segments and reconstructs object instances in 3D from multi-view RGB images without any additional manual annotations.
1 code implementation • CVPR 2023 • Benran Hu, Junkai Huang, Yichen Liu, Yu-Wing Tai, Chi-Keung Tang
This paper presents the first significant object detection framework, NeRF-RPN, which directly operates on NeRF.
1 code implementation • 21 Nov 2022 • Hao Zhang, Tianyuan Dai, Yu-Wing Tai, Chi-Keung Tang
This paper presents the first significant work on directly predicting 3D face landmarks on neural radiance fields (NeRFs), without using intermediate representations such as 2D images, depth maps, or point clouds.
no code implementations • 21 Nov 2022 • Changlin Li, Guangyang Wu, Yanan sun, Xin Tao, Chi-Keung Tang, Yu-Wing Tai
The learnt deformable kernel is then utilized in convolving the input frames for predicting the interpolated frame.
no code implementations • 8 Nov 2022 • Qi Fan, Mattia Segu, Yu-Wing Tai, Fisher Yu, Chi-Keung Tang, Bernt Schiele, Dengxin Dai
Thus, we propose to perturb the channel statistics of source domain features to synthesize various latent styles, so that the trained deep model can perceive diverse potential domains and generalizes well even without observations of target domain data in training.
no code implementations • 4 Nov 2022 • Kepeng Xu, Li Xu, Gang He, Chang Wu, Zijia Ma, Ming Sun, Yu-Wing Tai
To evaluate the performance of the proposed method, we construct a corresponding multi-frame dataset using HDR video of the HDR10 standard to conduct a comprehensive evaluation of different methods.
no code implementations • 13 Oct 2022 • Rui Qin, Bin Wang, Yu-Wing Tai
The CP Loss supervises the text reconstruction with content semantics by multi-scale text recognition features, which effectively incorporates content awareness into the framework.
no code implementations • 2 Oct 2022 • Xinhang Liu, Jiaben Chen, Huai Yu, Yu-Wing Tai, Chi-Keung Tang
The core of our method is a novel propagation strategy for individual objects' radiance fields with a bidirectional photometric loss, enabling an unsupervised partitioning of a scene into salient or meaningful regions corresponding to different object instances.
no code implementations • 28 Sep 2022 • Jiayin Cai, Changlin Li, Xin Tao, Chun Yuan, Yu-Wing Tai
This paper proposes a novel video inpainting method.
1 code implementation • 8 Aug 2022 • Lei Ke, Yu-Wing Tai, Chi-Keung Tang
Unlike previous instance segmentation methods, we model image formation as a composition of two overlapping layers, and propose Bilayer Convolutional Network (BCNet), where the top layer detects occluding objects (occluders) and the bottom layer infers partially occluded instances (occludees).
1 code implementation • 28 Jul 2022 • Lei Ke, Henghui Ding, Martin Danelljan, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu
While Video Instance Segmentation (VIS) has seen rapid progress, current approaches struggle to predict high-quality masks with accurate boundary details.
Ranked #1 on
Video Instance Segmentation
on HQ-YTVIS
1 code implementation • 23 Jul 2022 • Qi Fan, Wenjie Pei, Yu-Wing Tai, Chi-Keung Tang
Motivated by the simple Gestalt principle that pixels belonging to the same object are more similar than those to different objects of same class, we propose a novel self-support matching strategy to alleviate this problem, which uses query prototypes to match query features, where the query prototypes are collected from high-confidence query predictions.
Ranked #9 on
Few-Shot Semantic Segmentation
on PASCAL-5i (5-Shot)
1 code implementation • 20 Jul 2022 • Wenjie Pei, Xin Feng, Canmiao Fu, Qiong Cao, Guangming Lu, Yu-Wing Tai
The key challenge of sequence representation learning is to capture the long-range temporal dependencies.
1 code implementation • 30 May 2022 • Peng Zheng, Huazhu Fu, Deng-Ping Fan, Qi Fan, Jie Qin, Yu-Wing Tai, Chi-Keung Tang, Luc van Gool
In this paper, we present a novel end-to-end group collaborative learning network, termed GCoNet+, which can effectively and efficiently (250 fps) identify co-salient objects in natural scenes.
Ranked #1 on
Co-Salient Object Detection
on CoSal2015
1 code implementation • CVPR 2022 • Yanan sun, Chi-Keung Tang, Yu-Wing Tai
A new instance matting metric called instance matting quality (IMQ) is proposed, which addresses the absence of a unified and fair means of evaluation emphasizing both instance recognition and matting quality.
1 code implementation • CVPR 2022 • Xinpeng Liu, Yong-Lu Li, Xiaoqian Wu, Yu-Wing Tai, Cewu Lu, Chi-Keung Tang
Human-Object Interaction (HOI) detection plays a core role in activity understanding.
1 code implementation • CVPR 2022 • Takashi Isobe, Xu Jia, Xin Tao, Changlin Li, Ruihuang Li, Yongjie Shi, Jing Mu, Huchuan Lu, Yu-Wing Tai
Instead of directly feeding consecutive frames into a VSR model, we propose to compute the temporal difference between frames and divide those pixels into two subsets according to the level of difference.
no code implementations • 15 Feb 2022 • Mu-Ruei Tseng, Abhishek Gupta, Chi-Keung Tang, Yu-Wing Tai
All training and testing 3D skeletons in HAA4D are globally aligned, using a deep alignment model to the same global space, making each skeleton face the negative z-direction.
1 code implementation • 15 Dec 2021 • Li Xu, Gang He, Jinjia Zhou, Jie Lei, Weiying Xie, Yunsong Li, Yu-Wing Tai
In most video platforms, such as Youtube, and TikTok, the played videos usually have undergone multiple video encodings such as hardware encoding by recording devices, software encoding by video editing apps, and single/multiple video transcoding by video application servers.
1 code implementation • 3 Dec 2021 • Chen Wang, Xian Wu, Yuan-Chen Guo, Song-Hai Zhang, Yu-Wing Tai, Shi-Min Hu
We present NeRF-SR, a solution for high-resolution (HR) novel view synthesis with mostly low-resolution (LR) inputs.
1 code implementation • CVPR 2022 • Lei Ke, Martin Danelljan, Xia Li, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu
Instead of operating on regular dense tensors, our Mask Transfiner decomposes and represents the image regions as a quadtree.
Ranked #1 on
Instance Segmentation
on BDD100K val
no code implementations • ICCV 2021 • Lei Ke, Yu-Wing Tai, Chi-Keung Tang
To facilitate this new research, we construct the first large-scale video object inpainting benchmark YouTube-VOI to provide realistic occlusion scenarios with both occluded and visible object masks available.
1 code implementation • NeurIPS 2021 • Lei Ke, Xia Li, Martin Danelljan, Yu-Wing Tai, Chi-Keung Tang, Fisher Yu
We propose Prototypical Cross-Attention Network (PCAN), capable of leveraging rich spatio-temporal information for online multiple object tracking and segmentation.
Ranked #1 on
Video Instance Segmentation
on BDD100K val
Multi-Object Tracking and Segmentation
Multiple Object Track and Segmentation
+1
3 code implementations • NeurIPS 2021 • Ho Kei Cheng, Yu-Wing Tai, Chi-Keung Tang
This paper presents a simple yet effective approach to modeling space-time correspondences in the context of video object segmentation.
Ranked #5 on
Semi-Supervised Video Object Segmentation
on MOSE
Semantic Segmentation
Semi-Supervised Video Object Segmentation
+1
1 code implementation • 30 Apr 2021 • Qi Fan, Chi-Keung Tang, Yu-Wing Tai
We introduce Few-Shot Video Object Detection (FSVOD) with three contributions to real-world visual learning challenge in our highly diverse and dynamic world: 1) a large-scale video dataset FSVOD-500 comprising of 500 classes with class-balanced videos in each category for few-shot learning; 2) a novel Tube Proposal Network (TPN) to generate high-quality video tube proposals for aggregating feature representation for the target video object which can be highly dynamic; 3) a strategically improved Temporal Matching Network (TMN+) for matching representative query tube features with better discriminative ability thus achieving higher diversity.
1 code implementation • CVPR 2021 • Yanan sun, Guanzhi Wang, Qiao Gu, Chi-Keung Tang, Yu-Wing Tai
Despite the significant progress made by deep learning in natural image matting, there has been so far no representative work on deep learning for video matting due to the inherent technical challenges in reasoning temporal domain and lack of large-scale video matting datasets.
no code implementations • 19 Apr 2021 • Zhen Wei, Bingkun Liu, Weinong Wang, Yu-Wing Tai
Thus, there is always a great demand in customized data annotations.
1 code implementation • CVPR 2021 • Yanan sun, Chi-Keung Tang, Yu-Wing Tai
Specifically, we consider and learn 20 classes of matting patterns, and propose to extend the conventional trimap to semantic trimap.
1 code implementation • CVPR 2021 • Lei Ke, Yu-Wing Tai, Chi-Keung Tang
Segmenting highly-overlapping objects is challenging, because typically no distinction is made between real object contours and occlusion boundaries.
Ranked #1 on
Instance Segmentation
on KINS
1 code implementation • CVPR 2021 • Qi Fan, Deng-Ping Fan, Huazhu Fu, Chi Keung Tang, Ling Shao, Yu-Wing Tai
We present a novel group collaborative learning framework (GCoNet) capable of detecting co-salient objects in real time (16ms), by simultaneously mining consensus representations at group level based on the two necessary criteria: 1) intra-group compactness to better formulate the consistency among co-salient objects by capturing their inherent shared attributes using our novel group affinity module; 2) inter-group separability to effectively suppress the influence of noisy objects on the output by introducing our new group collaborating module conditioning the inconsistent consensus.
Ranked #4 on
Co-Salient Object Detection
on CoSOD3k
5 code implementations • CVPR 2021 • Ho Kei Cheng, Yu-Wing Tai, Chi-Keung Tang
We present Modular interactive VOS (MiVOS) framework which decouples interaction-to-mask and mask propagation, allowing for higher generalizability and better performance.
Ranked #1 on
Interactive Video Object Segmentation
on DAVIS 2017
(using extra training data)
Interactive Video Object Segmentation
Semantic Segmentation
+2
2 code implementations • 24 Feb 2021 • Yang You, Yujing Lou, Ruoxi Shi, Qi Liu, Yu-Wing Tai, Lizhuang Ma, Weiming Wang, Cewu Lu
Spherical Voxel Convolution and Point Re-sampling are proposed to extract rotation invariant features for each point.
1 code implementation • 17 Nov 2020 • Xiaoyuan Ni, Sizhe Song, Yu-Wing Tai, Chi-Keung Tang
Despite excellent progress has been made, the performance on action recognition still heavily relies on specific datasets, which are difficult to extend new action classes due to labor-intensive labeling.
no code implementations • ICCV 2021 • Jihoon Chung, Cheng-hsin Wuu, Hsuan-ru Yang, Yu-Wing Tai, Chi-Keung Tang
We contribute HAA500, a manually annotated human-centric atomic action dataset for action recognition on 500 classes with over 591K labeled frames.
Ranked #1 on
Action Recognition
on HAA500
no code implementations • 27 Aug 2020 • Ji Liu, Heshan Liu, Mang-Tik Chiu, Yu-Wing Tai, Chi-Keung Tang
We propose a novel pose-guided appearance transfer network for transferring a given reference appearance to a target pose in unprecedented image resolution (1024 * 1024), given respectively an image of the reference and target person.
1 code implementation • ECCV 2020 • Lei Ke, Shichao Li, Yanan sun, Yu-Wing Tai, Chi-Keung Tang
GSNet utilizes a unique four-way feature extraction and fusion scheme and directly regresses 6DoF poses and shapes in a single forward pass.
Ranked #1 on
3D Car Instance Understanding
on ApolloCar3D
no code implementations • ECCV 2020 • Ka Leong Cheng, Zhaoyang Yang, Qifeng Chen, Yu-Wing Tai
Continuous sign language recognition (SLR) is a challenging task that requires learning on both spatial and temporal dimensions of signing frame sequences.
1 code implementation • ECCV 2020 • Qi Fan, Lei Ke, Wenjie Pei, Chi-Keung Tang, Yu-Wing Tai
We propose to learn the underlying class-agnostic commonalities that can be generalized from mask-annotated categories to novel categories.
Ranked #76 on
Instance Segmentation
on COCO test-dev
2 code implementations • ECCV 2020 • Jian-Feng Yan, Zizhuang Wei, Hongwei Yi, Mingyu Ding, Runze Zhang, Yisong Chen, Guoping Wang, Yu-Wing Tai
In this paper, we propose an efficient and effective dense hybrid recurrent multi-view stereo net with dynamic consistency checking, namely $D^{2}$HC-RMVSNet, for accurate dense point cloud reconstruction.
no code implementations • ECCV 2020 • Ran Chen, Yong liu, Mengdan Zhang, Shu Liu, Bei Yu, Yu-Wing Tai
Anchor free methods have defined the new frontier in state-of-the-art object detection researches where accurate bounding box estimation is the key to the success of these methods.
no code implementations • CVPR 2020 • Xuhua Huang, Jiarui Xu, Yu-Wing Tai, Chi-Keung Tang
Significant progress has been made in Video Object Segmentation (VOS), the video object tracking task in its finest level.
Ranked #69 on
Semi-Supervised Video Object Segmentation
on DAVIS 2016
1 code implementation • CVPR 2020 • Shichao Li, Lei Ke, Kevin Pratama, Yu-Wing Tai, Chi-Keung Tang, Kwang-Ting Cheng
End-to-end deep representation learning has achieved remarkable accuracy for monocular 3D human pose estimation, yet these models may fail for unseen poses with limited and fixed training data.
Ranked #11 on
Weakly-supervised 3D Human Pose Estimation
on Human3.6M
1 code implementation • 8 May 2020 • Xiang Li, Lin Zhang, Yau Pun Chen, Yu-Wing Tai, Chi-Keung Tang
Deep learning has revolutionized object detection thanks to large-scale datasets, but their object categories are still arguably very limited.
1 code implementation • CVPR 2020 • Ho Kei Cheng, Jihoon Chung, Yu-Wing Tai, Chi-Keung Tang
In this paper, we propose a novel approach to address the high-resolution segmentation problem without using any high-resolution training data.
Ranked #1 on
Semantic Segmentation
on BIG
(using extra training data)
1 code implementation • CVPR 2020 • Xiankai Lu, Wenguan Wang, Jianbing Shen, Yu-Wing Tai, David Crandall, Steven C. H. Hoi
We propose a new method for video object segmentation (VOS) that addresses object pattern learning from unlabeled videos, unlike most existing methods which rely heavily on extensive annotated data.
no code implementations • 5 Jan 2020 • Lizhao Gao, Hai-Hua Xu, Chong Sun, Junling Liu, Yu-Wing Tai
Existing approaches for fine-grained visual recognition focus on learning marginal region-based representations while neglecting the spatial and scale misalignments, leading to inferior performance.
1 code implementation • ECCV 2020 • Hongwei Yi, Zizhuang Wei, Mingyu Ding, Runze Zhang, Yisong Chen, Guoping Wang, Yu-Wing Tai
n this paper, we propose an effective and efficient pyramid multi-view stereo (MVS) net with self-adaptive view aggregation for accurate and complete dense point cloud reconstruction.
no code implementations • ICCV 2019 • Lei Ke, Wenjie Pei, Ruiyu Li, Xiaoyong Shen, Yu-Wing Tai
State-of-the-art image captioning methods mostly focus on improving visual features, less attention has been paid to utilizing the inherent properties of language to boost captioning performance.
Ranked #4 on
Image Captioning
on COCO
no code implementations • 28 Aug 2019 • Weinong Wang, Wenjie Pei, Qiong Cao, Shu Liu, Yu-Wing Tai
Person re-identification aims to identify whether pairs of images belong to the same person or not.
no code implementations • ICCV 2019 • Canmiao Fu, Wenjie Pei, Qiong Cao, Chaopeng Zhang, Yong Zhao, Xiaoyong Shen, Yu-Wing Tai
Typical methods for supervised sequence modeling are built upon the recurrent neural networks to capture temporal dependencies.
no code implementations • ICCV 2019 • Jinkun Cao, Hongyang Tang, Hao-Shu Fang, Xiaoyong Shen, Cewu Lu, Yu-Wing Tai
Therefore, the easily available human pose dataset, which is of a much larger scale than our labeled animal dataset, provides important prior knowledge to boost up the performance on animal pose estimation.
3 code implementations • CVPR 2020 • Qi Fan, Wei Zhuo, Chi-Keung Tang, Yu-Wing Tai
To train our network, we contribute a new dataset that contains 1000 categories of various objects with high-quality annotations.
Ranked #17 on
Few-Shot Object Detection
on MS-COCO (10-shot)
no code implementations • 4 Aug 2019 • Zhaoyang Yang, Zhenmei Shi, Xiaoyong Shen, Yu-Wing Tai
The proposed SF-Net extracts features in a structured manner and gradually encodes information at the frame level, the gloss level and the sentence level into the feature representation.
no code implementations • 2 Aug 2019 • Zhenmei Shi, Haoyang Fang, Yu-Wing Tai, Chi-Keung Tang
Our Dual Augmented Memory Network (DAWN) is unique in remembering both target and background, and using an improved attention LSTM memory to guide the focus on memorized features.
1 code implementation • CVPR 2020 • Xiang Li, Tianhan Wei, Yau Pun Chen, Yu-Wing Tai, Chi-Keung Tang
In this paper, we are interested in few-shot object segmentation where the number of annotated training examples are limited to 5 only.
Ranked #18 on
Few-Shot Semantic Segmentation
on FSS-1000 (5-shot)
no code implementations • 24 Jul 2019 • Chia-Hung Huang, Hang Yin, Yu-Wing Tai, Chi-Keung Tang
Video stabilization algorithms are of greater importance nowadays with the prevalence of hand-held devices which unavoidably produce videos with undesirable shaky motions.
no code implementations • 2 Jul 2019 • Ruizheng Wu, Xiaodong Gu, Xin Tao, Xiaoyong Shen, Yu-Wing Tai, Jiaya Jia
In this paper, we are interested in generating an cartoon face of a person by using unpaired training data between real faces and cartoon ones.
1 code implementation • CVPR 2019 • Wenjie Pei, Jiyuan Zhang, Xiangrong Wang, Lei Ke, Xiaoyong Shen, Yu-Wing Tai
Typical techniques for video captioning follow the encoder-decoder framework, which can only focus on one source video being processed.
1 code implementation • ICCV 2019 • Qiao Gu, Guanzhi Wang, Mang Tik Chiu, Yu-Wing Tai, Chi-Keung Tang
Central to our method are multiple and overlapping local adversarial discriminators in a content-style disentangling network for achieving local detail transfer between facial images, with the use of asymmetric loss functions for dramatic makeup styles with high-frequency details.
1 code implementation • 23 Nov 2018 • Yang You, Yujing Lou, Qi Liu, Yu-Wing Tai, Lizhuang Ma, Cewu Lu, Weiming Wang
Point cloud analysis without pose priors is very challenging in real applications, as the orientations of point clouds are often unknown.
no code implementations • 2 Aug 2018 • Jinshan Pan, Jiangxin Dong, Yang Liu, Jiawei Zhang, Jimmy Ren, Jinhui Tang, Yu-Wing Tai, Ming-Hsuan Yang
We present an algorithm to directly solve numerous image restoration problems (e. g., image deblurring, image dehazing, image deraining, etc.).
1 code implementation • ECCV 2018 • Hao-Shu Fang, Jinkun Cao, Yu-Wing Tai, Cewu Lu
We propose a new pairwise body-part attention model which can learn to focus on crucial parts, and their correlations for HOI recognition.
Ranked #5 on
Human-Object Interaction Detection
on HICO
no code implementations • CVPR 2018 • Jinshan Pan, Sifei Liu, Deqing Sun, Jiawei Zhang, Yang Liu, Jimmy Ren, Zechao Li, Jinhui Tang, Huchuan Lu, Yu-Wing Tai, Ming-Hsuan Yang
These problems usually involve the estimation of two components of the target signals: structures and details.
1 code implementation • CVPR 2018 • Hao-Shu Fang, Guansong Lu, Xiaolin Fang, Jianwen Xie, Yu-Wing Tai, Cewu Lu
In this paper, we present a novel method to generate synthetic human part segmentation data using easily-obtained human keypoint annotations.
Ranked #4 on
Human Part Segmentation
on PASCAL-Part
(using extra training data)
no code implementations • 26 Nov 2017 • Boyu Liu, Yanzhao Wang, Yu-Wing Tai, Chi-Keung Tang
We introduce a one-shot learning approach for video object tracking.
1 code implementation • ECCV 2018 • Shangzhe Wu, Jiarui Xu, Yu-Wing Tai, Chi-Keung Tang
In state-of-the-art deep HDR imaging, input images are first aligned using optical flows before merging, which are still error-prone due to occlusion and large motions.
1 code implementation • ECCV 2018 • Yongyi Lu, Shangzhe Wu, Yu-Wing Tai, Chi-Keung Tang
We train a generated adversarial network, i. e, contextual GAN to learn the joint distribution of sketch and the corresponding image by using joint images.
no code implementations • ECCV 2018 • Haoye Cai, Chunyan Bai, Yu-Wing Tai, Chi-Keung Tang
In the second stage, a skeleton-to-image network is trained, which is used to generate a human action video given the complete human pose sequence generated in the first stage.
Ranked #5 on
Human action generation
on NTU RGB+D 2D
no code implementations • CVPR 2019 • Xiaohui Zeng, Chenxi Liu, Yu-Siang Wang, Weichao Qiu, Lingxi Xie, Yu-Wing Tai, Chi Keung Tang, Alan L. Yuille
Though image-space adversaries can be interpreted as per-pixel albedo change, we verify that they cannot be well explained along these physically meaningful dimensions, which often have a non-local effect.
no code implementations • ICCV 2017 • Jinshan Pan, Jiangxin Dong, Yu-Wing Tai, Zhixun Su, Ming-Hsuan Yang
Solving blind image deblurring usually requires defining a data fitting function and image priors.
no code implementations • 1 Oct 2017 • Hui Yang, Jinshan Pan, Qiong Yan, Wenxiu Sun, Jimmy Ren, Yu-Wing Tai
In this paper, we introduce a bilinear composition loss function to address the problem of image dehazing.
3 code implementations • ICCV 2017 • Donghyeon Cho, Jinsun Park, Tae-Hyun Oh, Yu-Wing Tai, In So Kweon
Our method implicitly learns an attention map, which leads to a content-aware shift map for image retargeting.
no code implementations • ECCV 2018 • Yongyi Lu, Yu-Wing Tai, Chi-Keung Tang
We are interested in attribute-guided face generation: given a low-res face input image, an attribute vector that can be extracted from a high-res image (attribute image), our new method generates a high-res face image for the low-res input that satisfies the given attributes.
1 code implementation • CVPR 2017 • Jinsun Park, Yu-Wing Tai, Donghyeon Cho, In So Kweon
In this paper, we introduce robust and synergetic hand-crafted features and a simple but efficient deep feature from a convolutional neural network (CNN) architecture for defocus estimation.
Ranked #2 on
Defocus Estimation
on CUHK - Blur Detection Dataset
2 code implementations • CVPR 2017 • Jimmy Ren, Xiaohao Chen, Jianbo Liu, Wenxiu Sun, Jiahao Pang, Qiong Yan, Yu-Wing Tai, Li Xu
In this paper, we proposed a novel single stage end-to-end trainable object detection network to overcome this limitation.
12 code implementations • ICCV 2017 • Hao-Shu Fang, Shuqin Xie, Yu-Wing Tai, Cewu Lu
In this paper, we propose a novel regional multi-person pose estimation (RMPE) framework to facilitate pose estimation in the presence of inaccurate human bounding boxes.
Ranked #1 on
Pose Estimation
on UAV-Human
no code implementations • 18 Aug 2016 • Gyeongmin Choe, Jaesik Park, Yu-Wing Tai, In So Kweon
To resolve the ambiguity in our model between the normals and distances, we utilize an initial 3D mesh from the Kinect fusion and multi-view information to reliably estimate surface details that were not captured and reconstructed by the Kinect fusion.
5 code implementations • 12 Jul 2016 • Hengyuan Hu, Rui Peng, Yu-Wing Tai, Chi-Keung Tang
We alternate the pruning and retraining to further reduce zero activations in a network.
no code implementations • CVPR 2016 • Jaesik Park, Yu-Wing Tai, Sudipta N. Sinha, In So Kweon
We present a robust low-rank matrix factorization method to estimate the unknown parameters of this model.
2 code implementations • CVPR 2016 • Gayoung Lee, Yu-Wing Tai, Junmo Kim
Recent advances in saliency detection have utilized deep learning to obtain high level features to detect salient regions in a scene.
no code implementations • 13 Feb 2016 • Jimmy Ren, Yongtao Hu, Yu-Wing Tai, Chuan Wang, Li Xu, Wenxiu Sun, Qiong Yan
This task not only requires collective perception over both visual and auditory signals, the robustness to handle severe quality degradations and unconstrained content variations are also indispensable.
no code implementations • ICCV 2015 • Hyeokhyen Kwon, Yu-Wing Tai
On the contrary, latest imaging sensors capture a RGB image with resolution of multiple times larger than a hyperspectral image.
no code implementations • 1 Sep 2015 • Tae-Hyun Oh, Yasuyuki Matsushita, Yu-Wing Tai, In So Kweon
The problems related to NNM, or WNNM, can be solved iteratively by applying a closed-form proximal operator, called Singular Value Thresholding (SVT), or Weighted SVT, but they suffer from high computational cost of Singular Value Decomposition (SVD) at each iteration.
no code implementations • CVPR 2015 • Hyeokhyen Kwon, Yu-Wing Tai, Stephen Lin
Depth maps captured by consumer-level depth cameras such as Kinect are usually degraded by noise, missing values, and quantization.
no code implementations • CVPR 2015 • Tae-Hyun Oh, Yasuyuki Matsushita, Yu-Wing Tai, In So Kweon
The problems related to NNM (or WNNM) can be solved iteratively by applying a closed-form proximal operator, called Singular Value Thresholding (SVT) (or Weighted SVT), but they suffer from high computational cost to compute a Singular Value Decomposition (SVD) at each iteration.
no code implementations • CVPR 2015 • Hae-Gon Jeon, Jaesik Park, Gyeongmin Choe, Jinsun Park, Yunsu Bok, Yu-Wing Tai, In So Kweon
This paper introduces an algorithm that accurately estimates depth maps using a lenslet light field camera.
no code implementations • 4 Mar 2015 • Tae-Hyun Oh, Yu-Wing Tai, Jean-Charles Bazin, Hyeongwoo Kim, In So Kweon
Robust Principal Component Analysis (RPCA) via rank minimization is a powerful tool for recovering underlying low-rank structure of clean data corrupted with sparse noise/outliers.
no code implementations • CVPR 2014 • Jaesik Park, Sudipta N. Sinha, Yasuyuki Matsushita, Yu-Wing Tai, In So Kweon
We show that a non-isotropic near point light source rigidly attached to a camera can be calibrated using multiple images of a weakly textured planar scene.
no code implementations • CVPR 2014 • Jiwhan Kim, Dongyoon Han, Yu-Wing Tai, Junmo Kim
By mapping a low dimensional RGB color to a feature vector in a high-dimensional color space, we show that we can linearly separate the salient regions from the background by finding an optimal linear combination of color coefficients in the high-dimensional color space.
no code implementations • CVPR 2014 • Gyeongmin Choe, Jaesik Park, Yu-Wing Tai, In So Kweon
To resolve ambiguity in our model between normals and distance, we utilize an initial 3D mesh from the Kinect fusion and multi-view information to reliably estimate surface details that were not reconstructed by the Kinect fusion.
no code implementations • CVPR 2013 • Lap-Fai Yu, Sai-Kit Yeung, Yu-Wing Tai, Stephen Lin
We present a shading-based shape refinement algorithm which uses a noisy, incomplete depth map from Kinect to help resolve ambiguities in shape-from-shading.