no code implementations • 27 Nov 2024 • Guangli Li, Zhehao Zhou, Tuo Sun, Ping Tan, Li Zhang, Zhen Liang
EEG signals have emerged as a powerful tool in affective brain-computer interfaces, playing a crucial role in emotion recognition.
1 code implementation • 26 Nov 2024 • Junyuan Deng, Wei Yin, Xiaoyang Guo, Qian Zhang, Xiaotao Hu, Weiqiang Ren, Xiaoxiao Long, Ping Tan
Monocular camera calibration is essential for many 3D vision tasks.
no code implementations • 5 Sep 2024 • Hanyang Yu, Xiaoxiao Long, Ping Tan
We aim to address sparse-view reconstruction of a 3D scene by leveraging priors from large-scale vision models.
no code implementations • 12 Aug 2024 • Yinhuai Wang, Qihan Zhao, Runyi Yu, Ailing Zeng, Jing Lin, Zhengyi Luo, Hok Wai Tsui, Jiwen Yu, Xiu Li, Qifeng Chen, Jian Zhang, Lei Zhang, Ping Tan
SkillMimic employs a unified configuration to learn diverse skills from human-ball motion datasets, with skill diversity and generalization improving as the dataset grows.
no code implementations • 3 Aug 2024 • Xiaodong Gu, Weihao Yuan, Heng Li, Zilong Dong, Ping Tan
To better represent 3D shapes, we introduce a volume encoding to explicitly encode the spatial information.
1 code implementation • 19 Jul 2024 • Suyi Chen, Hao Xu, Haipeng Li, Kunming Luo, Guanghui Liu, Chi-Wing Fu, Ping Tan, Shuaicheng Liu
To enhance the data realism, we formulate a generative model as a depth inpainting diffusion to process the target depth map with the re-projected source depth map as the condition.
1 code implementation • 16 Jul 2024 • Jingwen Yu, Hanjing Ye, Jianhao Jiao, Ping Tan, Hong Zhang
Existing feature matching methods focus on homography and pose estimation in long-term visual localization, lacking references for geometric verification.
1 code implementation • 3 Jun 2024 • Baowen Zhang, Chuan Fang, Rakesh Shrestha, Yixun Liang, Xiaoxiao Long, Ping Tan
Our method is a significant advancement in Gaussian Splatting and can be directly integrated into existing Gaussian Splatting-based methods.
1 code implementation • 23 May 2024 • Weiyu Li, Jiarui Liu, Rui Chen, Yixun Liang, Xuelin Chen, Ping Tan, Xiaoxiao Long
We present a novel generative 3D modeling system, coined CraftsMan, which can generate high-fidelity 3D geometries with highly varied shapes, regular mesh topologies, and detailed surfaces, and, notably, allows for refining the geometry in an interactive manner.
no code implementations • 19 May 2024 • Peng Li, YuAn Liu, Xiaoxiao Long, Feihu Zhang, Cheng Lin, Mengfei Li, Xingqun Qi, Shanghang Zhang, Wenhan Luo, Ping Tan, Wenping Wang, Qifeng Liu, Yike Guo
Specifically, these methods assume that the input images should comply with a predefined camera type, e. g. a perspective camera with a fixed focal length, leading to distorted shapes when the assumption fails.
1 code implementation • 23 Apr 2024 • Rui Chen, Mingyi Shi, Shaoli Huang, Ping Tan, Taku Komura, Xuelin Chen
We present a novel character control framework that effectively utilizes motion diffusion probabilistic models to generate high-quality and diverse character animations, responding in real-time to a variety of dynamic user-supplied control signals.
no code implementations • CVPR 2024 • Xiangyue Liu, Han Xue, Kunming Luo, Ping Tan, Li Yi
We present GenN2N, a unified NeRF-to-NeRF translation framework for various NeRF translation tasks such as text-driven NeRF editing, colorization, super-resolution, inpainting, etc.
no code implementations • CVPR 2024 • Ziqian Bai, Feitong Tan, Sean Fanello, Rohit Pandey, Mingsong Dou, Shichen Liu, Ping Tan, yinda zhang
To address these challenges, we propose a novel fast 3D neural implicit head avatar model that achieves real-time rendering while maintaining fine-grained controllability and high rendering quality.
no code implementations • 18 Mar 2024 • Xiao Fu, Wei Yin, Mu Hu, Kaixuan Wang, Yuexin Ma, Ping Tan, Shaojie Shen, Dahua Lin, Xiaoxiao Long
We introduce GeoWizard, a new generative foundation model designed for estimating geometric attributes, e. g., depth and normals, from single images.
1 code implementation • CVPR 2024 • Jie Tang, Fei-Peng Tian, Boshi An, Jian Li, Ping Tan
Depth completion aims to derive a dense depth map from sparse depth measurements with a synchronized color image.
no code implementations • 26 Feb 2024 • Jin Ding, Jie-Chao Zhao, Yong-Zhi Sun, Ping Tan, Jia-Wei Wang, Ji-En Ma, You-Tong Fang
Inspired by the principal way that human eyes recognize objects, i. e., largely relying on the shape features, this paper first employs the edge detectors as layer kernels and designs a binary edge feature branch (BEFB for short) to learn the binary edge features, which can be easily integrated into any popular backbone.
no code implementations • CVPR 2024 • Fan Fei, Jiajun Tang, Ping Tan, Boxin Shi
This paper introduces a versatile multi-view inverse rendering framework with near- and far-field light sources.
no code implementations • 5 Oct 2023 • Chuan Fang, Yuan Dong, Kunming Luo, Xiaotao Hu, Rakesh Shrestha, Ping Tan
Next, the Appearance Generation Stage employs a fine-tuned ControlNet to produce a vivid panoramic image of the room guided by the 3D scene layout and text prompt.
1 code implementation • 4 Oct 2023 • Weiyu Li, Rui Chen, Xuelin Chen, Ping Tan
Therefore, we improve the consistency by aligning the 2D geometric priors in diffusion models with well-defined 3D shapes during the lifting, addressing the vast majority of the problem.
no code implementations • 25 Sep 2023 • Xiongfeng Peng, Zhihua Liu, Weiming Li, Ping Tan, SoonYong Cho, Qiang Wang
Recent deep learning based visual simultaneous localization and mapping (SLAM) methods have made significant progress.
no code implementations • 7 Aug 2023 • Xiang Feng, Kaizhang Kang, Fan Pei, Huakeng Ding, Jinjiang You, Ping Tan, Kun Zhou, Hongzhi Wu
We propose a novel framework to automatically learn to aggregate and transform photometric measurements from multiple unstructured views into spatially distinctive and view-invariant low-level features, which are fed to a multi-view stereo method to enhance 3D reconstruction.
no code implementations • 25 Jul 2023 • Sicong Tang, Guangyuan Wang, Qing Ran, Lingzhi Li, Li Shen, Ping Tan
We present a novel method for reconstructing clothed humans from a sparse set of, e. g., 1 to 6 RGB images.
no code implementations • 29 May 2023 • Lingzhi Li, Zhongshu Wang, Zhen Shen, Li Shen, Ping Tan
Reconstructing neural radiance fields with explicit volumetric representations, demonstrated by Plenoxels, has shown remarkable advantages on training and rendering efficiency, while grid-based representations typically induce considerable overhead for storage and transmission.
no code implementations • CVPR 2024 • Yuan Dong, Chuan Fang, Liefeng Bo, Zilong Dong, Ping Tan
Panoramic image enables deeper understanding and more holistic perception of $360^\circ$ surrounding environment, which can naturally encode enriched scene context information compared to standard perspective image.
no code implementations • CVPR 2023 • Ziqian Bai, Feitong Tan, Zeng Huang, Kripasindhu Sarkar, Danhang Tang, Di Qiu, Abhimitra Meka, Ruofei Du, Mingsong Dou, Sergio Orts-Escolano, Rohit Pandey, Ping Tan, Thabo Beeler, Sean Fanello, yinda zhang
The learnt avatar is driven by a parametric face model to achieve user-controlled facial expressions and head poses.
no code implementations • ICCV 2023 • Xinglong Luo, Kunming Luo, Ao Luo, Zhengning Wang, Ping Tan, Shuaicheng Liu
Previous datasets are created by either capturing real scenes by event cameras or synthesizing from images with pasted foreground objects.
no code implementations • 11 Mar 2023 • Jin Ding, Jie-Chao Zhao, Yong-Zhi Sun, Ping Tan, Ji-En Ma, You-Tong Fang
To answer this question, this paper makes a beginning effort by proposing a shallow binary feature module (SBFM for short), which can be integrated into any popular backbone.
no code implementations • 21 Jan 2023 • Heng Li, Xiaodong Gu, Weihao Yuan, Luwei Yang, Zilong Dong, Ping Tan
To reach this challenging goal without depth input, we introduce a hierarchical feature volume to facilitate the implicit map decoder.
no code implementations • ICCV 2023 • Chaoran Tian, Weihong Pan, Zimo Wang, Mao Mao, Guofeng Zhang, Hujun Bao, Ping Tan, Zhaopeng Cui
Stereo depth estimation usually struggles to deal with textureless scenes for both traditional and learning-based methods due to the inherent dependence on image correspondence matching.
1 code implementation • CVPR 2022 • Heng Li, Zhaopeng Cui, Shuaicheng Liu, Ping Tan
Our graph optimizer iteratively refines the global camera rotations by minimizing each node's single rotation objective function.
1 code implementation • ICCV 2023 • Zhuofan Zhang, Zhen Liu, Ping Tan, Bing Zeng, Shuaicheng Liu
In this work, we adopt recent off-the-shelf high-quality deep motion models for motion estimation to recover the camera trajectory and focus on the latter two steps.
1 code implementation • CVPR 2023 • Shitao Tang, Sicong Tang, Andrea Tagliasacchi, Ping Tan, Yasutaka Furukawa
State-of-the-art feature matching methods require each scene to be stored as a 3D point cloud with per-point features, consuming several gigabytes of storage per scene.
1 code implementation • 26 Oct 2022 • Lingzhi Li, Zhen Shen, Zhongshu Wang, Li Shen, Ping Tan
Instead of training a single model that combines all the frames, we formulate the dynamic modeling problem with an incremental learning paradigm in which per-frame model difference is trained to complement the adaption of a base model on the current frame.
1 code implementation • 14 Oct 2022 • Daiheng Gao, Yuliang Xiu, Kailin Li, Lixin Yang, Feng Wang, Peng Zhang, Bang Zhang, Cewu Lu, Ping Tan
Unity GUI is also provided to generate synthetic hand data with user-defined settings, e. g., pose, camera, background, lighting, textures, and accessories.
1 code implementation • 7 Aug 2022 • Qiyu Dai, Jiyao Zhang, Qiwei Li, Tianhao Wu, Hao Dong, Ziyuan Liu, Ping Tan, He Wang
Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks.
no code implementations • 26 Jul 2022 • Jiahui Zhang, Shitao Tang, Kejie Qiu, Rui Huang, Chuan Fang, Le Cui, Zilong Dong, Siyu Zhu, Ping Tan
Visual relocalization has been a widely discussed problem in 3D vision: given a pre-constructed 3D visual map, the 6 DoF (Degrees-of-Freedom) pose of a query image is estimated.
no code implementations • 23 Jun 2022 • Minghao Gou, Haolin Pan, Hao-Shu Fang, Ziyuan Liu, Cewu Lu, Ping Tan
In this paper, we propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
no code implementations • 23 May 2022 • Xiaodong Gu, Chengzhou Tang, Weihao Yuan, Zuozhuo Dai, Siyu Zhu, Ping Tan
In the experiments, we evaluate the proposed method on both the 3D scene flow estimation and the point cloud registration task.
1 code implementation • 29 Mar 2022 • Jian Cheng, Yanguang Wan, Dexin Zuo, Cuixia Ma, Jian Gu, Ping Tan, Hongan Wang, Xiaoming Deng, yinda zhang
3D hand pose estimation from single depth is a fundamental problem in computer vision, and has wide applications. However, the existing methods still can not achieve satisfactory hand pose estimation results due to view variation and occlusion of human hand.
Ranked #1 on Hand Pose Estimation on ICVL Hands
1 code implementation • 25 Mar 2022 • Ziqian Bai, Timur Bagautdinov, Javier Romero, Michael Zollhöfer, Ping Tan, Shunsuke Saito
In this work, for the first time, we enable autoregressive modeling of implicit avatars.
no code implementations • 22 Mar 2022 • Rakesh Shrestha, Siqi Hu, Minghao Gou, Ziyuan Liu, Ping Tan
We present a dataset of 998 3D models of everyday tabletop objects along with their 847, 000 real world RGB and depth images.
1 code implementation • CVPR 2022 • Weihao Yuan, Xiaodong Gu, Zuozhuo Dai, Siyu Zhu, Ping Tan
While recent works design increasingly complicated and powerful networks to directly regress the depth map, we take the path of CRFs optimization.
Ranked #1 on Depth Prediction on Matterport3D
no code implementations • 13 Jan 2022 • Feitong Tan, Sean Fanello, Abhimitra Meka, Sergio Orts-Escolano, Danhang Tang, Rohit Pandey, Jonathan Taylor, Ping Tan, yinda zhang
We propose VoLux-GAN, a generative framework to synthesize 3D-aware faces with convincing relighting.
1 code implementation • ICLR 2022 • Shitao Tang, Jiahui Zhang, Siyu Zhu, Ping Tan
Transformers have been successful in many vision tasks, thanks to their capability of capturing long-range dependency.
no code implementations • CVPR 2022 • Zhaohua Zheng, Jianfang Li, Lingjie Zhu, Honghua Li, Frank Petzold, Ping Tan
Spotting graphical symbols from the computer-aided design (CAD) drawings is essential to many industrial applications.
no code implementations • CVPR 2022 • Weihao Yuan, Xiaodong Gu, Zuozhuo Dai, Siyu Zhu, Ping Tan
Estimating the accurate depth from a single image is challenging since it is inherently ambiguous and ill-posed.
no code implementations • CVPR 2022 • Luwei Yang, Rakesh Shrestha, Wenbo Li, Shuaicheng Liu, Guofeng Zhang, Zhaopeng Cui, Ping Tan
Standard visual localization methods build a priori 3D model of a scene which is used to establish correspondences against the 2D keypoints in a query image.
1 code implementation • CVPR 2022 • Xiaodong Gu, Chengzhou Tang, Weihao Yuan, Zuozhuo Dai, Siyu Zhu, Ping Tan
In the experiments, we evaluate the proposed method on both the 3D scene flow estimation and the point cloud registration task.
no code implementations • CVPR 2022 • Chengzhou Tang, Yuqiang Yang, Bing Zeng, Ping Tan, Shuaicheng Liu
To these ends, we design a method that receives a low-resolution RAW as the input and estimates the desired higher-resolution RAW jointly with the degradation model.
no code implementations • 22 Nov 2021 • Lizhe Liu, Mingqiang Chen, Xiaohao Chen, Siyu Zhu, Ping Tan
Our GB-CosFace introduces an adaptive global boundary to determine whether two face samples belong to the same identity so that the optimization objective is aligned with the testing process from the perspective of open set classification.
1 code implementation • CVPR 2021 • Luwei Yang, Heng Li, Jamal Ahmed Rahim, Zhaopeng Cui, Ping Tan
These methods can suffer from bad initializations due to the noisy spanning tree or outliers in input relative rotations.
no code implementations • ICCV 2021 • Zhiwen Fan, Lingjie Zhu, Honghua Li, Xiaohao Chen, Siyu Zhu, Ping Tan
The proposed CNN-GCN method achieved state-of-the-art (SOTA) performance on the task of semantic symbol spotting, and help us build a baseline network for the panoptic symbol spotting task.
4 code implementations • ICCV 2021 • Lizhe Liu, Xiaohao Chen, Siyu Zhu, Ping Tan
Modern deep-learning-based lane detection methods are successful in most scenarios but struggling for lane lines with complex topologies.
Ranked #8 on Lane Detection on CurveLanes (using extra training data)
no code implementations • 9 Apr 2021 • Weihao Yuan, Yazhan Zhang, Bingkun Wu, Siyu Zhu, Ping Tan, Michael Yu Wang, Qifeng Chen
Self-supervised learning for depth estimation possesses several advantages over supervised learning.
1 code implementation • CVPR 2021 • Ziqian Bai, Zhaopeng Cui, Xiaoming Liu, Ping Tan
This paper presents a method for riggable 3D face reconstruction from monocular images, which jointly estimates a personalized face rig and per-image parameters including expressions, poses, and illuminations.
1 code implementation • CVPR 2021 • Shitao Tang, Chengzhou Tang, Rui Huang, Siyu Zhu, Ping Tan
We present a new method for scene agnostic camera localization using dense scene matching (DSM), where a cost volume is constructed between a query image and a scene.
1 code implementation • CVPR 2021 • Feitong Tan, Danhang Tang, Mingsong Dou, Kaiwen Guo, Rohit Pandey, Cem Keskin, Ruofei Du, Deqing Sun, Sofien Bouaziz, Sean Fanello, Ping Tan, yinda zhang
In this paper, we address the problem of building dense correspondences between human images under arbitrary camera viewpoints and body poses.
no code implementations • 27 Mar 2021 • Rui Huang, Chuan Fang, Kejie Qiu, Le Cui, Zilong Dong, Siyu Zhu, Ping Tan
Secondly, we propose an AR mapping pipeline which takes the input from the scanning device and produces accurate AR Maps.
no code implementations • ICCV 2021 • Kaizhang Kang, Cihui Xie, Ruisheng Zhu, Xiaohe Ma, Ping Tan, Hongzhi Wu, Kun Zhou
We present a novel framework to learn to convert the perpixel photometric information at each view into spatially distinctive and view-invariant low-level features, which can be plugged into existing multi-view stereo pipeline for enhanced 3D reconstruction.
1 code implementation • 24 Mar 2021 • Xiaodong Gu, Weihao Yuan, Zuozhuo Dai, Siyu Zhu, Chengzhou Tang, Zilong Dong, Ping Tan
There are increasing interests of studying the video-to-depth (V2D) problem with machine learning techniques.
3 code implementations • 22 Mar 2021 • Zuozhuo Dai, Guangyuan Wang, Weihao Yuan, Xiaoli Liu, Siyu Zhu, Ping Tan
Thus, our method can solve the problem of cluster inconsistency and be applicable to larger data sets.
Ranked #1 on Unsupervised Person Re-Identification on PersonX
no code implementations • ICCV 2021 • Baowen Zhang, Yangang Wang, Xiaoming Deng, yinda zhang, Ping Tan, Cuixia Ma, Hongan Wang
In this paper, we propose a novel deep learning framework to reconstruct 3D hand poses and shapes of two interacting hands from a single color image.
Ranked #6 on 3D Interacting Hand Pose Estimation on InterHand2.6M
no code implementations • 17 Oct 2020 • Rakesh Shrestha, Zhiwen Fan, Qingkun Su, Zuozhuo Dai, Siyu Zhu, Ping Tan
Deep learning based 3D shape generation methods generally utilize latent features extracted from color images to encode the semantics of objects and guide the shape generation process.
no code implementations • ECCV 2020 • Boren Li, Po-Yu Zhuang, Jian Gu, Mingyang Li, Ping Tan
As for the proposed method, we first train a foreground encoder to learn representations of interchangeable foregrounds.
1 code implementation • CVPR 2020 • Feitong Tan, Hao Zhu, Zhaopeng Cui, Siyu Zhu, Marc Pollefeys, Ping Tan
Previous methods on estimating detailed human depth often require supervised training with `ground truth' depth data.
no code implementations • CVPR 2020 • Chengzhou Tang, Lu Yuan, Ping Tan
We study the energy minimization problem in low-level vision tasks from a novel perspective.
1 code implementation • CVPR 2020 • Lei Li, Siyu Zhu, Hongbo Fu, Ping Tan, Chiew-Lan Tai
In this work, we propose an end-to-end framework to learn local multi-view descriptors for 3D point clouds.
Ranked #5 on Point Cloud Registration on 3DMatch Benchmark
1 code implementation • ICML 2020 • Wenqi Shao, Shitao Tang, Xingang Pan, Ping Tan, Xiaogang Wang, Ping Luo
Unlike prior arts that simply removed the inhibited channels, we propose to "wake them up" during training by designing a novel neural building block, termed Channel Equilibrium (CE) block, which enables channels at the same layer to contribute equally to the learned representation.
no code implementations • 22 Feb 2020 • Qian Zhang, Wei Feng, Liang Wan, Fei-Peng Tian, Xiaowei Wang, Ping Tan
Besides, we also theoretically prove the invariance of our ALR approach to the ambiguity of normal and lighting decomposition.
no code implementations • 18 Jan 2020 • Min Li, Zhenglong Zhou, Zhe Wu, Boxin Shi, Changyu Diao, Ping Tan
From a single viewpoint, we use a set of photometric stereo images to identify surface points with the same distance to the camera.
4 code implementations • CVPR 2020 • Xiaodong Gu, Zhiwen Fan, Zuozhuo Dai, Siyu Zhu, Feitong Tan, Ping Tan
The deep multi-view stereo (MVS) and stereo matching approaches generally construct 3D cost volumes to regularize and regress the output depth or disparity.
Ranked #12 on Point Clouds on Tanks and Temples
no code implementations • 17 Nov 2019 • Renjiao Yi, Ping Tan, Stephen Lin
We present an unsupervised approach for factorizing object appearance into highlight, shading, and albedo layers, trained by multi-view real images.
1 code implementation • ICCV 2019 • Sicong Tang, Feitong Tan, Kelvin Cheng, Zhaoyang Li, Siyu Zhu, Ping Tan
To achieve this goal, we separate the depth map into a smooth base shape and a residual detail shape and design a network with two branches to regress them respectively.
no code implementations • 25 Sep 2019 • Wenqi Shao, Shitao Tang, Xingang Pan, Ping Tan, Xiaogang Wang, Ping Luo
However, over-sparse CNNs have many collapsed channels (i. e. many channels with undesired zero values), impeding their learning ability.
2 code implementations • 3 Aug 2019 • Jie Tang, Fei-Peng Tian, Wei Feng, Jian Li, Ping Tan
It is thus necessary to complete the sparse LiDAR data, where a synchronized guidance RGB image is often used to facilitate this completion.
no code implementations • ICLR 2019 • Chengzhou Tang, Ping Tan
The network first generates several basis depth maps according to the input image, and optimizes the final depth as a linear combination of these basis depth maps via feature-metric BA.
5 code implementations • ICCV 2019 • Zuozhuo Dai, Mingqiang Chen, Xiaodong Gu, Siyu Zhu, Ping Tan
In this paper, we propose the Batch DropBlock (BDB) Network which is a two branch network composed of a conventional ResNet-50 as the global branch and a feature dropping branch.
Ranked #8 on Person Re-Identification on Market-1501-C
1 code implementation • 13 Jun 2018 • Chengzhou Tang, Ping Tan
The network first generates several basis depth maps according to the input image and optimizes the final depth as a linear combination of these basis depth maps via feature-metric BA.
no code implementations • CVPR 2018 • Siyu Zhu, Runze Zhang, Lei Zhou, Tianwei Shen, Tian Fang, Ping Tan, Long Quan
This work proposes a divide-and-conquer framework to solve very large global SfM at the scale of millions of images.
no code implementations • CVPR 2018 • Luwei Yang, Feitong Tan, Ao Li, Zhaopeng Cui, Yasutaka Furukawa, Ping Tan
This paper presents a novel polarimetric dense monocular SLAM (PDMS) algorithm based on a polarization camera.
no code implementations • ECCV 2018 • Renjiao Yi, Chenyang Zhu, Ping Tan, Stephen Lin
We present a method for estimating detailed scene illumination using human faces in a single image.
2 code implementations • ECCV 2018 • Ligeng Zhu, Ruizhi Deng, Michael Maire, Zhiwei Deng, Greg Mori, Ping Tan
We explore a key architectural aspect of deep convolutional neural networks: the pattern of internal skip connections used to aggregate outputs of earlier layers for consumption by deeper layers.
no code implementations • 16 Aug 2017 • Chengzhou Tang, Oliver Wang, Ping Tan
Many monocular visual SLAM algorithms are derived from incremental structure-from-motion (SfM) methods.
no code implementations • CVPR 2017 • Zhaopeng Cui, Jinwei Gu, Boxin Shi, Ping Tan, Jan Kautz
Multi-view stereo relies on feature correspondences for 3D reconstruction, and thus is fundamentally flawed in dealing with featureless scenes.
no code implementations • 2 May 2017 • Rui Huang, Danping Zou, Richard Vaughan, Ping Tan
Image-based modeling techniques can now generate photo-realistic 3D models from images.
6 code implementations • ICCV 2017 • Zili Yi, Hao Zhang, Ping Tan, Minglun Gong
Depending on the task complexity, thousands to millions of labeled image pairs are needed to train a conditional GAN.
Ranked #2 on Image-to-Image Translation on Aerial-to-Map
no code implementations • 7 Apr 2017 • Xiaoming Deng, Shuo Yang, yinda zhang, Ping Tan, Liang Chang, Hongan Wang
We propose a novel 3D neural network architecture for 3D hand pose estimation from a single depth image.
no code implementations • 8 Dec 2016 • Xiaoming Deng, Ye Yuan, Yinda Zhang, Ping Tan, Liang Chang, Shuo Yang, Hongan Wang
Hand detection is essential for many hand related tasks, e. g. parsing hand pose, understanding gesture, which are extremely useful for robotics and human-computer interaction.
no code implementations • 5 Jul 2016 • Luwei Yang, Ligen Zhu, Yichen Wei, Shuang Liang, Ping Tan
Previous part-based attribute recognition approaches perform part detection and attribute recognition in separate steps.
no code implementations • CVPR 2016 • Boxin Shi, Zhe Wu, Zhipeng Mo, Dinglong Duan, Sai-Kit Yeung, Ping Tan
Recent progress on photometric stereo extends the technique to deal with general materials and unknown illumination conditions.
no code implementations • CVPR 2016 • Renjiao Yi, Jue Wang, Ping Tan
We present a fully automatic approach to detect and segment fence-like occluders from a video clip.
no code implementations • 12 Mar 2016 • Zhe Wu, Sai-Kit Yeung, Ping Tan
We present a portable device to capture both shape and reflectance of an indoor scene.
no code implementations • ICCV 2015 • Zhaopeng Cui, Ping Tan
Depth images help to upgrade an essential matrix to a similarity transformation, which can determine the scale of relative translation.
no code implementations • CVPR 2015 • Zhuwen Li, Ping Tan, Robby T. Tan, Danping Zou, Steven Zhiying Zhou, Loong-Fah Cheong
We present a method to jointly estimate scene depth and recover the clear latent image from a foggy video sequence.
no code implementations • 6 Mar 2015 • Zhaopeng Cui, Nianjuan Jiang, Chengzhou Tang, Ping Tan
This paper derives a novel linear position constraint for cameras seeing a common scene point, which leads to a direct linear method for global camera translation estimation.
no code implementations • CVPR 2014 • Rakesh Shiradkar, Li Shen, George Landon, Sim Heng Ong, Ping Tan
The surface bi-directional reflectance distribution function (BRDF) can be used to distinguish different materials.
no code implementations • CVPR 2014 • Shuaicheng Liu, Lu Yuan, Ping Tan, Jian Sun
We propose a novel motion model, SteadyFlow, to represent the motion between neighboring video frames for stabilization.
no code implementations • CVPR 2013 • Yinda Zhang, Jianxiong Xiao, James Hays, Ping Tan
We analyze the self-similarity of the guide image to generate a set of allowable local transformations and apply them to the input image.
no code implementations • CVPR 2013 • Xiaowu Chen, Dongqing Zou, Steven Zhiying Zhou, Qinping Zhao, Ping Tan
This nonlocal smooth prior and the well known local smooth prior from matting Laplacian complement each other.
no code implementations • CVPR 2013 • Zhe Wu, Ping Tan
Under unknown directional lighting, the uncalibrated Lambertian photometric stereo algorithm recovers the shape of a smooth surface up to the generalized bas-relief (GBR) ambiguity.
no code implementations • CVPR 2013 • Zhenglong Zhou, Zhe Wu, Ping Tan
We present a method to capture both 3D shape and spatially varying reflectance with a multi-view photometric stereo technique that works for general isotropic materials.
no code implementations • ACM Transactions on Graphics 2009 • Tao Chen, Ming-Ming Cheng, Ping Tan, Ariel Shamir, Shi-Min Hu
The composed picture is generated by seamlessly stitching several photographs in agreement with the sketch and text labels; these are found by searching the Internet.