no code implementations • 25 Sep 2023 • Xiongfeng Peng, Zhihua Liu, Weiming Li, Ping Tan, SoonYong Cho, Qiang Wang
Recent deep learning based visual simultaneous localization and mapping (SLAM) methods have made significant progress.
no code implementations • 7 Aug 2023 • Xiang Feng, Kaizhang Kang, Fan Pei, Huakeng Ding, Jinjiang You, Ping Tan, Kun Zhou, Hongzhi Wu
We propose a novel framework to automatically learn to aggregate and transform photometric measurements from multiple unstructured views into spatially distinctive and view-invariant low-level features, which are fed to a multi-view stereo method to enhance 3D reconstruction.
no code implementations • 25 Jul 2023 • Sicong Tang, Guangyuan Wang, Qing Ran, Lingzhi Li, Li Shen, Ping Tan
We present a novel method for reconstructing clothed humans from a sparse set of, e. g., 1 to 6 RGB images.
no code implementations • 29 May 2023 • Lingzhi Li, Zhongshu Wang, Zhen Shen, Li Shen, Ping Tan
Reconstructing neural radiance fields with explicit volumetric representations, demonstrated by Plenoxels, has shown remarkable advantages on training and rendering efficiency, while grid-based representations typically induce considerable overhead for storage and transmission.
no code implementations • 21 May 2023 • Yuan Dong, Chuan Fang, Liefeng Bo, Zilong Dong, Ping Tan
Panoramic image enables deeper understanding and more holistic perception of $360^\circ$ surrounding environment, which can naturally encode enriched scene context information compared to standard perspective image.
no code implementations • CVPR 2023 • Ziqian Bai, Feitong Tan, Zeng Huang, Kripasindhu Sarkar, Danhang Tang, Di Qiu, Abhimitra Meka, Ruofei Du, Mingsong Dou, Sergio Orts-Escolano, Rohit Pandey, Ping Tan, Thabo Beeler, Sean Fanello, yinda zhang
The learnt avatar is driven by a parametric face model to achieve user-controlled facial expressions and head poses.
no code implementations • 20 Mar 2023 • Xinglong Luo, Kunming Luo, Ao Luo, Zhengning Wang, Ping Tan, Shuaicheng Liu
Previous datasets are created by either capturing real scenes by event cameras or synthesizing from images with pasted foreground objects.
no code implementations • 11 Mar 2023 • Jin Ding, Jie-Chao Zhao, Yong-Zhi Sun, Ping Tan, Ji-En Ma, You-Tong Fang
To answer this question, this paper makes a beginning effort by proposing a shallow binary feature module (SBFM for short), which can be integrated into any popular backbone.
no code implementations • 21 Jan 2023 • Heng Li, Xiaodong Gu, Weihao Yuan, Luwei Yang, Zilong Dong, Ping Tan
To reach this challenging goal without depth input, we introduce a hierarchical feature volume to facilitate the implicit map decoder.
1 code implementation • CVPR 2022 • Heng Li, Zhaopeng Cui, Shuaicheng Liu, Ping Tan
Our graph optimizer iteratively refines the global camera rotations by minimizing each node's single rotation objective function.
1 code implementation • 5 Dec 2022 • Zhuofan Zhang, Zhen Liu, Ping Tan, Bing Zeng, Shuaicheng Liu
In this work, we adopt recent off-the-shelf high-quality deep motion models for motion estimation to recover the camera trajectory and focus on the latter two steps.
1 code implementation • CVPR 2023 • Shitao Tang, Sicong Tang, Andrea Tagliasacchi, Ping Tan, Yasutaka Furukawa
State-of-the-art feature matching methods require each scene to be stored as a 3D point cloud with per-point features, consuming several gigabytes of storage per scene.
1 code implementation • 26 Oct 2022 • Lingzhi Li, Zhen Shen, Zhongshu Wang, Li Shen, Ping Tan
Instead of training a single model that combines all the frames, we formulate the dynamic modeling problem with an incremental learning paradigm in which per-frame model difference is trained to complement the adaption of a base model on the current frame.
1 code implementation • 14 Oct 2022 • Daiheng Gao, Yuliang Xiu, Kailin Li, Lixin Yang, Feng Wang, Peng Zhang, Bang Zhang, Cewu Lu, Ping Tan
Unity GUI is also provided to generate synthetic hand data with user-defined settings, e. g., pose, camera, background, lighting, textures, and accessories.
1 code implementation • 7 Aug 2022 • Qiyu Dai, Jiyao Zhang, Qiwei Li, Tianhao Wu, Hao Dong, Ziyuan Liu, Ping Tan, He Wang
Commercial depth sensors usually generate noisy and missing depths, especially on specular and transparent objects, which poses critical issues to downstream depth or point cloud-based tasks.
no code implementations • 26 Jul 2022 • Jiahui Zhang, Shitao Tang, Kejie Qiu, Rui Huang, Chuan Fang, Le Cui, Zilong Dong, Siyu Zhu, Ping Tan
Visual relocalization has been a widely discussed problem in 3D vision: given a pre-constructed 3D visual map, the 6 DoF (Degrees-of-Freedom) pose of a query image is estimated.
no code implementations • 23 Jun 2022 • Minghao Gou, Haolin Pan, Hao-Shu Fang, Ziyuan Liu, Cewu Lu, Ping Tan
In this paper, we propose a new task that enables and facilitates algorithms to estimate the 6D pose estimation of novel objects during testing.
no code implementations • 23 May 2022 • Xiaodong Gu, Chengzhou Tang, Weihao Yuan, Zuozhuo Dai, Siyu Zhu, Ping Tan
In the experiments, we evaluate the proposed method on both the 3D scene flow estimation and the point cloud registration task.
1 code implementation • 29 Mar 2022 • Jian Cheng, Yanguang Wan, Dexin Zuo, Cuixia Ma, Jian Gu, Ping Tan, Hongan Wang, Xiaoming Deng, yinda zhang
3D hand pose estimation from single depth is a fundamental problem in computer vision, and has wide applications. However, the existing methods still can not achieve satisfactory hand pose estimation results due to view variation and occlusion of human hand.
Ranked #1 on
Hand Pose Estimation
on ICVL Hands
1 code implementation • 25 Mar 2022 • Ziqian Bai, Timur Bagautdinov, Javier Romero, Michael Zollhöfer, Ping Tan, Shunsuke Saito
In this work, for the first time, we enable autoregressive modeling of implicit avatars.
no code implementations • 22 Mar 2022 • Rakesh Shrestha, Siqi Hu, Minghao Gou, Ziyuan Liu, Ping Tan
We present a dataset of 998 3D models of everyday tabletop objects along with their 847, 000 real world RGB and depth images.
1 code implementation • CVPR 2022 • Weihao Yuan, Xiaodong Gu, Zuozhuo Dai, Siyu Zhu, Ping Tan
While recent works design increasingly complicated and powerful networks to directly regress the depth map, we take the path of CRFs optimization.
Ranked #1 on
Depth Prediction
on Matterport3D
no code implementations • 13 Jan 2022 • Feitong Tan, Sean Fanello, Abhimitra Meka, Sergio Orts-Escolano, Danhang Tang, Rohit Pandey, Jonathan Taylor, Ping Tan, yinda zhang
We propose VoLux-GAN, a generative framework to synthesize 3D-aware faces with convincing relighting.
1 code implementation • ICLR 2022 • Shitao Tang, Jiahui Zhang, Siyu Zhu, Ping Tan
Transformers have been successful in many vision tasks, thanks to their capability of capturing long-range dependency.
no code implementations • CVPR 2022 • Zhaohua Zheng, Jianfang Li, Lingjie Zhu, Honghua Li, Frank Petzold, Ping Tan
Spotting graphical symbols from the computer-aided design (CAD) drawings is essential to many industrial applications.
1 code implementation • CVPR 2022 • Xiaodong Gu, Chengzhou Tang, Weihao Yuan, Zuozhuo Dai, Siyu Zhu, Ping Tan
In the experiments, we evaluate the proposed method on both the 3D scene flow estimation and the point cloud registration task.
no code implementations • CVPR 2022 • Weihao Yuan, Xiaodong Gu, Zuozhuo Dai, Siyu Zhu, Ping Tan
Estimating the accurate depth from a single image is challenging since it is inherently ambiguous and ill-posed.
no code implementations • CVPR 2022 • Luwei Yang, Rakesh Shrestha, Wenbo Li, Shuaicheng Liu, Guofeng Zhang, Zhaopeng Cui, Ping Tan
Standard visual localization methods build a priori 3D model of a scene which is used to establish correspondences against the 2D keypoints in a query image.
no code implementations • CVPR 2022 • Chengzhou Tang, Yuqiang Yang, Bing Zeng, Ping Tan, Shuaicheng Liu
To these ends, we design a method that receives a low-resolution RAW as the input and estimates the desired higher-resolution RAW jointly with the degradation model.
no code implementations • 22 Nov 2021 • Lizhe Liu, Mingqiang Chen, Xiaohao Chen, Siyu Zhu, Ping Tan
Our GB-CosFace introduces an adaptive global boundary to determine whether two face samples belong to the same identity so that the optimization objective is aligned with the testing process from the perspective of open set classification.
1 code implementation • CVPR 2021 • Luwei Yang, Heng Li, Jamal Ahmed Rahim, Zhaopeng Cui, Ping Tan
These methods can suffer from bad initializations due to the noisy spanning tree or outliers in input relative rotations.
no code implementations • ICCV 2021 • Zhiwen Fan, Lingjie Zhu, Honghua Li, Xiaohao Chen, Siyu Zhu, Ping Tan
The proposed CNN-GCN method achieved state-of-the-art (SOTA) performance on the task of semantic symbol spotting, and help us build a baseline network for the panoptic symbol spotting task.
3 code implementations • ICCV 2021 • Lizhe Liu, Xiaohao Chen, Siyu Zhu, Ping Tan
Modern deep-learning-based lane detection methods are successful in most scenarios but struggling for lane lines with complex topologies.
Ranked #5 on
Lane Detection
on CurveLanes
(using extra training data)
no code implementations • 9 Apr 2021 • Weihao Yuan, Yazhan Zhang, Bingkun Wu, Siyu Zhu, Ping Tan, Michael Yu Wang, Qifeng Chen
Self-supervised learning for depth estimation possesses several advantages over supervised learning.
1 code implementation • CVPR 2021 • Ziqian Bai, Zhaopeng Cui, Xiaoming Liu, Ping Tan
This paper presents a method for riggable 3D face reconstruction from monocular images, which jointly estimates a personalized face rig and per-image parameters including expressions, poses, and illuminations.
1 code implementation • CVPR 2021 • Shitao Tang, Chengzhou Tang, Rui Huang, Siyu Zhu, Ping Tan
We present a new method for scene agnostic camera localization using dense scene matching (DSM), where a cost volume is constructed between a query image and a scene.
1 code implementation • CVPR 2021 • Feitong Tan, Danhang Tang, Mingsong Dou, Kaiwen Guo, Rohit Pandey, Cem Keskin, Ruofei Du, Deqing Sun, Sofien Bouaziz, Sean Fanello, Ping Tan, yinda zhang
In this paper, we address the problem of building dense correspondences between human images under arbitrary camera viewpoints and body poses.
no code implementations • ICCV 2021 • Kaizhang Kang, Cihui Xie, Ruisheng Zhu, Xiaohe Ma, Ping Tan, Hongzhi Wu, Kun Zhou
We present a novel framework to learn to convert the perpixel photometric information at each view into spatially distinctive and view-invariant low-level features, which can be plugged into existing multi-view stereo pipeline for enhanced 3D reconstruction.
no code implementations • 27 Mar 2021 • Rui Huang, Chuan Fang, Kejie Qiu, Le Cui, Zilong Dong, Siyu Zhu, Ping Tan
Secondly, we propose an AR mapping pipeline which takes the input from the scanning device and produces accurate AR Maps.
1 code implementation • 24 Mar 2021 • Xiaodong Gu, Weihao Yuan, Zuozhuo Dai, Siyu Zhu, Chengzhou Tang, Zilong Dong, Ping Tan
There are increasing interests of studying the video-to-depth (V2D) problem with machine learning techniques.
3 code implementations • 22 Mar 2021 • Zuozhuo Dai, Guangyuan Wang, Weihao Yuan, Xiaoli Liu, Siyu Zhu, Ping Tan
Thus, our method can solve the problem of cluster inconsistency and be applicable to larger data sets.
Ranked #1 on
Unsupervised Person Re-Identification
on PersonX
no code implementations • ICCV 2021 • Baowen Zhang, Yangang Wang, Xiaoming Deng, yinda zhang, Ping Tan, Cuixia Ma, Hongan Wang
In this paper, we propose a novel deep learning framework to reconstruct 3D hand poses and shapes of two interacting hands from a single color image.
no code implementations • 17 Oct 2020 • Rakesh Shrestha, Zhiwen Fan, Qingkun Su, Zuozhuo Dai, Siyu Zhu, Ping Tan
Deep learning based 3D shape generation methods generally utilize latent features extracted from color images to encode the semantics of objects and guide the shape generation process.
no code implementations • ECCV 2020 • Boren Li, Po-Yu Zhuang, Jian Gu, Mingyang Li, Ping Tan
As for the proposed method, we first train a foreground encoder to learn representations of interchangeable foregrounds.
1 code implementation • CVPR 2020 • Feitong Tan, Hao Zhu, Zhaopeng Cui, Siyu Zhu, Marc Pollefeys, Ping Tan
Previous methods on estimating detailed human depth often require supervised training with `ground truth' depth data.
no code implementations • CVPR 2020 • Chengzhou Tang, Lu Yuan, Ping Tan
We study the energy minimization problem in low-level vision tasks from a novel perspective.
1 code implementation • CVPR 2020 • Lei Li, Siyu Zhu, Hongbo Fu, Ping Tan, Chiew-Lan Tai
In this work, we propose an end-to-end framework to learn local multi-view descriptors for 3D point clouds.
Ranked #5 on
Point Cloud Registration
on 3DMatch Benchmark
1 code implementation • ICML 2020 • Wenqi Shao, Shitao Tang, Xingang Pan, Ping Tan, Xiaogang Wang, Ping Luo
Unlike prior arts that simply removed the inhibited channels, we propose to "wake them up" during training by designing a novel neural building block, termed Channel Equilibrium (CE) block, which enables channels at the same layer to contribute equally to the learned representation.
no code implementations • 22 Feb 2020 • Qian Zhang, Wei Feng, Liang Wan, Fei-Peng Tian, Xiaowei Wang, Ping Tan
Besides, we also theoretically prove the invariance of our ALR approach to the ambiguity of normal and lighting decomposition.
no code implementations • 18 Jan 2020 • Min Li, Zhenglong Zhou, Zhe Wu, Boxin Shi, Changyu Diao, Ping Tan
From a single viewpoint, we use a set of photometric stereo images to identify surface points with the same distance to the camera.
4 code implementations • CVPR 2020 • Xiaodong Gu, Zhiwen Fan, Zuozhuo Dai, Siyu Zhu, Feitong Tan, Ping Tan
The deep multi-view stereo (MVS) and stereo matching approaches generally construct 3D cost volumes to regularize and regress the output depth or disparity.
Ranked #7 on
Point Clouds
on Tanks and Temples
no code implementations • 17 Nov 2019 • Renjiao Yi, Ping Tan, Stephen Lin
We present an unsupervised approach for factorizing object appearance into highlight, shading, and albedo layers, trained by multi-view real images.
1 code implementation • ICCV 2019 • Sicong Tang, Feitong Tan, Kelvin Cheng, Zhaoyang Li, Siyu Zhu, Ping Tan
To achieve this goal, we separate the depth map into a smooth base shape and a residual detail shape and design a network with two branches to regress them respectively.
no code implementations • 25 Sep 2019 • Wenqi Shao, Shitao Tang, Xingang Pan, Ping Tan, Xiaogang Wang, Ping Luo
However, over-sparse CNNs have many collapsed channels (i. e. many channels with undesired zero values), impeding their learning ability.
2 code implementations • 3 Aug 2019 • Jie Tang, Fei-Peng Tian, Wei Feng, Jian Li, Ping Tan
It is thus necessary to complete the sparse LiDAR data, where a synchronized guidance RGB image is often used to facilitate this completion.
no code implementations • ICLR 2019 • Chengzhou Tang, Ping Tan
The network first generates several basis depth maps according to the input image, and optimizes the final depth as a linear combination of these basis depth maps via feature-metric BA.
5 code implementations • ICCV 2019 • Zuozhuo Dai, Mingqiang Chen, Xiaodong Gu, Siyu Zhu, Ping Tan
In this paper, we propose the Batch DropBlock (BDB) Network which is a two branch network composed of a conventional ResNet-50 as the global branch and a feature dropping branch.
Ranked #8 on
Person Re-Identification
on Market-1501-C
1 code implementation • 13 Jun 2018 • Chengzhou Tang, Ping Tan
The network first generates several basis depth maps according to the input image and optimizes the final depth as a linear combination of these basis depth maps via feature-metric BA.
no code implementations • CVPR 2018 • Siyu Zhu, Runze Zhang, Lei Zhou, Tianwei Shen, Tian Fang, Ping Tan, Long Quan
This work proposes a divide-and-conquer framework to solve very large global SfM at the scale of millions of images.
no code implementations • CVPR 2018 • Luwei Yang, Feitong Tan, Ao Li, Zhaopeng Cui, Yasutaka Furukawa, Ping Tan
This paper presents a novel polarimetric dense monocular SLAM (PDMS) algorithm based on a polarization camera.
no code implementations • ECCV 2018 • Renjiao Yi, Chenyang Zhu, Ping Tan, Stephen Lin
We present a method for estimating detailed scene illumination using human faces in a single image.
2 code implementations • ECCV 2018 • Ligeng Zhu, Ruizhi Deng, Michael Maire, Zhiwei Deng, Greg Mori, Ping Tan
We explore a key architectural aspect of deep convolutional neural networks: the pattern of internal skip connections used to aggregate outputs of earlier layers for consumption by deeper layers.
no code implementations • 16 Aug 2017 • Chengzhou Tang, Oliver Wang, Ping Tan
Many monocular visual SLAM algorithms are derived from incremental structure-from-motion (SfM) methods.
no code implementations • CVPR 2017 • Zhaopeng Cui, Jinwei Gu, Boxin Shi, Ping Tan, Jan Kautz
Multi-view stereo relies on feature correspondences for 3D reconstruction, and thus is fundamentally flawed in dealing with featureless scenes.
no code implementations • 2 May 2017 • Rui Huang, Danping Zou, Richard Vaughan, Ping Tan
Image-based modeling techniques can now generate photo-realistic 3D models from images.
6 code implementations • ICCV 2017 • Zili Yi, Hao Zhang, Ping Tan, Minglun Gong
Depending on the task complexity, thousands to millions of labeled image pairs are needed to train a conditional GAN.
Ranked #2 on
Image-to-Image Translation
on Aerial-to-Map
no code implementations • 7 Apr 2017 • Xiaoming Deng, Shuo Yang, yinda zhang, Ping Tan, Liang Chang, Hongan Wang
We propose a novel 3D neural network architecture for 3D hand pose estimation from a single depth image.
no code implementations • 8 Dec 2016 • Xiaoming Deng, Ye Yuan, Yinda Zhang, Ping Tan, Liang Chang, Shuo Yang, Hongan Wang
Hand detection is essential for many hand related tasks, e. g. parsing hand pose, understanding gesture, which are extremely useful for robotics and human-computer interaction.
no code implementations • 5 Jul 2016 • Luwei Yang, Ligen Zhu, Yichen Wei, Shuang Liang, Ping Tan
Previous part-based attribute recognition approaches perform part detection and attribute recognition in separate steps.
no code implementations • CVPR 2016 • Renjiao Yi, Jue Wang, Ping Tan
We present a fully automatic approach to detect and segment fence-like occluders from a video clip.
no code implementations • CVPR 2016 • Boxin Shi, Zhe Wu, Zhipeng Mo, Dinglong Duan, Sai-Kit Yeung, Ping Tan
Recent progress on photometric stereo extends the technique to deal with general materials and unknown illumination conditions.
no code implementations • 12 Mar 2016 • Zhe Wu, Sai-Kit Yeung, Ping Tan
We present a portable device to capture both shape and reflectance of an indoor scene.
no code implementations • ICCV 2015 • Zhaopeng Cui, Ping Tan
Depth images help to upgrade an essential matrix to a similarity transformation, which can determine the scale of relative translation.
no code implementations • CVPR 2015 • Zhuwen Li, Ping Tan, Robby T. Tan, Danping Zou, Steven Zhiying Zhou, Loong-Fah Cheong
We present a method to jointly estimate scene depth and recover the clear latent image from a foggy video sequence.
no code implementations • 6 Mar 2015 • Zhaopeng Cui, Nianjuan Jiang, Chengzhou Tang, Ping Tan
This paper derives a novel linear position constraint for cameras seeing a common scene point, which leads to a direct linear method for global camera translation estimation.
no code implementations • CVPR 2014 • Shuaicheng Liu, Lu Yuan, Ping Tan, Jian Sun
We propose a novel motion model, SteadyFlow, to represent the motion between neighboring video frames for stabilization.
no code implementations • CVPR 2014 • Rakesh Shiradkar, Li Shen, George Landon, Sim Heng Ong, Ping Tan
The surface bi-directional reflectance distribution function (BRDF) can be used to distinguish different materials.
no code implementations • CVPR 2013 • Yinda Zhang, Jianxiong Xiao, James Hays, Ping Tan
We analyze the self-similarity of the guide image to generate a set of allowable local transformations and apply them to the input image.
no code implementations • CVPR 2013 • Zhe Wu, Ping Tan
Under unknown directional lighting, the uncalibrated Lambertian photometric stereo algorithm recovers the shape of a smooth surface up to the generalized bas-relief (GBR) ambiguity.
no code implementations • CVPR 2013 • Xiaowu Chen, Dongqing Zou, Steven Zhiying Zhou, Qinping Zhao, Ping Tan
This nonlocal smooth prior and the well known local smooth prior from matting Laplacian complement each other.
no code implementations • CVPR 2013 • Zhenglong Zhou, Zhe Wu, Ping Tan
We present a method to capture both 3D shape and spatially varying reflectance with a multi-view photometric stereo technique that works for general isotropic materials.
no code implementations • ACM Transactions on Graphics 2009 • Tao Chen, Ming-Ming Cheng, Ping Tan, Ariel Shamir, Shi-Min Hu
The composed picture is generated by seamlessly stitching several photographs in agreement with the sketch and text labels; these are found by searching the Internet.