no code implementations • 2 Nov 2024 • Takeshi Noda, Chao Chen, Weiqi Zhang, Xinhai Liu, Yu-Shen Liu, Zhizhong Han
Recent methods usually train neural networks to overfit on single point clouds to infer signed distance functions (SDFs).
no code implementations • 25 Oct 2024 • Junsheng Zhou, Weiqi Zhang, Yu-Shen Liu
Through the novel disentanglement of 3DGS, we represent the discrete and unstructured 3DGS with continuous Gaussian Splatting functions, where we then train a latent diffusion model with the target of generating these Gaussian Splatting functions both unconditionally and conditionally.
1 code implementation • 25 Oct 2024 • Chao Chen, Yu-Shen Liu, Zhizhong Han
To resolve this issue, we propose a method to promote pros of both data-driven based and overfitting-based methods for better generalization, faster inference, and higher accuracy in learning neural SDFs.
no code implementations • 24 Oct 2024 • Congcong Wen, Yisiyuan Huang, Hao Huang, Yanjia Huang, Shuaihang Yuan, Yu Hao, Hui Lin, Yu-Shen Liu, Yi Fang
Zero-shot object navigation (ZSON) aims to address this challenge, allowing robots to interact with unknown objects without specific training data.
no code implementations • 24 Oct 2024 • Liang Han, Junsheng Zhou, Yu-Shen Liu, Zhizhong Han
In this paper, We propose a novel method for synthesizing novel views from sparse views with Gaussian Splatting that does not require external prior as supervision.
no code implementations • 21 Oct 2024 • Junsheng Zhou, Yu-Shen Liu, Zhizhong Han
Large language and vision models have been leading a revolution in visual computing.
no code implementations • 18 Oct 2024 • Wenyuan Zhang, Yu-Shen Liu, Zhizhong Han
Although 3DGS provides a promising neural rendering option, it is still hard to infer SDFs for surface reconstruction with 3DGS due to the discreteness, the sparseness, and the off-surface drift of 3D Gaussians.
no code implementations • 26 Aug 2024 • Chao Chen, Yu-Shen Liu, Zhizhong Han
Instead of object-centered coordinate system, current methods generalized global priors learned in seen classes to reconstruct 3D shapes from unseen classes in viewer-centered coordinate system.
no code implementations • 23 Jul 2024 • Wenyuan Zhang, Kanle Shi, Yu-Shen Liu, Zhizhong Han
To infer a UDF for an unseen scene from multiple RGB images, we generalize the learned volume rendering priors to map inferred unsigned distances in alpha blending for RGB image rendering.
no code implementations • 18 Jul 2024 • Shengtao Li, Ge Gao, Yudong Liu, Ming Gu, Yu-Shen Liu
The neural network typically fits the shape with a rough surface and omits fine-grained geometric details such as shape edges and corners.
no code implementations • CVPR 2024 • Junsheng Zhou, Weiqi Zhang, Baorui Ma, Kanle Shi, Yu-Shen Liu, Zhizhong Han
In this work, we present UDiFF, a 3D diffusion model for unsigned distance fields (UDFs) which is capable to generate textured 3D shapes with open surfaces from text conditions or unconditionally.
1 code implementation • 4 Jan 2024 • Shengtao Li, Ge Gao, Yudong Liu, Yu-Shen Liu, Ming Gu
Our method maximizes the spatial expressiveness of grid features and maintains computational efficiency.
no code implementations • 23 Dec 2023 • Shujuan Li, Junsheng Zhou, Baorui Ma, Yu-Shen Liu, Zhizhong Han
At inference time, we randomly sample queries around the sparse point cloud, and project these query points onto the zero-level set of the learned implicit field to generate a dense point cloud.
no code implementations • 21 Dec 2023 • Han Huang, Yulun Wu, Junsheng Zhou, Ge Gao, Ming Gu, Yu-Shen Liu
To achieve this, we train a neural network to learn a global implicit field from the on-surface points obtained from SfM and then leverage it as a coarse geometric constraint.
1 code implementation • NeurIPS 2023 • Junsheng Zhou, Baorui Ma, Wenyuan Zhang, Yi Fang, Yu-Shen Liu, Zhizhong Han
To address these problems, we propose to learn a structured cross-modality latent space to represent pixel features and 3D features via a differentiable probabilistic PnP solver.
2 code implementations • 29 Nov 2023 • Baorui Ma, Haoge Deng, Junsheng Zhou, Yu-Shen Liu, Tiejun Huang, Xinlong Wang
We justify that the refined 3D geometric priors aid in the 3D-aware capability of 2D diffusion priors, which in turn provides superior guidance for the refinement of 3D geometric priors.
1 code implementation • NeurIPS 2023 • Qing Li, Huifang Feng, Kanle Shi, Yue Gao, Yi Fang, Yu-Shen Liu, Zhizhong Han
Specifically, we introduce loss functions to facilitate query points to iteratively reach the moving targets and aggregate onto the approximated surface, thereby learning a global surface representation of the data.
2 code implementations • 10 Oct 2023 • Junsheng Zhou, Jinsheng Wang, Baorui Ma, Yu-Shen Liu, Tiejun Huang, Xinlong Wang
Scaling up representations for images or text has been extensively investigated in the past few years and has led to revolutions in learning vision and language.
Ranked #1 on Zero-shot 3D classification on Objaverse LVIS (using extra training data)
1 code implementation • 17 Sep 2023 • Qing Li, Huifang Feng, Kanle Shi, Yi Fang, Yu-Shen Liu, Zhizhong Han
We propose Neural Gradient Learning (NGL), a deep learning approach to learn gradient vectors with consistent orientation from 3D point clouds for normal estimation.
1 code implementation • ICCV 2023 • Chao Chen, Yu-Shen Liu, Zhizhong Han
However, these methods suffer from a slow inference due to the slow convergence of neural networks and the extensive calculation of distances to surface points, which limits them to small scale points.
2 code implementations • ICCV 2023 • Junsheng Zhou, Baorui Ma, Shujuan Li, Yu-Shen Liu, Zhizhong Han
We pull the non-zero level sets onto the zero level set with gradient constraints which align gradients over different level sets and correct unsigned distance errors on the zero level set, leading to a smoother and more continuous unsigned distance field.
1 code implementation • ICCV 2023 • Peng Xiang, Xin Wen, Yu-Shen Liu, HUI ZHANG, Yi Fang, Zhizhong Han
In this way, the categorization of each point is conditioned on its local semantic pattern.
1 code implementation • 2 Jun 2023 • Baorui Ma, Yu-Shen Liu, Zhizhong Han
However, without ground truth signed distances, point normals or clean point clouds, current methods still struggle from learning SDFs from noisy point clouds.
1 code implementation • CVPR 2023 • Baorui Ma, Junsheng Zhou, Yu-Shen Liu, Zhizhong Han
Our insight is to propagate the zero level set to everywhere in the field through consistent gradients to eliminate uncertainty in the field that is caused by the discreteness of 3D point clouds or the lack of observations from multi-view images.
1 code implementation • CVPR 2023 • Qing Li, Huifang Feng, Kanle Shi, Yue Gao, Yi Fang, Yu-Shen Liu, Zhizhong Han
In this work, we introduce signed hyper surfaces (SHS), which are parameterized by multi-layer perceptron (MLP) layers, to learn to estimate oriented normals from point clouds in an end-to-end manner.
no code implementations • 10 May 2023 • Xinhai Liu, Zhizhong Han, Sanghuk Lee, Yan-Pei Cao, Yu-Shen Liu
Most of early methods selected the important points on 3D shapes by analyzing the intrinsic geometric properties of every single shape, which fails to capture the importance of points that distinguishes a shape from objects of other classes, i. e., the distinction of points.
1 code implementation • CVPR 2023 • Haiping Wang, YuAn Liu, Zhen Dong, Yulan Guo, Yu-Shen Liu, Wenping Wang, Bisheng Yang
Previous multiview registration methods rely on exhaustive pairwise registration to construct a densely-connected pose graph and apply Iteratively Reweighted Least Square (IRLS) on the pose graph to compute the scan poses.
1 code implementation • CVPR 2023 • Chao Chen, Yu-Shen Liu, Zhizhong Han
To resolve this issue, we present a neural network to directly infer SDFs from single sparse point clouds without using signed distance supervision, learned priors or even normals.
no code implementations • CVPR 2023 • Meng Wang, Yu-Shen Liu, Yue Gao, Kanle Shi, Yi Fang, Zhizhong Han
To capture geometry details, current mainstream methods divide 3D shapes into local regions and then learn each one with a local latent code via a decoder, where the decoder shares the geometric similarities among different local regions.
1 code implementation • 30 Nov 2022 • Shujuan Li, Junsheng Zhou, Baorui Ma, Yu-Shen Liu, Zhizhong Han
To resolve these issues, we propose an implicit function to learn an angle field around the normal of each point in the spherical coordinate system, which is dubbed as Neural Angle Fields (NeAF).
Ranked #3 on Surface Normals Estimation on PCPNet
1 code implementation • 13 Oct 2022 • Qing Li, Yu-Shen Liu, Jin-San Cheng, Cheng Wang, Yi Fang, Zhizhong Han
To address these issues, we introduce hyper surface fitting to implicitly learn hyper surfaces, which are represented by multi-layer perceptron (MLP) layers that take point features as input and output surface patterns in a high dimensional feature space.
Ranked #2 on Surface Normals Estimation on PCPNet
1 code implementation • 6 Oct 2022 • Junsheng Zhou, Baorui Ma, Shujuan Li, Yu-Shen Liu, Yi Fang, Zhizhong Han
Some other methods tried to represent open surfaces using unsigned distance functions (UDF) which are learned from ground truth distances.
1 code implementation • 14 Aug 2022 • Wenyuan Zhang, Ruofan Xing, Yunfan Zeng, Yu-Shen Liu, Kanle Shi, Zhizhong Han
Experimental results show that our method achieves comparable accuracy to the state-of-the-art with much faster training.
1 code implementation • 18 Jul 2022 • Chao Chen, Yu-Shen Liu, Zhizhong Han
Our insight here is that both the part learning and the part blending can be conducted much easier in the latent space than in the spatial space.
1 code implementation • CVPR 2022 • Baorui Ma, Yu-Shen Liu, Matthias Zwicker, Zhizhong Han
To reconstruct a surface at a specific query location at inference time, these methods then match the local reconstruction target by searching for the best match in the local prior space (by optimizing the parameters encoding the local context) at the given query location.
2 code implementations • CVPR 2022 • Baorui Ma, Yu-Shen Liu, Zhizhong Han
Our key idea is to infer signed distances by pushing both the query projections to be on the surface and the projection distance to be the minimum.
1 code implementation • CVPR 2022 • Xin Wen, Junsheng Zhou, Yu-Shen Liu, Zhen Dong, Zhizhong Han
Reconstructing 3D shape from a single 2D image is a challenging task, which needs to estimate the detailed 3D structures based on the semantic attributes from 2D image.
1 code implementation • 26 Mar 2022 • Junsheng Zhou, Xin Wen, Baorui Ma, Yu-Shen Liu, Yue Gao, Yi Fang, Zhizhong Han
To address this problem, we present a novel and efficient self-supervised point cloud representation learning framework, named 3D Occlusion Auto-Encoder (3D-OAE), to facilitate the detailed supervision inherited in local regions and global shapes.
1 code implementation • CVPR 2022 • Tianyang Li, Xin Wen, Yu-Shen Liu, Hua Su, Zhizhong Han
However, the local codes are constrained at discrete and regular positions like grid points, which makes the code positions difficult to be optimized and limits their representation ability.
no code implementations • 4 Mar 2022 • Han Liu, Xiaoyu Song, Ge Gao, Hehua Zhang, Yu-Shen Liu, Ming Gu
Semantic rule checking on RDFS/OWL data has been widely used in the construction industry.
1 code implementation • 19 Feb 2022 • Xin Wen, Peng Xiang, Zhizhong Han, Yan-Pei Cao, Pengfei Wan, Wen Zheng, Yu-Shen Liu
It moves each point of incomplete input to obtain a complete point cloud, where total distance of point moving paths (PMPs) should be the shortest.
Ranked #1 on Point Cloud Completion on Completion3D
1 code implementation • 18 Feb 2022 • Peng Xiang, Xin Wen, Yu-Shen Liu, Yan-Pei Cao, Pengfei Wan, Wen Zheng, Zhizhong Han
Our insight into the detailed geometry is to introduce a skip-transformer in the SPD to learn the point splitting patterns that can best fit the local regions.
Ranked #5 on Point Cloud Completion on ShapeNet
2 code implementations • 22 Dec 2021 • Liang Pan, Tong Wu, Zhongang Cai, Ziwei Liu, Xumin Yu, Yongming Rao, Jiwen Lu, Jie zhou, Mingye Xu, Xiaoyuan Luo, Kexue Fu, Peng Gao, Manning Wang, Yali Wang, Yu Qiao, Junsheng Zhou, Xin Wen, Peng Xiang, Yu-Shen Liu, Zhizhong Han, Yuanjie Yan, Junyi An, Lifa Zhu, Changwei Lin, Dongrui Liu, Xin Li, Francisco Gómez-Fernández, Qinlong Wang, Yang Yang
Based on the MVP dataset, this paper reports methods and results in the Multi-View Partial Point Cloud Challenge 2021 on Completion and Registration.
2 code implementations • 23 Nov 2021 • Zhen Cao, Wenxiao Zhang, Xin Wen, Zhen Dong, Yu-Shen Liu, Xiongwu Xiao, Bisheng Yang
The student network takes the incomplete one as input and restores the corresponding complete shape.
2 code implementations • ICCV 2021 • Peng Xiang, Xin Wen, Yu-Shen Liu, Yan-Pei Cao, Pengfei Wan, Wen Zheng, Zhizhong Han
However, previous methods usually suffered from discrete nature of point cloud and unstructured prediction of points in local regions, which makes it hard to reveal fine local geometric details on the complete shape.
1 code implementation • 8 Aug 2021 • Chen Chao, Zhizhong Han, Yu-Shen Liu, Matthias Zwicker
Our method pushes the neural network to generate a 3D point cloud whose 2D projections match the irregular point supervision from different view angles.
no code implementations • 8 Aug 2021 • Zhizhong Han, Xiyang Wang, Yu-Shen Liu, Matthias Zwicker
To mine highly discriminative information from unordered views, HVP performs a novel hierarchical view prediction over a view pair, and aggregates the knowledge learned from the predictions in all view pairs into a global feature.
1 code implementation • CVPR 2021 • Xin Wen, Zhizhong Han, Yan-Pei Cao, Pengfei Wan, Wen Zheng, Yu-Shen Liu
We provide a comprehensive evaluation in experiments, which shows that our model with the learned bidirectional geometry correspondence outperforms state-of-the-art unpaired completion methods.
1 code implementation • ICCV 2021 • Chao Chen, Zhizhong Han, Yu-Shen Liu, Matthias Zwicker
Our method pushes the neural network to generate a 3D point cloud whose 2D projections match the irregular point supervision from different view angles.
1 code implementation • 8 Dec 2020 • Xinhai Liu, Xinchen Liu, Yu-Shen Liu, Zhizhong Han
The task of point cloud upsampling aims to acquire dense and uniform point sets from sparse and irregular point sets.
1 code implementation • CVPR 2021 • Xin Wen, Peng Xiang, Zhizhong Han, Yan-Pei Cao, Pengfei Wan, Wen Zheng, Yu-Shen Liu
As a result, the network learns a strict and unique correspondence on point-level, which can capture the detailed topology and structure relationships between the incomplete shape and the complete target, and thus improves the quality of the predicted complete shape.
2 code implementations • 26 Nov 2020 • Baorui Ma, Zhizhong Han, Yu-Shen Liu, Matthias Zwicker
Specifically, we train a neural network to pull query 3D locations to their closest points on the surface using the predicted signed distance values and the gradient at the query locations, both of which are computed by the network itself.
no code implementations • ICML 2020 • Zhizhong Han, Chao Chen, Yu-Shen Liu, Matthias Zwicker
To optimize 3D shape parameters, current renderers rely on pixel-wise losses between rendered images of 3D reconstructions and ground truth images from corresponding viewpoints.
1 code implementation • 26 May 2020 • Xinhai Liu, Zhizhong Han, Yu-Shen Liu, Matthias Zwicker
According to our experiments under this fine-grained dataset, we find that state-of-the-art methods are significantly limited by the small variance among subcategories in the same category.
no code implementations • CVPR 2020 • Xin Wen, Tianyang Li, Zhizhong Han, Yu-Shen Liu
Point cloud completion aims to infer the complete geometries for missing regions of 3D objects from incomplete ones.
no code implementations • 18 Mar 2020 • Xinhai Liu, Zhizhong Han, Fangzhou Hong, Yu-Shen Liu, Matthias Zwicker
However, due to the irregularity and sparsity in sampled point clouds, it is hard to encode the fine-grained geometry of local regions and their spatial relationships when only using the fixed-size filters and individual local feature integration, which limit the ability to learn discriminative features.
no code implementations • ECCV 2020 • Zhizhong Han, Guanhui Qiao, Yu-Shen Liu, Matthias Zwicker
To avoid dense and irregular sampling in 3D, we propose to represent shapes using 2D functions, where the output of the function at each 2D location is a sequence of line segments inside the shape.
1 code implementation • NeurIPS 2019 • Han Liu, Zhizhong Han, Yu-Shen Liu, Ming Gu
Low-rank metric learning aims to learn better discrimination of data subject to low-rank constraints.
no code implementations • 29 Aug 2019 • Xin Wen, Zhizhong Han, Xinhai Liu, Yu-Shen Liu
Compared to the previous capsule network based methods, the feature routing on the spatial-aware capsules can learn more discriminative spatial relationships among local regions for point clouds, which establishes a direct mapping between log priors and the spatial locations through feature clusters.
no code implementations • 2 Aug 2019 • Xinhai Liu, Zhizhong Han, Xin Wen, Yu-Shen Liu, Matthias Zwicker
Specifically, L2G-AE employs an encoder to encode the geometry information of multiple scales in a local region at the same time.
no code implementations • 31 Jul 2019 • Zhizhong Han, Chao Chen, Yu-Shen Liu, Matthias Zwicker
Specifically, ShapeCaptioner aggregates the parts detected in multiple colored views using our novel part class specific aggregation to represent a 3D shape, and then, employs a sequence to sequence model to generate the caption.
no code implementations • ICCV 2019 • Zhizhong Han, Xiyang Wang, Yu-Shen Liu, Matthias Zwicker
To resolve this issue, we propose MAP-VAE to enable the learning of global and local geometry by jointly leveraging global and local self-supervision.
Ranked #16 on 3D Point Cloud Linear Classification on ModelNet40
3D Point Cloud Linear Classification Unsupervised 3D Point Cloud Linear Evaluation
no code implementations • 18 May 2019 • Zhizhong Han, Xinhai Liu, Yu-Shen Liu, Matthias Zwicker
In contrast, we propose a deep neural network, called Parts4Feature, to learn 3D global features from part-level information in multiple views.
no code implementations • 17 May 2019 • Zhizhong Han, Xiyang Wang, Chi-Man Vong, Yu-Shen Liu, Matthias Zwicker, C. L. Philip Chen
Then, the content and spatial information of each pair of view nodes are encoded by a novel spatial pattern correlation, where the correlation is computed among latent semantic patterns.
no code implementations • 17 Apr 2019 • Xin Wen, Zhizhong Han, Xinyu Yin, Yu-Shen Liu
Cross-modal retrieval aims to retrieve relevant data across different modalities (e. g., texts vs. images).
no code implementations • 7 Nov 2018 • Zhizhong Han, Mingyang Shang, Xiyang Wang, Yu-Shen Liu, Matthias Zwicker
A recent method employs 3D voxels to represent 3D shapes, but this limits the approach to low resolutions due to the computational cost caused by the cubic complexity of 3D voxels.
no code implementations • 7 Nov 2018 • Zhizhong Han, Mingyang Shang, Yu-Shen Liu, Matthias Zwicker
Intuitively, this memory enables the system to aggregate information that is useful to better solve the view inter-prediction tasks for each shape, and to leverage the memory as a view-independent shape representation.
Ranked #14 on 3D Point Cloud Linear Classification on ModelNet40
3D Point Cloud Linear Classification Representation Learning
no code implementations • 6 Nov 2018 • Xinhai Liu, Zhizhong Han, Yu-Shen Liu, Matthias Zwicker
However, it is hard to capture fine-grained contextual information in hand-crafted or explicit manners, such as the correlation between different areas in a local region, which limits the discriminative ability of learned features.
Ranked #49 on 3D Part Segmentation on ShapeNet-Part