Search Results for author: Manyi Li

Found 11 papers, 4 papers with code

CLIP-based Point Cloud Classification via Point Cloud to Image Translation

no code implementations7 Aug 2024 Shuvozit Ghose, Manyi Li, Yiming Qian, Yang Wang

Motivated by this observation, we propose a Pretrained Point Cloud to Image Translation Network (PPCITNet) that produces generalized colored images along with additional salient visual cues to the point cloud depth maps so that it can achieve promising performance on point cloud classification and understanding.

Classification Point Cloud Classification +1

Varying Manifolds in Diffusion: From Time-varying Geometries to Visual Saliency

no code implementations7 Jun 2024 JunHao Chen, Manyi Li, Zherong Pan, Xifeng Gao, Changhe Tu

Our key contribution is the introduction of generation rate, which corresponds to the local deformation of manifold over time around an image component.

Image Manipulation

Learning based 2D Irregular Shape Packing

no code implementations19 Sep 2023 Zeshi Yang, Zherong Pan, Manyi Li, Kui Wu, Xifeng Gao

2D irregular shape packing is a necessary step to arrange UV patches of a 3D model within a texture atlas for memory-efficient appearance rendering in computer graphics.

Decision Making

AffordPose: A Large-scale Dataset of Hand-Object Interactions with Affordance-driven Hand Pose

1 code implementation ICCV 2023 Juntao Jian, Xiuping Liu, Manyi Li, Ruizhen Hu, Jian Liu

We collect a total of 26. 7K hand-object interactions, each including the 3D object shape, the part-level affordance label, and the manually adjusted hand poses.

Diversity Object

Laplacian2Mesh: Laplacian-Based Mesh Understanding

1 code implementation1 Feb 2022 Qiujie Dong, Zixiong Wang, Manyi Li, Junjie Gao, Shuangmin Chen, Zhenyu Shu, Shiqing Xin, Changhe Tu, Wenping Wang

Geometric deep learning has sparked a rising interest in computer graphics to perform shape understanding tasks, such as shape classification and semantic segmentation.

Semantic Segmentation Surface Reconstruction

RIM-Net: Recursive Implicit Fields for Unsupervised Learning of Hierarchical Shape Structures

1 code implementation CVPR 2022 Chengjie Niu, Manyi Li, Kai Xu, Hao Zhang

Each level of the tree corresponds to an assembly of shape parts, represented as implicit functions, to reconstruct the input shape.

Decoder

D2IM-Net: Learning Detail Disentangled Implicit Fields From Single Images

no code implementations CVPR 2021 Manyi Li, Hao Zhang

We present the first single-view 3D reconstruction network aimed at recovering geometric details from an input image which encompass both topological shape structures and surface features.

3D Reconstruction Decoder +1

CAPRI-Net: Learning Compact CAD Shapes with Adaptive Primitive Assembly

no code implementations CVPR 2022 Fenggen Yu, Zhiqin Chen, Manyi Li, Aditya Sanghi, Hooman Shayani, Ali Mahdavi-Amiri, Hao Zhang

We introduce CAPRI-Net, a neural network for learning compact and interpretable implicit representations of 3D computer-aided design (CAD) models, in the form of adaptive primitive assemblies.

CAD Reconstruction

LayoutGMN: Neural Graph Matching for Structural Layout Similarity

1 code implementation CVPR 2021 Akshay Gadi Patil, Manyi Li, Matthew Fisher, Manolis Savva, Hao Zhang

In particular, retrieval results by our network better match human judgement of structural layout similarity compared to both IoUs and other baselines including a state-of-the-art method based on graph neural networks and image convolution.

Graph Matching Metric Learning +2

D$^2$IM-Net: Learning Detail Disentangled Implicit Fields from Single Images

no code implementations11 Dec 2020 Manyi Li, Hao Zhang

We present the first single-view 3D reconstruction network aimed at recovering geometric details from an input image which encompass both topological shape structures and surface features.

3D Reconstruction Decoder +1

GRAINS: Generative Recursive Autoencoders for INdoor Scenes

no code implementations24 Jul 2018 Manyi Li, Akshay Gadi Patil, Kai Xu, Siddhartha Chaudhuri, Owais Khan, Ariel Shamir, Changhe Tu, Baoquan Chen, Daniel Cohen-Or, Hao Zhang

We present a generative neural network which enables us to generate plausible 3D indoor scenes in large quantities and varieties, easily and highly efficiently.

Graphics

Cannot find the paper you are looking for? You can Submit a new open access paper.