Search Results for author: Gengshan Yang

Found 15 papers, 12 papers with code

Reconstructing Animatable Categories from Videos

2 code implementations CVPR 2023 Gengshan Yang, Chaoyang Wang, N Dinesh Reddy, Deva Ramanan

Building animatable 3D models is challenging due to the need for 3D scans, laborious registration, and manual rigging, which are difficult to scale to arbitrary categories.

3D Shape Reconstruction from Videos Dynamic Reconstruction +1

Total-Recon: Deformable Scene Reconstruction for Embodied View Synthesis

no code implementations ICCV 2023 Chonghyuk Song, Gengshan Yang, Kangle Deng, Jun-Yan Zhu, Deva Ramanan

Given a minute-long RGBD video of people interacting with their pets, we render the scene from novel camera trajectories derived from the in-scene motion of actors: (1) egocentric cameras that simulate the point of view of a target actor and (2) 3rd-person cameras that follow the actor.

3D-aware Conditional Image Synthesis

2 code implementations CVPR 2023 Kangle Deng, Gengshan Yang, Deva Ramanan, Jun-Yan Zhu

We propose pix2pix3D, a 3D-aware conditional generative model for controllable photorealistic image synthesis.

Image Generation

Distilling Neural Fields for Real-Time Articulated Shape Reconstruction

no code implementations CVPR 2023 Jeff Tan, Gengshan Yang, Deva Ramanan

We present a method for reconstructing articulated 3D models from videos in real-time, without test-time optimization or manual 3D supervision at training time.

motion prediction

PPR: Physically Plausible Reconstruction from Monocular Videos

no code implementations ICCV 2023 Gengshan Yang, Shuo Yang, John Z. Zhang, Zachary Manchester, Deva Ramanan

Given monocular videos, we build 3D models of articulated objects and environments whose 3D configurations satisfy dynamics and contact constraints.

BANMo: Building Animatable 3D Neural Models from Many Casual Videos

1 code implementation CVPR 2022 Gengshan Yang, Minh Vo, Natalia Neverova, Deva Ramanan, Andrea Vedaldi, Hanbyul Joo

Our key insight is to merge three schools of thought; (1) classic deformable shape models that make use of articulated bones and blend skinning, (2) volumetric neural radiance fields (NeRFs) that are amenable to gradient-based optimization, and (3) canonical embeddings that generate correspondences between pixels and an articulated model.

3D Shape Reconstruction from Videos Dynamic Reconstruction

ViSER: Video-Specific Surface Embeddings for Articulated 3D Shape Reconstruction

1 code implementation NeurIPS 2021 Gengshan Yang, Deqing Sun, Varun Jampani, Daniel Vlasic, Forrester Cole, Ce Liu, Deva Ramanan

The surface embeddings are implemented as coordinate-based MLPs that are fit to each video via consistency and contrastive reconstruction losses. Experimental results show that ViSER compares favorably against prior work on challenging videos of humans with loose clothing and unusual poses as well as animals videos from DAVIS and YTVOS.

3D Shape Reconstruction from Videos

NeRS: Neural Reflectance Surfaces for Sparse-view 3D Reconstruction in the Wild

1 code implementation NeurIPS 2021 Jason Y. Zhang, Gengshan Yang, Shubham Tulsiani, Deva Ramanan

NeRS learns a neural shape representation of a closed surface that is diffeomorphic to a sphere, guaranteeing water-tight reconstructions.

3D Reconstruction Neural Rendering

Learning to Segment Rigid Motions from Two Frames

1 code implementation CVPR 2021 Gengshan Yang, Deva Ramanan

Geometric motion segmentation algorithms, however, generalize to novel scenes, but have yet to achieve comparable performance to appearance-based ones, due to noisy motion estimations and degenerate motion configurations.

Motion Segmentation Scene Flow Estimation +2

Volumetric Correspondence Networks for Optical Flow

2 code implementations NeurIPS 2019 Gengshan Yang, Deva Ramanan

As a result, SOTA networks also employ various heuristics designed to limit volumetric processing, leading to limited accuracy and overfitting.

Optical Flow Estimation

Cannot find the paper you are looking for? You can Submit a new open access paper.