Search Results for author: Lizhen Wang

Found 19 papers, 10 papers with code

DreamActor-H1: High-Fidelity Human-Product Demonstration Video Generation via Motion-designed Diffusion Transformers

no code implementations12 Jun 2025 Lizhen Wang, Zhurong Xia, Tianshu Hu, Pengrui Wang, Pengfei Wang, Zerong Zheng, Ming Zhou

Trained on a hybrid dataset with extensive data augmentation strategies, our approach outperforms state-of-the-art techniques in maintaining the identity integrity of both humans and products and generating realistic demonstration motions.

Data Augmentation Marketing +1

DreamActor-M1: Holistic, Expressive and Robust Human Image Animation with Hybrid Guidance

no code implementations2 Apr 2025 Yuxuan Luo, Zhengkun Rong, Lizhen Wang, Longhao Zhang, Tianshu Hu, Yongming Zhu

For motion guidance, our hybrid control signals that integrate implicit facial representations, 3D head spheres, and 3D body skeletons achieve robust control of facial expressions and body movements, while producing expressive and identity-preserving animations.

Human Animation Image Animation +1

Stereo-Talker: Audio-driven 3D Human Synthesis with Prior-Guided Mixture-of-Experts

no code implementations31 Oct 2024 Xiang Deng, Youxin Pang, Xiaochen Zhao, Chao Xu, Lizhen Wang, Hongjiang Xiao, Shi Yan, Hongwen Zhang, Yebin Liu

This paper introduces Stereo-Talker, a novel one-shot audio-driven human video synthesis system that generates 3D talking videos with precise lip synchronization, expressive body gestures, temporally consistent photo-realistic quality, and continuous viewpoint control.

Language Modeling Language Modelling +3

RAM-Avatar: Real-time Photo-Realistic Avatar from Monocular Videos with Full-body Control

no code implementations CVPR 2024 Xiang Deng, Zerong Zheng, Yuxiang Zhang, Jingxiang Sun, Chao Xu, Xiaodong Yang, Lizhen Wang, Yebin Liu

This paper focuses on advancing the applicability of human avatar learning methods by proposing RAM-Avatar which learns a Real-time photo-realistic Avatar that supports full-body control from Monocular videos.

GMTalker: Gaussian Mixture-based Audio-Driven Emotional Talking Video Portraits

no code implementations12 Dec 2023 Yibo Xia, Lizhen Wang, Xiang Deng, Xiaoyan Luo, Yunhong Wang, Yebin Liu

Finally, we propose a personalized emotion-guided head generator with an emotion mapping network that can synthesize high-fidelity and faithful emotional video portraits.

Diversity

MonoGaussianAvatar: Monocular Gaussian Point-based Head Avatar

no code implementations7 Dec 2023 Yufan Chen, Lizhen Wang, Qijing Li, Hongjiang Xiao, Shengping Zhang, Hongxun Yao, Yebin Liu

In response to these challenges, we propose MonoGaussianAvatar (Monocular Gaussian Point-based Head Avatar), a novel approach that harnesses 3D Gaussian point representation coupled with a Gaussian deformation field to learn explicit head avatars from monocular portrait videos.

InvertAvatar: Incremental GAN Inversion for Generalized Head Avatars

1 code implementation3 Dec 2023 Xiaochen Zhao, Jingxiang Sun, Lizhen Wang, Jinli Suo, Yebin Liu

While high fidelity and efficiency are central to the creation of digital head avatars, recent methods relying on 2D or 3D generative models often experience limitations such as shape distortion, expression inaccuracy, and identity flickering.

Image-to-Image Translation

Animatable and Relightable Gaussians for High-fidelity Human Avatar Modeling

1 code implementation27 Nov 2023 Zhe Li, Yipengjing Sun, Zerong Zheng, Lizhen Wang, Shengping Zhang, Yebin Liu

To associate 3D Gaussians with the animatable avatar, we learn a parametric template from the input videos, and then parameterize the template on two front & back canonical Gaussian maps where each pixel represents a 3D Gaussian.

NeRF

DreamCraft3D: Hierarchical 3D Generation with Bootstrapped Diffusion Prior

1 code implementation25 Oct 2023 Jingxiang Sun, Bo Zhang, Ruizhi Shao, Lizhen Wang, Wen Liu, Zhenda Xie, Yebin Liu

The score distillation from this 3D-aware diffusion prior provides view-consistent guidance for the scene.

3D Generation

HAvatar: High-fidelity Head Avatar via Facial Model Conditioned Neural Radiance Field

no code implementations29 Sep 2023 Xiaochen Zhao, Lizhen Wang, Jingxiang Sun, Hongwen Zhang, Jinli Suo, Yebin Liu

The problem of modeling an animatable 3D human head avatar under light-weight setups is of significant importance but has not been well solved.

Image-to-Image Translation NeRF

LatentAvatar: Learning Latent Expression Code for Expressive Neural Head Avatar

no code implementations2 May 2023 Yuelang Xu, Hongwen Zhang, Lizhen Wang, Xiaochen Zhao, Han Huang, GuoJun Qi, Yebin Liu

Existing approaches to animatable NeRF-based head avatars are either built upon face templates or use the expression coefficients of templates as the driving signal.

NeRF

StyleAvatar: Real-time Photo-realistic Portrait Avatar from a Single Video

1 code implementation1 May 2023 Lizhen Wang, Xiaochen Zhao, Jingxiang Sun, Yuxiang Zhang, Hongwen Zhang, Tao Yu, Yebin Liu

Results and experiments demonstrate the superiority of our method in terms of image quality, full portrait video generation, and real-time re-animation compared to existing facial reenactment methods.

Face Reenactment Translation +1

AvatarMAV: Fast 3D Head Avatar Reconstruction Using Motion-Aware Neural Voxels

no code implementations23 Nov 2022 Yuelang Xu, Lizhen Wang, Xiaochen Zhao, Hongwen Zhang, Yebin Liu

AvatarMAV is the first to model both the canonical appearance and the decoupled expression motion by neural voxels for head avatar.

NeRF

Next3D: Generative Neural Texture Rasterization for 3D-Aware Head Avatars

2 code implementations CVPR 2023 Jingxiang Sun, Xuan Wang, Lizhen Wang, Xiaoyu Li, Yong Zhang, Hongwen Zhang, Yebin Liu

We propose a novel 3D GAN framework for unsupervised learning of generative, high-quality and 3D-consistent facial avatars from unstructured 2D images.

Face Model

IDE-3D: Interactive Disentangled Editing for High-Resolution 3D-aware Portrait Synthesis

1 code implementation31 May 2022 Jingxiang Sun, Xuan Wang, Yichun Shi, Lizhen Wang, Jue Wang, Yebin Liu

Existing 3D-aware facial generation methods face a dilemma in quality versus editability: they either generate editable results in low resolution or high-quality ones with no editing flexibility.

3D-Aware Image Synthesis

FaceVerse: a Fine-grained and Detail-controllable 3D Face Morphable Model from a Hybrid Dataset

1 code implementation CVPR 2022 Lizhen Wang, ZhiYuan Chen, Tao Yu, Chenguang Ma, Liang Li, Yebin Liu

In the coarse module, we generate a base parametric model from large-scale RGB-D images, which is able to predict accurate rough 3D face models in different genders, ages, etc.

2k 3D Face Reconstruction +1

NormalGAN: Learning Detailed 3D Human from a Single RGB-D Image

1 code implementation ECCV 2020 Lizhen Wang, Xiaochen Zhao, Tao Yu, Songtao Wang, Yebin Liu

We propose NormalGAN, a fast adversarial learning-based method to reconstruct the complete and detailed 3D human from a single RGB-D image.

3D Human Reconstruction Denoising

DDRNet: Depth Map Denoising and Refinement for Consumer Depth Cameras Using Cascaded CNNs

3 code implementations ECCV 2018 Shi Yan, Chenglei Wu, Lizhen Wang, Feng Xu, Liang An, Kaiwen Guo, Yebin Liu

Consumer depth sensors are more and more popular and come to our daily lives marked by its recent integration in the latest Iphone X.

Denoising

Cannot find the paper you are looking for? You can Submit a new open access paper.