Search Results for author: Longhao Zhang

Found 7 papers, 2 papers with code

DreamActor-M1: Holistic, Expressive and Robust Human Image Animation with Hybrid Guidance

no code implementations2 Apr 2025 Yuxuan Luo, Zhengkun Rong, Lizhen Wang, Longhao Zhang, Tianshu Hu, Yongming Zhu

For motion guidance, our hybrid control signals that integrate implicit facial representations, 3D head spheres, and 3D body skeletons achieve robust control of facial expressions and body movements, while producing expressive and identity-preserving animations.

Human Animation Image Animation +1

INFP: Audio-Driven Interactive Head Generation in Dyadic Conversations

no code implementations5 Dec 2024 Yongming Zhu, Longhao Zhang, Zhengkun Rong, Tianshu Hu, Shuang Liang, Zhipeng Ge

The second stage learns the mapping from the input dyadic audio to motion latent codes through denoising, leading to the audio-driven head generation in interactive scenarios.

Denoising Motion Generation

PersonaTalk: Bring Attention to Your Persona in Visual Dubbing

no code implementations9 Sep 2024 Longhao Zhang, Shuang Liang, Zhipeng Ge, Tianshu Hu

In this paper, we present PersonaTalk, an attention-based two-stage framework, including geometry construction and face rendering, for high-fidelity and personalized visual dubbing.

MaTe3D: Mask-guided Text-based 3D-aware Portrait Editing

1 code implementation12 Dec 2023 Kangneng Zhou, Daiheng Gao, Xuan Wang, Jie Zhang, Peng Zhang, Xusen Sun, Longhao Zhang, Shiqi Yang, Bang Zhang, Liefeng Bo, Yaxing Wang, Ming-Ming Cheng

This enhances masked-based editing in local areas; second, we present a novel distillation strategy: Conditional Distillation on Geometry and Texture (CDGT).

VividTalk: One-Shot Audio-Driven Talking Head Generation Based on 3D Hybrid Prior

no code implementations4 Dec 2023 Xusen Sun, Longhao Zhang, Hao Zhu, Peng Zhang, Bang Zhang, Xinya Ji, Kangneng Zhou, Daiheng Gao, Liefeng Bo, Xun Cao

Audio-driven talking head generation has drawn much attention in recent years, and many efforts have been made in lip-sync, expressive facial expressions, natural head pose generation, and high video quality.

Talking Head Generation

Depth-Aware Generative Adversarial Network for Talking Head Video Generation

1 code implementation CVPR 2022 Fa-Ting Hong, Longhao Zhang, Li Shen, Dan Xu

In a more dense way, the depth is also utilized to learn 3D-aware cross-modal (i. e. appearance and depth) attention to guide the generation of motion fields for warping source image representations.

3D geometry Generative Adversarial Network +2

Cannot find the paper you are looking for? You can Submit a new open access paper.