Search Results for author: GuoJun Qi

Found 6 papers, 2 papers with code

DiffSpeaker: Speech-Driven 3D Facial Animation with Diffusion Transformer

1 code implementation8 Feb 2024 Zhiyuan Ma, Xiangyu Zhu, GuoJun Qi, Chen Qian, Zhaoxiang Zhang, Zhen Lei

We suspect this is due to a shortage of paired audio-4D data, which is crucial for the Transformer to effectively perform as a denoiser within the Diffusion framework.

Lightweight high-resolution Subject Matting in the Real World

no code implementations12 Dec 2023 Peng Liu, Fanyi Wang, Jingwen Su, Yanhao Zhang, GuoJun Qi

To alleviate these issues, we propose to construct a saliency object matting dataset HRSOM and a lightweight network PSUNet.

Image Matting object-detection +1

LatentAvatar: Learning Latent Expression Code for Expressive Neural Head Avatar

no code implementations2 May 2023 Yuelang Xu, Hongwen Zhang, Lizhen Wang, Xiaochen Zhao, Han Huang, GuoJun Qi, Yebin Liu

Existing approaches to animatable NeRF-based head avatars are either built upon face templates or use the expression coefficients of templates as the driving signal.

OTAvatar: One-shot Talking Face Avatar with Controllable Tri-plane Rendering

1 code implementation CVPR 2023 Zhiyuan Ma, Xiangyu Zhu, GuoJun Qi, Zhen Lei, Lei Zhang

In this paper, we propose One-shot Talking face Avatar (OTAvatar), which constructs face avatars by a generalized controllable tri-plane rendering solution so that each personalized avatar can be constructed from only one portrait as the reference.

Causal Attention for Vision-Language Tasks

no code implementations CVPR 2021 Xu Yang, Hanwang Zhang, GuoJun Qi, Jianfei Cai

Specifically, CATT is implemented as a combination of 1) In-Sample Attention (IS-ATT) and 2) Cross-Sample Attention (CS-ATT), where the latter forcibly brings other samples into every IS-ATT, mimicking the causal intervention.

Cannot find the paper you are looking for? You can Submit a new open access paper.