1 code implementation • 14 Oct 2024 • Yue Zhang, Minhao Liu, Zhaokang Chen, Bin Wu, Yubin Zeng, Chao Zhan, Yingjie He, Junxin Huang, Wenjiang Zhou
We propose MuseTalk, which generates lip-sync targets in a latent space encoded by a Variational Autoencoder, enabling high-fidelity talking face video generation with efficient inference.
3 code implementations • 10 Feb 2020 • Didan Deng, Zhaokang Chen, Bertram E. Shi
We use the soft labels and the ground truth to train the student model.
2 code implementations • 25 Jan 2020 • Zhaokang Chen, Bertram E. Shi
Appearance-based gaze estimation from RGB images provides relatively unconstrained gaze tracking.
2 code implementations • 21 Nov 2019 • Didan Deng, Zhaokang Chen, Yuqian Zhou, Bertram Shi
Spatial-temporal feature learning is of vital importance for video emotion recognition.
no code implementations • 11 May 2019 • Zhaokang Chen, Bertram E. Shi
To improve estimation, we propose a novel gaze decomposition method and a single gaze point calibration method, motivated by our finding that the inter-subject squared bias exceeds the intra-subject variance for a subject-independent estimator.
no code implementations • 18 Mar 2019 • Zhaokang Chen, Bertram E. Shi
Appearance-based gaze estimation has attracted more and more attention because of its wide range of applications.