no code implementations • 11 Jan 2024 • Juni Kim, Zhikang Dong, Pawel Polak
We introduce a novel method that combines differential geometry, kernels smoothing, and spectral analysis to quantify facial muscle activity from widely accessible video recordings, such as those captured on personal smartphones.
1 code implementation • 10 Oct 2023 • Xiulong Liu, Zhikang Dong, Peng Zhang
In recent years, there has been a growing emphasis on the intersection of audio, vision, and text modalities, driving forward the advancements in multimodal research.
no code implementations • 10 Oct 2023 • Zhikang Dong, Bin Chen, Xiulong Liu, Pawel Polak, Peng Zhang
The reasoning module, equipped with the power of Large Language Model (Vicuna-7B) and extended to multi-modal inputs, is able to provide reasonable explanation for the recommended music.
no code implementations • 1 Nov 2022 • Juni Kim, Zhikang Dong, Eric Guan, Judah Rosenthal, Shi Fu, Miriam Rafailovich, Pawel Polak
Although the original FAN model achieves very high out-of-sample performance on the original CK++ videos, it does not perform so well on hidden emotions videos.
no code implementations • 18 Aug 2022 • Zhikang Dong, Pawel Polak
However, when changepoints are present, our approach yields superior parameter estimation, improved model fitting, and reduced training error compared to the original PINNs model.