1 code implementation • 11 Jun 2024 • Payal Mohapatra, Shamika Likhite, Subrata Biswas, Bashima Islam, Qi Zhu
In experiments across five disfluency-detection tasks, our unified multimodal approach significantly outperforms Audio-only unimodal methods, yielding an average absolute improvement of 10% (i. e., 10 percentage point increase) when both video and audio modalities are always available, and 7% even when video modality is missing in half of the samples.
no code implementations • 5 Feb 2024 • Payal Mohapatra, Lixu Wang, Qi Zhu
Monitoring and recognizing patterns in continuous sensing data is crucial for many practical applications.
1 code implementation • 28 Aug 2023 • Payal Mohapatra, Akash Pandey, Yueyuan Sui, Qi Zhu
different share of people perceive the same speech segment as a non-unanimous emotion.