no code implementations • 4 Feb 2024 • Alexandra Saliba, Yuanchao Li, Ramon Sanabria, Catherine Lai
Through a comparative experiment and a layer-wise accuracy analysis on two distinct corpora, IEMOCAP and ESD, we explore differences between AWEs and raw self-supervised representations, as well as the proper utilization of AWEs alone and in combination with word embeddings.
no code implementations • 25 May 2023 • Yuanchao Li, Zeyu Zhao, Ondrej Klejch, Peter Bell, Catherine Lai
To overcome this challenge, we investigate how Automatic Speech Recognition (ASR) performs on emotional speech by analyzing the ASR performance on emotion corpora and examining the distribution of word errors and confidence scores in ASR transcripts to gain insight into how emotion affects ASR.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 25 May 2023 • Yuanchao Li, Peter Bell, Catherine Lai
In this work, we investigate the relationship between two affective attributes: personality and emotion, from a transfer learning perspective.
no code implementations • 23 May 2023 • Yaoting Wang, Yuanchao Li, Paul Pu Liang, Louis-Philippe Morency, Peter Bell, Catherine Lai
Fusing multiple modalities has proven effective for multimodal information processing.
no code implementations • 5 Oct 2022 • Yuanchao Li, Yumnah Mohamied, Peter Bell, Catherine Lai
Self-supervised speech models have grown fast during the past few years and have proven feasible for use in various downstream tasks.
no code implementations • 25 Mar 2022 • Yuanchao Li, Catherine Lai
In this paper, we perform impression recognition using a proposed cross-domain architecture on the dyadic IMPRESSION dataset.
no code implementations • 17 Mar 2022 • Yuanchao Li, Catherine Lai
In recent years, many works have investigated the feasibility of conversational robots for performing specific tasks, such as healthcare and interview.
no code implementations • 29 Oct 2021 • Yuanchao Li, Peter Bell, Catherine Lai
However, due to the scarcity of emotion labelled data and the difficulty of recognizing emotional speech, it is hard to obtain reliable linguistic features and models in this research area.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2