1 code implementation • 14 Oct 2024 • Yu Wu, Ting Dang, Dimitris Spathis, Hong Jia, Cecilia Mascolo
However, existing CL methods often introduce false negative pairs (FNPs) by neglecting inherent characteristics and then randomly selecting distinct segments as dissimilar pairs, leading to erroneous representation learning, reduced model performance, and overall inefficiency.
no code implementations • 26 Sep 2024 • Xin Hong, Yuan Gong, Vidhyasaharan Sethu, Ting Dang
Recent advancements in Large Language Models (LLMs) have demonstrated great success in many Natural Language Processing (NLP) tasks.
no code implementations • 17 Sep 2024 • Xin Wang, Ting Dang, Vassilis Kostakos, Hong Jia
Healthcare monitoring is crucial for early detection, timely intervention, and the ongoing management of health conditions, ultimately improving individuals' quality of life.
no code implementations • 17 Sep 2024 • Zheng Nan, Ting Dang, Vidhyasaharan Sethu, Beena Ahmed
Despite the crucial role relational thinking plays in human understanding of speech, it has yet to be leveraged in any artificial speech recognition systems.
1 code implementation • 31 Jul 2024 • Jingyao Wu, Ting Dang, Vidhyasaharan Sethu, Eliathamby Ambikairajah
There has been a significant focus on modelling emotion ambiguity in recent years, with advancements made in representing emotions as distributions to capture ambiguity.
1 code implementation • 7 Mar 2024 • Xiaoyu Tang, Yixin Lin, Ting Dang, Yuanfang Zhang, Jintao Cheng
In this paper, to model local and global information at different levels of granularity in speech and capture temporal, spatial and channel dependencies in speech signals, we propose a Speech Emotion Recognition network based on CNN-Transformer and multi-dimensional attention mechanisms.
no code implementations • 21 Sep 2023 • Zheng Nan, Ting Dang, Vidhyasaharan Sethu, Beena Ahmed
Connectionist temporal classification (CTC) is commonly adopted for sequence modeling tasks like speech recognition, where it is necessary to preserve order between the input and target sequences.
no code implementations • 4 Jan 2022 • Ting Dang, Jing Han, Tong Xia, Dimitris Spathis, Erika Bondareva, Chloë Siegele-Brown, Jagmohan Chauhan, Andreas Grammenos, Apinan Hasthanasombat, Andres Floto, Pietro Cicuta, Cecilia Mascolo
Recent work has shown the potential of using audio data (eg, cough, breathing, and voice) in the screening for COVID-19.
no code implementations • 10 Aug 2021 • Jingyao Wu, Ting Dang, Vidhyasaharan Sethu, Eliathamby Ambikairajah
We propose a Markovian framework referred to as Dynamic Ordinal Markov Model (DOMM) that makes use of both absolute and relative ordinal information, to improve speech based ordinal emotion prediction.
no code implementations • 29 Jun 2021 • Jing Han, Tong Xia, Dimitris Spathis, Erika Bondareva, Chloë Brown, Jagmohan Chauhan, Ting Dang, Andreas Grammenos, Apinan Hasthanasombat, Andres Floto, Pietro Cicuta, Cecilia Mascolo
In this paper, we explore the realistic performance of audio-based digital testing of COVID-19.
no code implementations • 5 Apr 2021 • Tong Xia, Jing Han, Lorena Qendro, Ting Dang, Cecilia Mascolo
To handle these issues, we propose an ensemble framework where multiple deep learning models for sound-based COVID-19 detection are developed from different but balanced subsets from original data.