no code implementations • 26 Feb 2024 • Tassadaq Hussain, Kia Dashtipour, Yu Tsao, Amir Hussain
By integrating emotional features, the proposed system achieves significant improvements in both objective and subjective assessments of speech quality and intelligibility, especially in challenging noise environments.
no code implementations • 15 Jul 2023 • Richard Lee Lai, Jen-Cheng Hou, Mandar Gogate, Kia Dashtipour, Amir Hussain, Yu Tsao
The aim of this study is to explore the effectiveness of audio-visual speech enhancement (AVSE) in enhancing the intelligibility of vocoded speech in cochlear implant (CI) simulations.
no code implementations • 24 Oct 2022 • Abhijeet Bishnu, Ankit Gupta, Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Amir Hussain, Mathini Sellathurai, Tharmalingam Ratnarajah
In this paper, we design a first of its kind transceiver (PHY layer) prototype for cloud-based audio-visual (AV) speech enhancement (SE) complying with high data rate and low latency requirements of future multimodal hearing assistive technology.
no code implementations • 11 Feb 2022 • Tassadaq Hussain, Muhammad Diyan, Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Yu Tsao, Amir Hussain
Current deep learning (DL) based approaches to speech intelligibility enhancement in noisy environments are often trained to minimise the feature distance between noise-free speech and enhanced speech signals.
no code implementations • 8 Feb 2022 • Tassadaq Hussain, Muhammad Diyan, Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Yu Tsao, Amir Hussain
Current deep learning (DL) based approaches to speech intelligibility enhancement in noisy environments are generally trained to minimise the distance between clean and enhanced speech features.
no code implementations • 24 Jan 2022 • Tassadaq Hussain, Wei-Chien Wang, Mandar Gogate, Kia Dashtipour, Yu Tsao, Xugang Lu, Adeel Ahsan, Amir Hussain
To address this problem, we propose to integrate a novel temporal attentive-pooling (TAP) mechanism into a conventional convolutional recurrent neural network, termed as TAP-CRNN.
no code implementations • 16 Dec 2021 • Mandar Gogate, Kia Dashtipour, Amir Hussain
The human brain contextually exploits heterogeneous sensory information to efficiently perform cognitive tasks including vision and hearing.
1 code implementation • 18 Nov 2021 • Tassadaq Hussain, Mandar Gogate, Kia Dashtipour, Amir Hussain
To the best of our knowledge, this is the first work that exploits the integration of AV modalities with an I-O based loss function for SE.
no code implementations • 3 Mar 2021 • Kia Dashtipour, Mandar Gogate, Erik Cambria, Amir Hussain
Most recent works on sentiment analysis have exploited the text modality.
no code implementations • 6 Aug 2020 • William Taylor, Syed Aziz Shah, Kia Dashtipour, Adnan Zahid, Qammer H. Abbasi, Muhammad Ali Imran
The dataset was used to create a machine learning model, which was used in a developed application to provide a quasi-real-time classification of standing or sitting state.
no code implementations • 30 Sep 2019 • Kia Dashtipour, Mandar Gogate, Jingpeng Li, Fengling Jiang, Bin Kong, Amir Hussain
When no pattern is triggered, the framework switches to its subsymbolic counterpart and leverages deep neural networks (DNN) to perform the classification.
no code implementations • 30 Sep 2019 • Mandar Gogate, Ahsan Adeel, Kia Dashtipour, Peter Derleth, Amir Hussain
This paper presents, a first of its kind, audio-visual (AV) speech enhacement challenge in real-noisy settings.
no code implementations • 23 Sep 2019 • Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Amir Hussain
In addition, our work challenges a popular belief that a scarcity of multi-language large vocabulary AV corpus and wide variety of noises is a major bottleneck to build a robust language, speaker and noise independent SE systems.
no code implementations • 15 Aug 2018 • Kia Dashtipour, Mandar Gogate, Ahsan Adeel, Cosimo Ieracitano, Hadi Larijani, Amir Hussain
The rise of social media is enabling people to freely express their opinions about products and services.