Search Results for author: Kia Dashtipour

Found 14 papers, 1 papers with code

Audio-Visual Speech Enhancement in Noisy Environments via Emotion-Based Contextual Cues

no code implementations26 Feb 2024 Tassadaq Hussain, Kia Dashtipour, Yu Tsao, Amir Hussain

By integrating emotional features, the proposed system achieves significant improvements in both objective and subjective assessments of speech quality and intelligibility, especially in challenging noise environments.

Speech Enhancement

Audio-Visual Speech Enhancement Using Self-supervised Learning to Improve Speech Intelligibility in Cochlear Implant Simulations

no code implementations15 Jul 2023 Richard Lee Lai, Jen-Cheng Hou, Mandar Gogate, Kia Dashtipour, Amir Hussain, Yu Tsao

The aim of this study is to explore the effectiveness of audio-visual speech enhancement (AVSE) in enhancing the intelligibility of vocoded speech in cochlear implant (CI) simulations.

Self-Supervised Learning Speech Enhancement

A Novel Frame Structure for Cloud-Based Audio-Visual Speech Enhancement in Multimodal Hearing-aids

no code implementations24 Oct 2022 Abhijeet Bishnu, Ankit Gupta, Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Amir Hussain, Mathini Sellathurai, Tharmalingam Ratnarajah

In this paper, we design a first of its kind transceiver (PHY layer) prototype for cloud-based audio-visual (AV) speech enhancement (SE) complying with high data rate and low latency requirements of future multimodal hearing assistive technology.

Lip Reading Speech Enhancement

A Novel Speech Intelligibility Enhancement Model based on CanonicalCorrelation and Deep Learning

no code implementations11 Feb 2022 Tassadaq Hussain, Muhammad Diyan, Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Yu Tsao, Amir Hussain

Current deep learning (DL) based approaches to speech intelligibility enhancement in noisy environments are often trained to minimise the feature distance between noise-free speech and enhanced speech signals.

Speech Enhancement

A Speech Intelligibility Enhancement Model based on Canonical Correlation and Deep Learning for Hearing-Assistive Technologies

no code implementations8 Feb 2022 Tassadaq Hussain, Muhammad Diyan, Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Yu Tsao, Amir Hussain

Current deep learning (DL) based approaches to speech intelligibility enhancement in noisy environments are generally trained to minimise the distance between clean and enhanced speech features.

Speech Enhancement

A Novel Temporal Attentive-Pooling based Convolutional Recurrent Architecture for Acoustic Signal Enhancement

no code implementations24 Jan 2022 Tassadaq Hussain, Wei-Chien Wang, Mandar Gogate, Kia Dashtipour, Yu Tsao, Xugang Lu, Adeel Ahsan, Amir Hussain

To address this problem, we propose to integrate a novel temporal attentive-pooling (TAP) mechanism into a conventional convolutional recurrent neural network, termed as TAP-CRNN.

Towards Robust Real-time Audio-Visual Speech Enhancement

no code implementations16 Dec 2021 Mandar Gogate, Kia Dashtipour, Amir Hussain

The human brain contextually exploits heterogeneous sensory information to efficiently perform cognitive tasks including vision and hearing.

Speech Enhancement

Towards Intelligibility-Oriented Audio-Visual Speech Enhancement

1 code implementation18 Nov 2021 Tassadaq Hussain, Mandar Gogate, Kia Dashtipour, Amir Hussain

To the best of our knowledge, this is the first work that exploits the integration of AV modalities with an I-O based loss function for SE.

Speech Enhancement

An Intelligent Non-Invasive Real Time Human Activity Recognition System for Next-Generation Healthcare

no code implementations6 Aug 2020 William Taylor, Syed Aziz Shah, Kia Dashtipour, Adnan Zahid, Qammer H. Abbasi, Muhammad Ali Imran

The dataset was used to create a machine learning model, which was used in a developed application to provide a quasi-real-time classification of standing or sitting state.

BIG-bench Machine Learning Human Activity Recognition +1

A Hybrid Persian Sentiment Analysis Framework: Integrating Dependency Grammar Based Rules and Deep Neural Networks

no code implementations30 Sep 2019 Kia Dashtipour, Mandar Gogate, Jingpeng Li, Fengling Jiang, Bin Kong, Amir Hussain

When no pattern is triggered, the framework switches to its subsymbolic counterpart and leverages deep neural networks (DNN) to perform the classification.

Persian Sentiment Analysis

AV Speech Enhancement Challenge using a Real Noisy Corpus

no code implementations30 Sep 2019 Mandar Gogate, Ahsan Adeel, Kia Dashtipour, Peter Derleth, Amir Hussain

This paper presents, a first of its kind, audio-visual (AV) speech enhacement challenge in real-noisy settings.

Speech Enhancement

CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement

no code implementations23 Sep 2019 Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Amir Hussain

In addition, our work challenges a popular belief that a scarcity of multi-language large vocabulary AV corpus and wide variety of noises is a major bottleneck to build a robust language, speaker and noise independent SE systems.

Speech Enhancement

Cannot find the paper you are looking for? You can Submit a new open access paper.