no code implementations • 2 Sep 2022 • Geoffrey Frost, Grant Theron, Thomas Niesler
In this work, we explore recurrent neural network architectures for tuberculosis (TB) cough classification.
no code implementations • 11 May 2022 • Madhurananda Pahar, Marisa Klopper, Byron Reeve, Rob Warren, Grant Theron, Andreas Diacon, Thomas Niesler
This cough data include 1. 68 hours of TB coughs, 18. 54 minutes of COVID-19 coughs and 1. 69 hours of healthy coughs from 47 TB patients, 229 COVID-19 patients and 1498 healthy patients and were used to train and evaluate a CNN, LSTM and Resnet50.
no code implementations • 8 Feb 2022 • Madhurananda Pahar, Igor Miranda, Andreas Diacon, Thomas Niesler
When integrated into a complete cough monitoring system, the daily cough rate of a patient undergoing TB treatment was determined over a period of 14 days.
no code implementations • 7 Oct 2021 • Madhurananda Pahar, Marisa Klopper, Byron Reeve, Rob Warren, Grant Theron, Andreas Diacon, Thomas Niesler
We present `wake-cough', an application of wake-word spotting to coughs using a Resnet50 and the identification of coughers using i-vectors, for the purpose of a long-term, personalised cough monitoring system.
no code implementations • 31 Aug 2021 • Madhurananda Pahar, Igor Miranda, Andreas Diacon, Thomas Niesler
We present an automatic non-invasive way of detecting cough events based on both accelerometer and audio signals.
no code implementations • 13 Aug 2021 • Ewald van der Westhuizen, Herman Kamper, Raghav Menon, John Quinn, Thomas Niesler
We show that, using these features, the CNN-DTW keyword spotter performs almost as well as the DTW keyword spotter while outperforming a baseline CNN trained only on the keyword templates.
no code implementations • 13 Aug 2021 • Ewald van der Westhuizen, Trideba Padhi, Thomas Niesler
We find that, although maximising the training pool by including all six additional languages provides improved speech recognition in both target languages, substantially better performance can be achieved by a more judicious choice.
no code implementations • 2 Apr 2021 • Madhurananda Pahar, Marisa Klopper, Robin Warren, Thomas Niesler
We present an experimental investigation into the effectiveness of transfer learning and bottleneck feature extraction in detecting COVID-19 from audio recordings of cough, breath and speech.
no code implementations • 23 Mar 2021 • Madhurananda Pahar, Marisa Klopper, Byron Reeve, Grant Theron, Rob Warren, Thomas Niesler
Objective: The automatic discrimination between the coughing sounds produced by patients with tuberculosis (TB) and those produced by patients with other lung ailments.
1 code implementation • 5 Mar 2021 • Nicholas Wilkinson, Thomas Niesler
We find that significantly smaller models with near optimal parameters perform on par with larger models trained with optimal parameters.
no code implementations • 9 Feb 2021 • Madhurananda Pahar, Igor Miranda, Andreas Diacon, Thomas Niesler
Since the need to gather audio is avoided and therefore privacy is inherently protected, and since the accelerometer is attached to the bed and not worn, this form of monitoring may represent a more convenient and readily accepted method of long-term patient cough monitoring.
no code implementations • 2 Dec 2020 • Madhurananda Pahar, Marisa Klopper, Robin Warren, Thomas Niesler
We present a machine learning based COVID-19 cough classifier which can discriminate COVID-19 positive coughs from both COVID-19 negative and healthy coughs recorded on a smartphone.
1 code implementation • 31 Oct 2020 • Trideba Padhi, Astik Biswas, Febe De Wet, Ewald van der Westhuizen, Thomas Niesler
In this work, we explore the benefits of using multilingual bottleneck features (mBNF) in acoustic modelling for the automatic speech recognition of code-switched (CS) speech in African languages.
no code implementations • LREC 2020 • Astik Biswas, Febe De Wet, Ewald van der Westhuizen, Thomas Niesler
We present an analysis of semi-supervised acoustic and language model training for English-isiZulu code-switched (CS) ASR using soap opera speech.
no code implementations • LREC 2020 • Nick Wilkinson, Astik Biswas, Emre Yilmaz, Febe De Wet, Ewald van der Westhuizen, Thomas Niesler
Automatic segmentation was applied in combination with automaticspeaker diarization.
no code implementations • LREC 2020 • Astik Biswas, Emre Yilmaz, Febe De Wet, Ewald van der Westhuizen, Thomas Niesler
This paper reports on the semi-supervised development of acoustic and language models for under-resourced, code-switched speech in five South African languages.
no code implementations • 6 Jul 2019 • Astik Biswas, Raghav Menon, Ewald van der Westhuizen, Thomas Niesler
The automatic transcriptions from the best performing pass were used for language model augmentation.
no code implementations • 20 Jun 2019 • Astik Biswas, Emre Yilmaz, Febe De Wet, Ewald van der Westhuizen, Thomas Niesler
Furthermore, because English is common to all language pairs in our data, it dominates when training a unified language model, leading to improved English ASR performance at the expense of the other languages.
no code implementations • 14 Nov 2018 • Raghav Menon, Herman Kamper, Ewald van der Westhuizen, John Quinn, Thomas Niesler
We compare features for dynamic time warping (DTW) when used to bootstrap keyword spotting (KWS) in an almost zero-resource setting.
no code implementations • 30 Oct 2018 • Lerato Lerato, Thomas Niesler
Agglomerative hierarchical clustering (AHC) requires only the similarity between objects to be known.
no code implementations • 30 Oct 2018 • Lerato Lerato, Thomas Niesler
Dynamic time warping (DTW) can be used to compute the similarity between two sequences of generally differing length.
no code implementations • 28 Jul 2018 • Emre Yilmaz, Astik Biswas, Ewald van der Westhuizen, Febe De Wet, Thomas Niesler
We present our first efforts towards building a single multilingual automatic speech recognition (ASR) system that can process code-switching (CS) speech in five languages spoken within the same population.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+1
no code implementations • 23 Jul 2018 • Raghav Menon, Herman Kamper, Emre Yilmaz, John Quinn, Thomas Niesler
We consider multilingual bottleneck features (BNFs) for nearly zero-resource keyword spotting.
no code implementations • 23 Jul 2018 • Raghav Menon, Astik Biswas, Armin Saeb, John Quinn, Thomas Niesler
We present our first efforts in building an automatic speech recognition system for Somali, an under-resourced language, using 1. 57 hrs of annotated speech for acoustic model training.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+5
no code implementations • 25 Jun 2018 • Raghav Menon, Herman Kamper, John Quinn, Thomas Niesler
While the resulting CNN keyword spotter cannot match the performance of the DTW-based system, it substantially outperforms a CNN classifier trained only on the keywords, improving the area under the ROC curve from 0. 54 to 0. 64.