no code implementations • 23 May 2024 • Pouria Saidi, Gautam Dasarathy, Visar Berisha
Validity concerns are underscored by findings of an inverse relationship between sample size and reported accuracy in published ML models, contrasting with the theory of learning curves where accuracy should improve or remain stable with increasing sample size.
1 code implementation • NeurIPS 2023 • Jianwei Zhang, Suren Jayasuriya, Visar Berisha
A good supervised embedding for a specific machine learning task is only sensitive to changes in the label of interest and is invariant to other confounding factors.
no code implementations • 4 Mar 2023 • Thomas B. Kaufmann, Mehdi Foroogozar, Julie Liss, Visar Berisha
Assistive listening systems (ALSs) dramatically increase speech intelligibility and reduce listening effort.
no code implementations • 17 Feb 2023 • Tyler Sypherd, Nathan Stromberg, Richard Nock, Visar Berisha, Lalitha Sankar
There is a growing need for models that are interpretable and have reduced energy and computational cost (e. g., in health care analytics and federated learning).
no code implementations • 30 Jan 2023 • Weizhi Li, Prad Kadambi, Pouria Saidi, Karthikeyan Natesan Ramamurthy, Gautam Dasarathy, Visar Berisha
The classification model is adaptively updated and used to predict where the (unlabelled) features have a high dependency on labels; labeling the ``high-dependency'' features leads to the increased power of the proposed testing framework.
1 code implementation • 17 Nov 2022 • Jianwei Zhang, Julie Liss, Suren Jayasuriya, Visar Berisha
In this paper, we propose a deep learning framework for generating acoustic feature embeddings sensitive to vocal quality and robust across different corpora.
1 code implementation • 17 Oct 2022 • Sean Kinahan, Julie Liss, Visar Berisha
The DIVA model is a computational model of speech motor control that combines a simulation of the brain regions responsible for speech production with a model of the human vocal tract.
no code implementations • 24 Mar 2022 • Leo Hsu, Visar Berisha
Researchers have observed that the frequencies of leading digits in many man-made and naturally occurring datasets follow a logarithmic curve, with digits that start with the number 1 accounting for $\sim 30\%$ of all numbers in the dataset and digits that start with the number 9 accounting for $\sim 5\%$ of all numbers in the dataset.
no code implementations • 18 Mar 2022 • Vikram C. Mathad, Julie M. Liss, Kathy Chapman, Nancy Scherer, Visar Berisha
Spectro-temporal dynamics of consonant-vowel (CV) transition regions are considered to provide robust cues related to articulation.
1 code implementation • 17 Nov 2021 • Weizhi Li, Gautam Dasarathy, Karthikeyan Natesan Ramamurthy, Visar Berisha
Two-sample tests evaluate whether two samples are realizations of the same distribution (the null hypothesis) or two different distributions (the alternative hypothesis).
no code implementations • 22 Apr 2021 • Jianwei Zhang, Suren Jayasuriya, Visar Berisha
We replace the mel-spectrum upsampler in DiffWave with a deep CNN upsampler, which is trained to alter the degraded speech mel-spectrum to match that of the original speech.
1 code implementation • NeurIPS 2020 • Weizhi Li, Gautam Dasarathy, Karthikeyan Natesan Ramamurthy, Visar Berisha
We theoretically analyze the proposed framework and show that the query complexity of our active learning algorithm depends naturally on the intrinsic complexity of the underlying manifold.
no code implementations • NeurIPS Workshop DL-IG 2020 • Prad Kadambi, Karthikeyan Natesan Ramamurthy, Visar Berisha
A large body of work addresses deep neural network (DNN) quantization and pruning to mitigate the high computational burden of deploying DNNs.
no code implementations • 7 Jan 2020 • Weizhi Li, Gautam Dasarathy, Visar Berisha
Regularization is an effective way to promote the generalization performance of machine learning models.
no code implementations • 26 Nov 2019 • Michael Saxon, Ayush Tripathi, Yishan Jiao, Julie Liss, Visar Berisha
To demonstrate that the features derived from these acoustic models are specific to hypernasal speech, we evaluate them across different dysarthria corpora.
no code implementations • 4 Jun 2019 • Rohit Voleti, Julie M. Liss, Visar Berisha
Broadly speaking, the review is split into two categories: language features based on natural language processing and speech features based on speech signal processing.
no code implementations • 24 Apr 2019 • Rohit Voleti, Stephanie Woolridge, Julie M. Liss, Melissa Milanovic, Christopher R. Bowie, Visar Berisha
Furthermore, the same feature set can be used to build a strong binary classifier to distinguish between healthy controls and a clinical group (AUC = 0. 96) and also between patients within the clinical group with schizophrenia and bipolar I disorder (AUC = 0. 83).
1 code implementation • 16 Nov 2018 • Rohit Voleti, Julie M. Liss, Visar Berisha
In this paper we investigate the effects of word substitution errors, such as those coming from automatic speech recognition errors (ASR), on several state-of-the-art sentence embedding methods.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +6
no code implementations • 4 Aug 2018 • Huan Song, Megan Willi, Jayaraman J. Thiagarajan, Visar Berisha, Andreas Spanias
In automatic speech processing systems, speaker diarization is a crucial front-end component to separate segments from different speakers.
no code implementations • 19 Apr 2018 • Shihui Yin, Gaurav Srivastava, Shreyas K. Venkataramanaiah, Chaitali Chakrabarti, Visar Berisha, Jae-sun Seo
Deep learning algorithms have shown tremendous success in many recognition tasks; however, these algorithms typically include a deep neural network (DNN) structure and a large number of parameters, which makes it challenging to implement them on power/area-constrained embedded platforms.
no code implementations • 21 Feb 2017 • Alan Wisler, Visar Berisha, Andreas Spanias, Alfred O. Hero
Typically, estimating these quantities requires complete knowledge of the underlying distribution followed by multi-dimensional integration.
no code implementations • 16 May 2016 • Ming Tu, Visar Berisha, Yu Cao, Jae-sun Seo
In this paper, we propose a method to compress deep neural networks by using the Fisher Information metric, which we estimate through a stochastic optimization method that keeps track of second-order information in the network.
no code implementations • 19 Dec 2014 • Visar Berisha, Alan Wisler, Alfred O. Hero, Andreas Spanias
Information divergence functions play a critical role in statistics and information theory.
1 code implementation • 6 Aug 2014 • Visar Berisha, Alfred O. Hero
Traditional approaches to estimating the FIM require estimating the probability distribution function (PDF), or its parameters, along with its gradient or Hessian.