no code implementations • EACL (AdaptNLP) 2021 • Abhinav Ramesh Kashyap, Laiba Mehnaz, Bhavitvya Malik, Abdul Waheed, Devamanyu Hazarika, Min-Yen Kan, Rajiv Ratn Shah
The robustness of pretrained language models(PLMs) is generally measured using performance drops on two or more domains.
no code implementations • 18 Feb 2025 • Abdul Waheed, Hanin Atwany, Rita Singh, Bhiksha Raj
Traditionally, ASR models are evaluated using metrics like Word Error Rate (WER) and Character Error Rate (CER), which depend on ground truth labels.
no code implementations • 18 Feb 2025 • Hanin Atwany, Abdul Waheed, Rita Singh, Monojit Choudhury, Bhiksha Raj
We examine how factors such as distribution shifts, model size, and model architecture influence the hallucination error rate (HER), a metric we introduce to quantify hallucinations.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+2
no code implementations • 19 Oct 2024 • Hao Chen, Abdul Waheed, Xiang Li, Yidong Wang, Jindong Wang, Bhiksha Raj, Marah I. Abdin
The rise of Large Language Models (LLMs) has accentuated the need for diverse, high-quality pre-training data.
no code implementations • 16 Oct 2024 • Abdul Waheed, Hanin Atwany, Bhiksha Raj, Rita Singh
The analysis of layer-wise features demonstrates that some models exhibit a convex relationship between the separability of the learned representations and model depth, with different layers capturing task-specific features.
1 code implementation • 1 Jul 2024 • Abdul Waheed, Karima Kadaoui, Bhiksha Raj, Muhammad Abdul-Mageed
Our models are also 25-50% more compute- and memory-efficient while maintaining performance equal to or better than that of the teacher model.
no code implementations • 24 Jun 2024 • Khai Duy Doan, Abdul Waheed, Muhammad Abdul-Mageed
We then evaluate our models on a dataset comprising 31 unseen speakers and an in-house dialectal dataset.
1 code implementation • 6 Jun 2024 • Abdul Waheed, Karima Kadaoui, Muhammad Abdul-Mageed
Our best-distilled model's overall performance ($45. 0$\% WER) surpasses that of a SoTA model twice its size (SeamlessM4T-large-v2, WER=$47. 0$\%) and its teacher model (Whisper-large-v2, WER=$55. 1$\%), and its average performance on our new dialectal data ($56. 9$\% WER) outperforms all other models.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+3
no code implementations • 17 Oct 2023 • Abdul Waheed, Bashar Talafha, Peter Sullivan, AbdelRahim Elmadany, Muhammad Abdul-Mageed
We train a wide range of models such as HuBERT (DID), Whisper, and XLS-R (ASR) in a supervised setting for Arabic DID and ASR tasks.
no code implementations • 12 Oct 2023 • Sadia Basar, Mushtaq Ali, Abdul Waheed, Muneer Ahmad, Mahdi H. Miraz
Therefore, it is important to detect in-focused objects in defocused-blurred images after the segmentation of blurred and non-blurred regions.
no code implementations • 6 Aug 2023 • Karima Kadaoui, Samar M. Magdy, Abdul Waheed, Md Tawkat Islam Khondaker, Ahmed Oumar El-Shangiti, El Moatez Billah Nagoudi, Muhammad Abdul-Mageed
Our evaluation covers diverse Arabic varieties such as Classical Arabic (CA), Modern Standard Arabic (MSA), and several country-level dialectal variants.
no code implementations • 5 Jun 2023 • Bashar Talafha, Abdul Waheed, Muhammad Abdul-Mageed
Whisper, the recently developed multilingual weakly supervised model, is reported to perform well on multiple speech recognition benchmarks in both monolingual and multilingual settings.
no code implementations • 24 May 2023 • Md Tawkat Islam Khondaker, Abdul Waheed, El Moatez Billah Nagoudi, Muhammad Abdul-Mageed
Although we further explore and confirm the utility of employing GPT-4 as a potential alternative for human evaluation, our work adds to a growing body of research underscoring the limitations of ChatGPT.
1 code implementation • 27 Apr 2023 • Minghao Wu, Abdul Waheed, Chiyu Zhang, Muhammad Abdul-Mageed, Alham Fikri Aji
The results demonstrate that our proposed LaMini-LM models are comparable to competitive baselines, while being much smaller in size.
Ranked #15 on
Word Sense Disambiguation
on Words in Context
1 code implementation • 12 Nov 2021 • Ganeshan Malhotra, Abdul Waheed, Aseem Srivastava, Md Shad Akhtar, Tanmoy Chakraborty
We identify the requirement of such conversation and propose twelve domain-specific dialogue-act (DAC) labels.
no code implementations • 16 Aug 2021 • Abdul Waheed, Muskan Goyal, Nimisha Mittal, Deepak Gupta, Ashish Khanna, Moolchand Sharma
For the optimization of educational programs, it is crucial to design course learning outcomes (CLOs) according to the different cognitive levels of Bloom Taxonomy.
no code implementations • 8 Mar 2021 • Abdul Waheed, Muskan Goyal, Nimisha Mittal, Deepak Gupta
We study automatic title generation and present a method for generating domain-controlled titles for scientific articles.
2 code implementations • 8 Mar 2021 • Abdul Waheed, Muskan Goyal, Deepak Gupta, Ashish Khanna, Fadi Al-Turjman, Placido Rogerio Pinheiro
This has led to the introduction of a variety of deep learning systems and studies have shown that the accuracy of COVID-19 patient detection through the use of chest X-rays is strongly optimistic.