Search Results for author: Yizhou Peng

Found 8 papers, 1 papers with code

The NUS-HLT System for ICASSP2024 ICMC-ASR Grand Challenge

no code implementations26 Dec 2023 Meng Ge, Yizhou Peng, Yidi Jiang, Jingru Lin, Junyi Ao, Mehmet Sinan Yildirim, Shuai Wang, Haizhou Li, Mengling Feng

This paper summarizes our team's efforts in both tracks of the ICMC-ASR Challenge for in-car multi-channel automatic speech recognition.

Automatic Speech Recognition Data Augmentation +2

Adapting OpenAI's Whisper for Speech Recognition on Code-Switch Mandarin-English SEAME and ASRU2019 Datasets

no code implementations29 Nov 2023 Yuhang Yang, Yizhou Peng, Xionghu Zhong, Hao Huang, Eng Siong Chng

The Mixed Error Rate results show that the amount of adaptation data may be as low as $1\sim10$ hours to achieve saturation in performance gain (SEAME) while the ASRU task continued to show performance with more adaptation data ($>$100 hours).

speech-recognition Speech Recognition

Mutual Information-Based Integrated Sensing and Communications: A WMMSE Framework

1 code implementation19 Oct 2023 Yizhou Peng, Songjie Yang, Wanting Lyu, Ya Li, Hongjun He, Zhongpei Zhang, Chadi Assi

In this letter, a weighted minimum mean square error (WMMSE) empowered integrated sensing and communication (ISAC) system is investigated.

Intermediate-layer output Regularization for Attention-based Speech Recognition with Shared Decoder

no code implementations9 Jul 2022 Jicheng Zhang, Yizhou Peng, HaiHua Xu, Yi He, Eng Siong Chng, Hao Huang

Intermediate layer output (ILO) regularization by means of multitask training on encoder side has been shown to be an effective approach to yielding improved results on a wide range of end-to-end ASR frameworks.

speech-recognition Speech Recognition

Internal Language Model Estimation based Language Model Fusion for Cross-Domain Code-Switching Speech Recognition

no code implementations9 Jul 2022 Yizhou Peng, Yufei Liu, Jicheng Zhang, HaiHua Xu, Yi He, Hao Huang, Eng Siong Chng

More importantly, we train an end-to-end (E2E) speech recognition model by means of merging two monolingual data sets and observe the efficacy of the proposed ILME-based LM fusion for CSSR.

Language Modelling speech-recognition +1

Minimum word error training for non-autoregressive Transformer-based code-switching ASR

no code implementations7 Oct 2021 Yizhou Peng, Jicheng Zhang, HaiHua Xu, Hao Huang, Eng Siong Chng

Non-autoregressive end-to-end ASR framework might be potentially appropriate for code-switching recognition task thanks to its inherent property that present output token being independent of historical ones.

E2E-based Multi-task Learning Approach to Joint Speech and Accent Recognition

no code implementations15 Jun 2021 Jicheng Zhang, Yizhou Peng, Pham Van Tung, HaiHua Xu, Hao Huang, Eng Siong Chng

In this paper, we propose a single multi-task learning framework to perform End-to-End (E2E) speech recognition (ASR) and accent recognition (AR) simultaneously.

Multi-Task Learning speech-recognition +1

Multilingual Approach to Joint Speech and Accent Recognition with DNN-HMM Framework

no code implementations22 Oct 2020 Yizhou Peng, Jicheng Zhang, Haobo Zhang, HaiHua Xu, Hao Huang, Eng Siong Chng

Experimental results on an 8-accent English speech recognition show both methods can yield WERs close to the conventional ASR systems that completely ignore the accent, as well as desired AR accuracy.

speech-recognition Speech Recognition +1

Cannot find the paper you are looking for? You can Submit a new open access paper.