1 code implementation • 31 Mar 2022 • Keyu An, Huahuan Zheng, Zhijian Ou, Hongyu Xiang, Ke Ding, Guanglu Wan
The simulation module is jointly trained with the ASR model using a self-supervised loss; the ASR model is optimized with the usual ASR loss, e. g., CTC-CRF as used in our experiments.
no code implementations • 31 Mar 2022 • Huahuan Zheng, Keyu An, Zhijian Ou, Chen Huang, Ke Ding, Guanglu Wan
Based on the DR method, we propose a low-order density ratio method (LODR) by replacing the estimation with a low-order weak language model.
1 code implementation • 11 Jul 2021 • Chengrui Zhu, Keyu An, Huahuan Zheng, Zhijian Ou
The use of phonological features (PFs) potentially allows language-specific phones to remain linked in training, which is highly desirable for information sharing for multilingual and crosslingual speech recognition methods for low-resourced languages.
1 code implementation • 7 Jul 2021 • Huahuan Zheng, Wenjie Peng, Zhijian Ou, Jinsong Zhang
Automatic speech recognition systems have been largely improved in the past few decades and current systems are mainly hybrid-based and end-to-end-based.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 11 Nov 2020 • Huahuan Zheng, Keyu An, Zhijian Ou
Using ST gradients to support sub-graph sampling is a core element to achieve efficient NAS beyond DARTS and SNAS.
Ranked #1 on Speech Recognition on WSJ dev93
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • 25 Oct 2020 • Yunfu Song, Huahuan Zheng, Zhijian Ou
In contrast, generative SSL methods involve unsupervised learning based on generative models by either joint-training or pre-training, and are more appealing from the perspective of being domain-agnostic, since they do not inherently require data augmentations.