no code implementations • 13 Mar 2024 • Wenjing Zhu, Sining Sun, Changhao Shan, Peng Fan, Qing Yang
Conformer-based attention models have become the de facto backbone model for Automatic Speech Recognition tasks.
1 code implementation • 23 Oct 2023 • Peng Fan, Changhao Shan, Sining Sun, Qing Yang, Jianwei Zhang
Following the initial encoder, we introduce an intermediate CTC loss function to compute the label frame, enabling us to extract the key frames and blank frames for KFSA.
no code implementations • 23 Jul 2023 • Ziwei Zhu, Changhao Shan, Bihong Zhang, Jian Yu
We combine the methods of meta learning and freeze of model parameters, which makes the recognition performance more stable in different cases and the training faster about 20%.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
3 code implementations • 29 Mar 2018 • Changhao Shan, Junbo Zhang, Yujun Wang, Lei Xie
In this paper, we propose an attention-based end-to-end neural approach for small-footprint keyword spotting (KWS), which aims to simplify the pipelines of building a production-quality KWS system.
no code implementations • 22 Jul 2017 • Changhao Shan, Junbo Zhang, Yujun Wang, Lei Xie
Previous attempts have shown that applying attention-based encoder-decoder to Mandarin speech recognition was quite difficult due to the logographic orthography of Mandarin, the large vocabulary and the conditional dependency of the attention model.