no code implementations • 29 Jun 2022 • Jesús Andrés-Ferrer, Dario Albesano, Puming Zhan, Paul Vozila
In this work, we propose a contextual density ratio approach for both training a contextual aware E2E model and adapting the language model to named entities.
no code implementations • 29 Jun 2022 • Dario Albesano, Jesús Andrés-Ferrer, Nicola Ferri, Puming Zhan
In contrast to some previous works, our results show that Transformer does not always outperform LSTM when used as prediction network along with Conformer encoder.
no code implementations • 22 Jun 2022 • Felix Weninger, Marco Gaudesi, Md Akmal Haidar, Nicola Ferri, Jesús Andrés-Ferrer, Puming Zhan
In the dual-mode Conformer Transducer model, layers can function in online or offline mode while sharing parameters, and in-place knowledge distillation from offline to online mode is applied in training to improve online accuracy.
no code implementations • 27 Jul 2020 • Felix Weninger, Franco Mana, Roberto Gemello, Jesús Andrés-Ferrer, Puming Zhan
In the result, the Noisy Student algorithm with soft labels and consistency regularization achieves 10. 4% word error rate (WER) reduction when adding 475h of unlabeled data, corresponding to a recovery rate of 92%.
no code implementations • 8 Jul 2019 • Felix Weninger, Jesús Andrés-Ferrer, Xinwei Li, Puming Zhan
Sequence-to-sequence (seq2seq) based ASR systems have shown state-of-the-art performances while having clear advantages in terms of simplicity.