no code implementations • 2 Apr 2024 • Jinxi Guo, Niko Moritz, Yingyi Ma, Frank Seide, Chunyang Wu, Jay Mahadeokar, Ozlem Kalinli, Christian Fuegen, Mike Seltzer
However, even with the adoption of factorized transducer models, limited improvement has been observed compared to shallow fusion.
no code implementations • 17 Oct 2023 • Yingyi Ma, Zhe Liu, Ozlem Kalinli
Language models (LMs) have been commonly adopted to boost the performance of automatic speech recognition (ASR) particularly in domain adaptation tasks.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
no code implementations • 12 Sep 2023 • Arpita Vats, Zhe Liu, Peng Su, Debjyoti Paul, Yingyi Ma, Yutong Pang, Zeeshan Ahmed, Ozlem Kalinli
To effectively perform adaptation, textual data of users is typically stored on servers or their local devices, where downstream natural language processing (NLP) models can be directly trained using such in-domain data.
no code implementations • 1 Sep 2023 • Chuanneng Sun, Zeeshan Ahmed, Yingyi Ma, Zhe Liu, Lucas Kabela, Yutong Pang, Ozlem Kalinli
We propose to leverage prompts for a LLM without fine tuning during rescoring which incorporate a biasing list and few-shot examples to serve as additional information when calculating the score for the hypothesis.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • 9 Nov 2022 • Yingyi Ma, Zhe Liu, Xuedong Zhang
Thus, the data sampling strategy is important to the adaptation performance.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +3
1 code implementation • NeurIPS 2020 • Mao Li, Yingyi Ma, Xinhua Zhang
Underpinning the success of deep learning is effective regularizations that allow a variety of priors in data to be modeled.
no code implementations • ICML 2020 • Yingyi Ma, Vignesh Ganapathiraman, Yao-Liang Yu, Xinhua Zhang
Invariance (defined in a general sense) has been one of the most effective priors for representation learning.