no code implementations • 11 Oct 2023 • Zijian Yang, Wei Zhou, Ralf Schlüter, Hermann Ney
In this work, we investigate the effect of language models (LMs) with different context lengths and label units (phoneme vs. word) used in sequence discriminative training for phoneme-based neural transducers.
no code implementations • 25 Sep 2023 • Zijian Yang, Wei Zhou, Ralf Schlüter, Hermann Ney
Empirically, we show that ILM subtraction and sequence discriminative training achieve similar effects across a wide range of experiments on Librispeech, including both MMI and minimum Bayes risk (MBR) criteria, as well as neural transducers and LMs of both full and limited context.
no code implementations • 14 Feb 2023 • Qi Zhang, Zijian Yang, Yilun Huang, Ze Chen, Zijian Cai, Kangxu Wang, Jiewen Zheng, Jiarong He, Jin Gao
In this paper, we present our solution to the Multilingual Information Retrieval Across a Continuum of Languages (MIRACL) challenge of WSDM CUP 2023\footnote{https://project-miracl. github. io/}.
no code implementations • 7 Dec 2022 • Zijian Yang, Wei Zhou, Ralf Schlüter, Hermann Ney
Compared to the N-best-list based minimum Bayes risk objectives, lattice-free methods gain 40% - 70% relative training time speedup with a small degradation in performance.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
no code implementations • 24 Oct 2022 • Christoph Lüscher, Mohammad Zeineldeen, Zijian Yang, Tina Raissi, Peter Vieting, Khai Le-Duc, Weiyue Wang, Ralf Schlüter, Hermann Ney
Language barriers present a great challenge in our increasingly connected and global world.
no code implementations • 21 Oct 2022 • Yingbo Gao, Christian Herold, Zijian Yang, Hermann Ney
Checkpoint averaging is a simple and effective method to boost the performance of converged neural machine translation models.
no code implementations • 21 Oct 2022 • Yingbo Gao, Christian Herold, Zijian Yang, Hermann Ney
Encoder-decoder architecture is widely adopted for sequence-to-sequence modeling tasks.
no code implementations • 5 Aug 2022 • Qi Zhang, Zijian Yang, Yilun Huang, Ze Chen, Zijian Cai, Kangxu Wang, Jiewen Zheng, Jiarong He, Jin Gao
Our models are all trained with cross-entropy loss to classify the query-product pairs into ESCI 4 categories at first, and then we use weighted sum with the 4-class probabilities to get the score for ranking.
1 code implementation • 2 Mar 2022 • Qi Zhang, Zijian Yang, Yilun Huang, Jiarong He, Lixiang Wang
The Cross-Market Recommendation task of WSDM CUP 2022 is about finding solutions to improve individual recommendation systems in resource-scarce target markets by leveraging data from similar high-resource source markets.
2 code implementations • 4 Jan 2022 • Zhiwei Li, Zijian Yang, Lu Sun, Mineichi Kudo, Kego Kimura
A variety of modern applications exhibit multi-view multi-label learning, where each sample has multi-view features, and multiple labels are correlated via common views.
no code implementations • 11 Nov 2021 • Zijian Yang, Yingbo Gao, Alexander Gerstenberger, Jintao Jiang, Ralf Schlüter, Hermann Ney
Compared to our previous work, the criteria considered in this work are self-normalized and there is no need to further conduct a correction step.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +2
no code implementations • ACL 2021 • Weiyue Wang, Zijian Yang, Yingbo Gao, Hermann Ney
The neural hidden Markov model has been proposed as an alternative to attention mechanism in machine translation with recurrent neural networks.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Yingbo Gao, Weiyue Wang, Christian Herold, Zijian Yang, Hermann Ney
In order to combat overfitting and in pursuit of better generalization, label smoothing is widely applied in modern neural machine translation systems.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Zijian Yang, Yingbo Gao, Weiyue Wang, Hermann Ney
Attention-based encoder-decoder models have achieved great success in neural machine translation tasks.