Search Results for author: Zijian Yang

Found 14 papers, 2 papers with code

Investigating the Effect of Language Models in Sequence Discriminative Training for Neural Transducers

no code implementations11 Oct 2023 Zijian Yang, Wei Zhou, Ralf Schlüter, Hermann Ney

In this work, we investigate the effect of language models (LMs) with different context lengths and label units (phoneme vs. word) used in sequence discriminative training for phoneme-based neural transducers.

On the Relation between Internal Language Model and Sequence Discriminative Training for Neural Transducers

no code implementations25 Sep 2023 Zijian Yang, Wei Zhou, Ralf Schlüter, Hermann Ney

Empirically, we show that ILM subtraction and sequence discriminative training achieve similar effects across a wide range of experiments on Librispeech, including both MMI and minimum Bayes risk (MBR) criteria, as well as neural transducers and LMs of both full and limited context.

Language Modelling Relation +2

Enhancing Model Performance in Multilingual Information Retrieval with Comprehensive Data Engineering Techniques

no code implementations14 Feb 2023 Qi Zhang, Zijian Yang, Yilun Huang, Ze Chen, Zijian Cai, Kangxu Wang, Jiewen Zheng, Jiarong He, Jin Gao

In this paper, we present our solution to the Multilingual Information Retrieval Across a Continuum of Languages (MIRACL) challenge of WSDM CUP 2023\footnote{https://project-miracl. github. io/}.

Data Augmentation Information Retrieval +1

Lattice-Free Sequence Discriminative Training for Phoneme-Based Neural Transducers

no code implementations7 Dec 2022 Zijian Yang, Wei Zhou, Ralf Schlüter, Hermann Ney

Compared to the N-best-list based minimum Bayes risk objectives, lattice-free methods gain 40% - 70% relative training time speedup with a small degradation in performance.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +1

Revisiting Checkpoint Averaging for Neural Machine Translation

no code implementations21 Oct 2022 Yingbo Gao, Christian Herold, Zijian Yang, Hermann Ney

Checkpoint averaging is a simple and effective method to boost the performance of converged neural machine translation models.

Machine Translation Translation

A Semantic Alignment System for Multilingual Query-Product Retrieval

no code implementations5 Aug 2022 Qi Zhang, Zijian Yang, Yilun Huang, Ze Chen, Zijian Cai, Kangxu Wang, Jiewen Zheng, Jiarong He, Jin Gao

Our models are all trained with cross-entropy loss to classify the query-product pairs into ESCI 4 categories at first, and then we use weighted sum with the 4-class probabilities to get the score for ranking.

Data Augmentation Retrieval +1

An Effective Way for Cross-Market Recommendation with Hybrid Pre-Ranking and Ranking Models

1 code implementation2 Mar 2022 Qi Zhang, Zijian Yang, Yilun Huang, Jiarong He, Lixiang Wang

The Cross-Market Recommendation task of WSDM CUP 2022 is about finding solutions to improve individual recommendation systems in resource-scarce target markets by leveraging data from similar high-resource source markets.

feature selection Recommendation Systems

Incomplete Multi-View Weak-Label Learning with Noisy Features and Imbalanced Labels

2 code implementations4 Jan 2022 Zhiwei Li, Zijian Yang, Lu Sun, Mineichi Kudo, Kego Kimura

A variety of modern applications exhibit multi-view multi-label learning, where each sample has multi-view features, and multiple labels are correlated via common views.

Multi-Label Learning

Self-Normalized Importance Sampling for Neural Language Modeling

no code implementations11 Nov 2021 Zijian Yang, Yingbo Gao, Alexander Gerstenberger, Jintao Jiang, Ralf Schlüter, Hermann Ney

Compared to our previous work, the criteria considered in this work are self-normalized and there is no need to further conduct a correction step.

Automatic Speech Recognition Automatic Speech Recognition (ASR) +2

Transformer-Based Direct Hidden Markov Model for Machine Translation

no code implementations ACL 2021 Weiyue Wang, Zijian Yang, Yingbo Gao, Hermann Ney

The neural hidden Markov model has been proposed as an alternative to attention mechanism in machine translation with recurrent neural networks.

Machine Translation Translation

Cannot find the paper you are looking for? You can Submit a new open access paper.