no code implementations • 3 Feb 2024 • Cunxiao Du, Hao Zhou, Zhaopeng Tu, Jing Jiang
In this paper, we re-examine the Markov property in the context of neural machine translation.
no code implementations • 3 Feb 2024 • Cunxiao Du, Jing Jiang, Xu Yuanchen, Jiawei Wu, Sicheng Yu, Yongqi Li, Shenggui Li, Kai Xu, Liqiang Nie, Zhaopeng Tu, Yang You
Speculative decoding is a relatively new decoding framework that leverages small and efficient draft models to reduce the latency of LLMs.
no code implementations • COLING 2022 • Cunxiao Du, Zhaopeng Tu, Longyue Wang, Jing Jiang
Recently, a new training oaxe loss has proven effective to ameliorate the effect of multimodality for non-autoregressive translation (NAT), which removes the penalty of word order errors in the standard cross-entropy loss.
1 code implementation • 9 Jun 2021 • Cunxiao Du, Zhaopeng Tu, Jing Jiang
We propose a new training objective named order-agnostic cross entropy (OaXE) for fully non-autoregressive translation (NAT) models.
1 code implementation • 23 Nov 2018 • Cunxiao Du, Zhaozheng Chin, Fuli Feng, Lei Zhu, Tian Gan, Liqiang Nie
To address this problem, we introduce the interaction mechanism to incorporate word-level matching signals into the text classification task.
Ranked #4 on Text Classification on Yahoo! Answers