no code implementations • 20 Feb 2023 • Jiasheng Ye, Zaixiang Zheng, Yu Bao, Lihua Qian, Mingxuan Wang
In this paper, we introduce DINOISER to facilitate diffusion models for sequence generation by manipulating noises.
no code implementations • 20 Dec 2022 • Yaoming Zhu, Zewei Sun, Shanbo Cheng, YuYang Huang, Liwei Wu, Mingxuan Wang
Therefore, this paper correspondingly establishes new methods and new datasets for MMT.
no code implementations • 20 Dec 2022 • Lihua Qian, Mingxuan Wang, Yang Liu, Hao Zhou
Autoregressive models can achieve high generation quality, but the sequential decoding scheme causes slow decoding speed.
1 code implementation • 19 Dec 2022 • Wenda Xu, Xian Qian, Mingxuan Wang, Lei LI, William Yang Wang
Existing learned metrics have gaps to human judgements, are model-dependent or are limited to the domains or tasks where human ratings are available.
1 code implementation • 17 Dec 2022 • Yifan Wang, Zewei Sun, Shanbo Cheng, Weiguo Zheng, Mingxuan Wang
First, we re-visit this task and propose a multiway stylized machine translation (MSMT) benchmark, which includes multiple categories of styles in four language directions to push the boundary of this task.
no code implementations • 17 Dec 2022 • Jiahuan Li, Shanbo Cheng, Zewei Sun, Mingxuan Wang, ShuJian Huang
The effectiveness of kNNMT directly depends on the quality of retrieved neighbors.
no code implementations • 7 Dec 2022 • Xuxin Cheng, Qianqian Dong, Fengpeng Yue, Tom Ko, Mingxuan Wang, Yuexian Zou
How to solve the data scarcity problem for end-to-end speech-to-text translation (ST)?
no code implementations • 20 Nov 2022 • Yunhao Gou, Tom Ko, Hansi Yang, James Kwok, Yu Zhang, Mingxuan Wang
(2) Under-utilization of the unmasked tokens: CMLM primarily focuses on the masked token but it cannot simultaneously leverage other tokens to learn vision-language associations.
no code implementations • 20 Oct 2022 • Xian Qian, Kai Hu, Jiaqiang Wang, Yifeng Liu, Xingyuan Pan, Jun Cao, Mingxuan Wang
This report describes our VolcTrans system for the WMT22 shared task on large-scale multilingual machine translation.
1 code implementation • 7 Oct 2022 • Jiangtao Feng, Yi Zhou, Jun Zhang, Xian Qian, Liwei Wu, Zhexi Zhang, Yanming Liu, Mingxuan Wang, Lei LI, Hao Zhou
PARAGEN is a PyTorch-based NLP toolkit for further development on parallel generation.
no code implementations • 23 Sep 2022 • Zewei Sun, Qingnan Jiang, ShuJian Huang, Jun Cao, Shanbo Cheng, Mingxuan Wang
Domain adaptation is an important challenge for neural machine translation.
1 code implementation • 18 May 2022 • Qianqian Dong, Fengpeng Yue, Tom Ko, Mingxuan Wang, Qibing Bai, Yu Zhang
Direct Speech-to-speech translation (S2ST) has drawn more and more attention recently.
1 code implementation • NAACL 2022 • Rong Ye, Mingxuan Wang, Lei LI
Learning similar representations for semantically similar speech and text is important for speech translation.
no code implementations • 8 Apr 2022 • Rong Ye, Chengqi Zhao, Tom Ko, Chutong Meng, Tao Wang, Mingxuan Wang, Jun Cao
The training set is translated by a strong machine translation system and the test set is translated by human.
no code implementations • 26 Mar 2022 • Sha Yuan, Hanyu Zhao, Shuai Zhao, Jiahong Leng, Yangxiao Liang, Xiaozhi Wang, Jifan Yu, Xin Lv, Zhou Shao, Jiaao He, Yankai Lin, Xu Han, Zhenghao Liu, Ning Ding, Yongming Rao, Yizhao Gao, Liang Zhang, Ming Ding, Cong Fang, Yisen Wang, Mingsheng Long, Jing Zhang, Yinpeng Dong, Tianyu Pang, Peng Cui, Lingxiao Huang, Zheng Liang, HuaWei Shen, HUI ZHANG, Quanshi Zhang, Qingxiu Dong, Zhixing Tan, Mingxuan Wang, Shuo Wang, Long Zhou, Haoran Li, Junwei Bao, Yingwei Pan, Weinan Zhang, Zhou Yu, Rui Yan, Chence Shi, Minghao Xu, Zuobai Zhang, Guoqiang Wang, Xiang Pan, Mengjie Li, Xiaoyu Chu, Zijun Yao, Fangwei Zhu, Shulin Cao, Weicheng Xue, Zixuan Ma, Zhengyan Zhang, Shengding Hu, Yujia Qin, Chaojun Xiao, Zheni Zeng, Ganqu Cui, Weize Chen, Weilin Zhao, Yuan YAO, Peng Li, Wenzhao Zheng, Wenliang Zhao, Ziyi Wang, Borui Zhang, Nanyi Fei, Anwen Hu, Zenan Ling, Haoyang Li, Boxi Cao, Xianpei Han, Weidong Zhan, Baobao Chang, Hao Sun, Jiawen Deng, Chujie Zheng, Juanzi Li, Lei Hou, Xigang Cao, Jidong Zhai, Zhiyuan Liu, Maosong Sun, Jiwen Lu, Zhiwu Lu, Qin Jin, Ruihua Song, Ji-Rong Wen, Zhouchen Lin, LiWei Wang, Hang Su, Jun Zhu, Zhifang Sui, Jiajun Zhang, Yang Liu, Xiaodong He, Minlie Huang, Jian Tang, Jie Tang
With the rapid development of deep learning, training Big Models (BMs) for multiple downstream tasks becomes a popular paradigm.
1 code implementation • ACL 2022 • Qingkai Fang, Rong Ye, Lei LI, Yang Feng, Mingxuan Wang
How to learn a better speech representation for end-to-end speech-to-text translation (ST) with limited labeled data?
1 code implementation • 24 Jan 2022 • Yaoming Zhu, Liwei Wu, Shanbo Cheng, Mingxuan Wang
The punctuation restoration task aims to correctly punctuate the output transcriptions of automatic speech recognition systems.
Automatic Speech Recognition
Automatic Speech Recognition (ASR)
+2
1 code implementation • 12 Oct 2021 • Xiaohui Wang, Yang Wei, Ying Xiong, Guyue Huang, Xian Qian, Yufei Ding, Mingxuan Wang, Lei LI
In this paper, we present LightSeq2, a system to accelerate training for a general family of Transformer models on GPUs.
no code implementations • ICLR 2022 • Zhenqiao Song, Hao Zhou, Lihua Qian, Jingjing Xu, Shanbo Cheng, Mingxuan Wang, Lei LI
Multilingual machine translation aims to develop a single model for multiple language directions.
no code implementations • WMT (EMNLP) 2021 • Lihua Qian, Yi Zhou, Zaixiang Zheng, Yaoming Zhu, Zehui Lin, Jiangtao Feng, Shanbo Cheng, Lei LI, Mingxuan Wang, Hao Zhou
This paper describes the Volctrans' submission to the WMT21 news translation shared task for German->English translation.
2 code implementations • EMNLP 2021 • Qingnan Jiang, Mingxuan Wang, Jun Cao, Shanbo Cheng, ShuJian Huang, Lei LI
How to effectively adapt neural machine translation (NMT) models according to emerging cases without retraining?
1 code implementation • ACL 2022 • Qianqian Dong, Yaoming Zhu, Mingxuan Wang, Lei LI
Given a usually long speech sequence, we develop an efficient monotonic segmentation module inside an encoder-decoder model to accumulate acoustic information incrementally and detect proper speech unit boundaries for the input in speech translation task.
1 code implementation • Findings (EMNLP) 2021 • Zewei Sun, Mingxuan Wang, Lei LI
Can pre-trained BERT for one language and GPT for another be glued together to translate texts?
no code implementations • Findings (EMNLP) 2021 • Tao Wang, Chengqi Zhao, Mingxuan Wang, Lei LI, Hang Li, Deyi Xiong
This paper presents Self-correcting Encoding (Secoco), a framework that effectively deals with input noise for robust neural machine translation by introducing self-correcting predictors.
no code implementations • ACL 2021 • Mingxuan Wang, Lei LI
This tutorial provides a comprehensive guide to make the most of pre-training for neural machine translation.
no code implementations • Findings (ACL) 2021 • Liwei Wu, Shanbo Cheng, Mingxuan Wang, Lei LI
Language tag (LT) strategies are often adopted to indicate the translation directions in MNMT.
no code implementations • NAACL 2021 • Tao Wang, Chengqi Zhao, Mingxuan Wang, Lei LI, Deyi Xiong
Automatic translation of dialogue texts is a much needed demand in many real life scenarios.
3 code implementations • ACL 2021 • Xiao Pan, Mingxuan Wang, Liwei Wu, Lei LI
Existing multilingual machine translation approaches mainly focus on English-centric directions, while the non-English directions still lag behind.
1 code implementation • ACL 2021 • Zehui Lin, Liwei Wu, Mingxuan Wang, Lei LI
These jointly trained models often suffer from performance degradation on rich-resource language pairs.
1 code implementation • ACL (IWSLT) 2021 • Chengqi Zhao, Zhicheng Liu, Jian Tong, Tao Wang, Mingxuan Wang, Rong Ye, Qianqian Dong, Jun Cao, Lei LI
For offline speech translation, our best end-to-end model achieves 8. 1 BLEU improvements over the benchmark on the MuST-C test set and is even approaching the results of a strong cascade solution.
2 code implementations • Findings (ACL) 2021 • Chi Han, Mingxuan Wang, Heng Ji, Lei LI
By projecting audio and text features to a common semantic representation, Chimera unifies MT and ST tasks and boosts the performance on ST benchmarks, MuST-C and Augmented Librispeech, to a new state-of-the-art.
1 code implementation • 21 Apr 2021 • Rong Ye, Mingxuan Wang, Lei LI
XSTNet takes both speech and text as input and outputs both transcription and translation text.
1 code implementation • Findings (EMNLP) 2021 • Yaoming Zhu, Jiangtao Feng, Chengqi Zhao, Mingxuan Wang, Lei LI
Developing a unified multilingual model has long been a pursuit for machine translation.
1 code implementation • 30 Mar 2021 • Tao Wang, Chengqi Zhao, Mingxuan Wang, Lei LI, Deyi Xiong
Automatic translation of dialogue texts is a much needed demand in many real life scenarios.
no code implementations • 1 Jan 2021 • Lihua Qian, Hao Zhou, Yu Bao, Mingxuan Wang, Lin Qiu, Weinan Zhang, Yong Yu, Lei LI
Although non-autoregressive models with one-iteration generation achieves remarkable inference speed-up, they still falls behind their autoregressive counterparts inprediction accuracy.
2 code implementations • 19 Dec 2020 • Jianze Liang, Chengqi Zhao, Mingxuan Wang, Xipeng Qiu, Lei LI
Neural machine translation often adopts the fine-tuning approach to adapt to specific domains.
1 code implementation • ACL 2021 • Chengqi Zhao, Mingxuan Wang, Qianqian Dong, Rong Ye, Lei LI
NeurST is an open-source toolkit for neural speech translation.
Ranked #1 on
Speech-to-Text Translation
on libri-trans
1 code implementation • 5 Dec 2020 • Minkai Xu, Mingxuan Wang, Zhouhan Lin, Hao Zhou, Weinan Zhang, Lei LI
Despite the recent success on image classification, self-training has only achieved limited gains on structured prediction tasks such as neural machine translation (NMT).
3 code implementations • EMNLP 2020 • Bohan Li, Hao Zhou, Junxian He, Mingxuan Wang, Yiming Yang, Lei LI
Pre-trained contextual representations like BERT have achieved great success in natural language processing.
Ranked #10 on
Semantic Textual Similarity
on STS16
no code implementations • WMT (EMNLP) 2020 • Liwei Wu, Xiao Pan, Zehui Lin, Yaoming Zhu, Mingxuan Wang, Lei LI
This paper describes our VolcTrans system on WMT20 shared news translation task.
no code implementations • WMT (EMNLP) 2020 • Runxin Xu, Zhuo Zhi, Jun Cao, Mingxuan Wang, Lei LI
In this paper, we describe our submissions to the WMT20 shared task on parallel corpus filtering and alignment for low-resource conditions.
1 code implementation • NAACL 2021 • Xiaohui Wang, Ying Xiong, Yang Wei, Mingxuan Wang, Lei LI
Transformer, BERT and their variants have achieved great success in natural language processing.
1 code implementation • Findings (ACL) 2022 • Zewei Sun, Mingxuan Wang, Hao Zhou, Chengqi Zhao, ShuJian Huang, Jiajun Chen, Lei LI
This paper does not aim at introducing a novel model for document-level neural machine translation.
1 code implementation • EMNLP 2020 • Zehui Lin, Xiao Pan, Mingxuan Wang, Xipeng Qiu, Jiangtao Feng, Hao Zhou, Lei LI
We investigate the following question for machine translation (MT): can we develop a single universal MT model to serve as the common seed and obtain derivative and improved models on arbitrary language pairs?
Ranked #3 on
Machine Translation
on WMT2014 English-French
(using extra training data)
1 code implementation • 21 Sep 2020 • Qianqian Dong, Rong Ye, Mingxuan Wang, Hao Zhou, Shuang Xu, Bo Xu, Lei LI
Can we build a system to fully utilize signals in a parallel ST corpus?
no code implementations • NAACL 2021 • Quanyu Long, Mingxuan Wang, Lei LI
Given a sentence in a source language, whether depicting the visual scene helps translation into a target language?
Ranked #6 on
Multimodal Machine Translation
on Multi30K
1 code implementation • 21 Sep 2020 • Qianqian Dong, Mingxuan Wang, Hao Zhou, Shuang Xu, Bo Xu, Lei LI
The key idea is to generate source transcript and target translation text with a single decoder.
no code implementations • ACL 2021 • Lihua Qian, Hao Zhou, Yu Bao, Mingxuan Wang, Lin Qiu, Wei-Nan Zhang, Yong Yu, Lei LI
With GLM, we develop Glancing Transformer (GLAT) for machine translation.
Ranked #67 on
Machine Translation
on WMT2014 English-German
no code implementations • ACL 2020 • Runxin Xu, Jun Cao, Mingxuan Wang, Jiaze Chen, Hao Zhou, Ying Zeng, Yu-Ping Wang, Li Chen, Xiang Yin, Xijin Zhang, Songcheng Jiang, Yuxuan Wang, Lei LI
This paper proposes the building of Xiaomingbot, an intelligent, multilingual and multimodal software robot equipped with four integral capabilities: news generation, news translation, news reading and avatar animation.
no code implementations • 12 Jul 2020 • Yuxuan Song, Ning Miao, Hao Zhou, Lantao Yu, Mingxuan Wang, Lei LI
Auto-regressive sequence generative models trained by Maximum Likelihood Estimation suffer the exposure bias problem in practical finite sample scenarios.
no code implementations • Findings of the Association for Computational Linguistics 2020 • Dongyu Ru, Jiangtao Feng, Lin Qiu, Hao Zhou, Mingxuan Wang, Wei-Nan Zhang, Yong Yu, Lei LI
We propose adversarial uncertainty sampling in discrete space (AUSDS) to retrieve informative unlabeled samples more efficiently.
no code implementations • NAACL 2021 • Mingxuan Wang, Hongxiao Bai, Hai Zhao, Lei LI
Neural machine translation~(NMT) is ineffective for zero-resource languages.
no code implementations • 25 Nov 2019 • Yu Bao, Hao Zhou, Jiangtao Feng, Mingxuan Wang, Shu-Jian Huang, Jia-Jun Chen, Lei LI
Non-autoregressive models are promising on various text generation tasks.
no code implementations • 25 Sep 2019 • Yu Bao, Hao Zhou, Jiangtao Feng, Mingxuan Wang, ShuJian Huang, Jiajun Chen, Lei LI
However, position modeling of output words is an essential problem in non-autoregressive text generation.
2 code implementations • 15 Aug 2019 • Jiacheng Yang, Mingxuan Wang, Hao Zhou, Chengqi Zhao, Yong Yu, Wei-Nan Zhang, Lei LI
Our experiments in machine translation show CTNMT gains of up to 3 BLEU score on the WMT14 English-German language pair which even surpasses the previous state-of-the-art pre-training aided NMT by 1. 4 BLEU score.
no code implementations • ACL 2019 • Bingzhen Wei, Mingxuan Wang, Hao Zhou, Junyang Lin, Jun Xie, Xu sun
Non-autoregressive translation models (NAT) have achieved impressive inference speedup.
no code implementations • IJCNLP 2019 • Mingxuan Wang, Jun Xie, Zhixing Tan, Jinsong Su, Deyi Xiong, Lei LI
In this study, we first investigate a novel capsule network with dynamic routing for linear time Neural Machine Translation (NMT), referred as \textsc{CapsNMT}.
no code implementations • WS 2018 • Mingxuan Wang, Li Gong, Wenhuan Zhu, Jun Xie, Chao Bian
We participated in the WMT 2018 shared news translation task on English↔Chinese language pair.
no code implementations • COLING 2018 • Mingxuan Wang, Jun Xie, Zhixing Tan, Jinsong Su, Deyi Xiong, Chao Bian
Neural machine translation with source-side attention have achieved remarkable performance.
1 code implementation • 5 Dec 2017 • Zhixing Tan, Mingxuan Wang, Jun Xie, Yidong Chen, Xiaodong Shi
Semantic Role Labeling (SRL) is believed to be a crucial step towards natural language understanding and has been widely studied.
Ranked #14 on
Semantic Role Labeling
on OntoNotes
no code implementations • ACL 2017 • Jinchao Zhang, Mingxuan Wang, Qun Liu, Jie zhou
This paper proposes three distortion models to explicitly incorporate the word reordering knowledge into attention-based Neural Machine Translation (NMT) for further improving translation performance.
no code implementations • ACL 2017 • Mingxuan Wang, Zhengdong Lu, Jie zhou, Qun Liu
Deep Neural Networks (DNNs) have provably enhanced the state-of-the-art Neural Machine Translation (NMT) with their capability in modeling complex functions and capturing complex linguistic structures.
no code implementations • EMNLP 2016 • Mingxuan Wang, Zhengdong Lu, Hang Li, Qun Liu
We propose to enhance the RNN decoder in a neural machine translator (NMT) with external memory, as a natural but powerful extension to the state in the decoding RNN.
no code implementations • 17 Mar 2015 • Mingxuan Wang, Zhengdong Lu, Hang Li, Wenbin Jiang, Qun Liu
Different from previous work on neural network-based language modeling and generation (e. g., RNN or LSTM), we choose not to greedily summarize the history of words as a fixed length vector.
no code implementations • 9 Mar 2015 • Mingxuan Wang, Zhengdong Lu, Hang Li, Qun Liu
Many tasks in natural language processing, ranging from machine translation to question answering, can be reduced to the problem of matching two sentences or more generally two short texts.
no code implementations • IJCNLP 2015 • Fandong Meng, Zhengdong Lu, Mingxuan Wang, Hang Li, Wenbin Jiang, Qun Liu
The recently proposed neural network joint model (NNJM) (Devlin et al., 2014) augments the n-gram target language model with a heuristically chosen source context window, achieving state-of-the-art performance in SMT.