1 code implementation • 20 Oct 2023 • Wenyu Guo, Qingkai Fang, Dong Yu, Yang Feng
Multimodal machine translation (MMT) simultaneously takes the source sentence and a relevant image as input for translation.
1 code implementation • NeurIPS 2023 • Qingkai Fang, Yan Zhou, Yang Feng
However, due to the presence of linguistic and acoustic diversity, the target speech follows a complex multimodal distribution, posing challenges to achieving both high-quality translations and fast decoding speeds for S2ST models.
1 code implementation • 19 Jun 2023 • Shaolei Zhang, Qingkai Fang, Zhuocheng Zhang, Zhengrui Ma, Yan Zhou, Langlin Huang, Mengyu Bu, Shangtong Gui, Yunji Chen, Xilin Chen, Yang Feng
To minimize human workload, we propose to transfer the capabilities of language generation and instruction following from English to other languages through an interactive translation task.
2 code implementations • 24 May 2023 • Yan Zhou, Qingkai Fang, Yang Feng
End-to-end speech translation (ST) is the task of translating speech signals in the source language into text in the target language.
1 code implementation • 15 May 2023 • Qingkai Fang, Yang Feng
Motivated by the remarkable success of back translation in MT, we develop a back translation algorithm for ST (BT4ST) to synthesize pseudo ST data from monolingual target data.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +4
1 code implementation • 15 May 2023 • Qingkai Fang, Yang Feng
However, due to the differences between speech and text, there is always a gap between ST and MT.
1 code implementation • 13 Oct 2022 • Zhe Yang, Qingkai Fang, Yang Feng
How to achieve neural machine translation with limited parallel data?
Contrastive Learning Low-Resource Neural Machine Translation +2
1 code implementation • ACL 2022 • Qingkai Fang, Rong Ye, Lei LI, Yang Feng, Mingxuan Wang
How to learn a better speech representation for end-to-end speech-to-text translation (ST) with limited labeled data?
1 code implementation • ACL 2022 • Qingkai Fang, Yang Feng
Multimodal machine translation (MMT) aims to improve neural machine translation (NMT) with additional visual information, but most existing MMT methods require paired input of source sentence and image, which makes them suffer from shortage of sentence-image pairs.