no code implementations • WMT (EMNLP) 2020 • Minghan Wang, Hao Yang, Hengchao Shang, Daimeng Wei, Jiaxin Guo, Lizhi Lei, Ying Qin, Shimin Tao, Shiliang Sun, Yimeng Chen, Liangyou Li
This paper presents our work in the WMT 2020 Word and Sentence-Level Post-Editing Quality Estimation (QE) Shared Task.
no code implementations • ACL (IWSLT) 2021 • Xingshan Zeng, Liangyou Li, Qun Liu
We use a unified transformer architecture for our MultiST model, so that the data from different modalities (i. e., speech and text) and different tasks (i. e., Speech Recognition, Machine Translation, and Speech Translation) can be exploited to enhance the model’s ability.
no code implementations • NAACL (AutoSimTrans) 2022 • Xingshan Zeng, Pengfei Li, Liangyou Li, Qun Liu
This paper describes the system submitted to AutoSimTrans 2022 from Huawei Noah’s Ark Lab, which won the first place in the audio input track of the Chinese-English translation task.
no code implementations • WMT (EMNLP) 2021 • Meng Zhang, Minghao Wu, Pengfei Li, Liangyou Li, Qun Liu
This paper describes the NoahNMT system submitted to the WMT 2021 shared task of Very Low Resource Supervised Machine Translation.
no code implementations • AACL (WAT) 2020 • Zhengzhe Yu, Zhanglin Wu, Xiaoyu Chen, Daimeng Wei, Hengchao Shang, Jiaxin Guo, Zongyao Li, Minghan Wang, Liangyou Li, Lizhi Lei, Hao Yang, Ying Qin
This paper describes our work in the WAT 2020 Indic Multilingual Translation Task.
no code implementations • WMT (EMNLP) 2020 • Wei Peng, Jianfeng Liu, Minghan Wang, Liangyou Li, Xupeng Meng, Hao Yang, Qun Liu
This paper describes Huawei’s submissions to the WMT20 biomedical translation shared task.
no code implementations • WMT (EMNLP) 2020 • Daimeng Wei, Hengchao Shang, Zhanglin Wu, Zhengzhe Yu, Liangyou Li, Jiaxin Guo, Minghan Wang, Hao Yang, Lizhi Lei, Ying Qin, Shiliang Sun
We also conduct experiment with similar language augmentation, which lead to positive results, although not used in our submission.
1 code implementation • 14 Aug 2024 • Yuxin Jiang, Bo Huang, YuFei Wang, Xingshan Zeng, Liangyou Li, Yasheng Wang, Xin Jiang, Lifeng Shang, Ruiming Tang, Wei Wang
Direct preference optimization (DPO), a widely adopted offline preference optimization algorithm, aims to align large language models (LLMs) with human-desired behaviors using pairwise preference data.
no code implementations • 23 Jun 2024 • Zezhong Wang, Xingshan Zeng, Weiwen Liu, YuFei Wang, Liangyou Li, Yasheng Wang, Lifeng Shang, Xin Jiang, Qun Liu, Kam-Fai Wong
To address these questions, we propose a method, namely Chain-of-Probe (CoP), to probe changes in the mind during the model's reasoning.
no code implementations • 17 Jun 2024 • Minda Hu, Bowei He, YuFei Wang, Liangyou Li, Chen Ma, Irwin King
Large language models (LLMs) have demonstrated remarkable performance on various natural language processing tasks.
1 code implementation • 19 Feb 2024 • Yuxin Jiang, YuFei Wang, Chuhan Wu, Wanjun Zhong, Xingshan Zeng, Jiahui Gao, Liangyou Li, Xin Jiang, Lifeng Shang, Ruiming Tang, Qun Liu, Wei Wang
Knowledge editing techniques, aiming to efficiently modify a minor proportion of knowledge in large language models (LLMs) without negatively impacting performance across other inputs, have garnered widespread attention.
1 code implementation • 30 Jan 2024 • Wai-Chung Kwan, Xingshan Zeng, Yuxin Jiang, YuFei Wang, Liangyou Li, Lifeng Shang, Xin Jiang, Qun Liu, Kam-Fai Wong
Large language models (LLMs) are increasingly relied upon for complex multi-turn conversations across diverse real-world applications.
1 code implementation • 31 Oct 2023 • Yuxin Jiang, YuFei Wang, Xingshan Zeng, Wanjun Zhong, Liangyou Li, Fei Mi, Lifeng Shang, Xin Jiang, Qun Liu, Wei Wang
To fill this research gap, in this paper, we propose FollowBench, a Multi-level Fine-grained Constraints Following Benchmark for LLMs.
1 code implementation • 30 Oct 2023 • Wai-Chung Kwan, Xingshan Zeng, YuFei Wang, Yusen Sun, Liangyou Li, Lifeng Shang, Qun Liu, Kam-Fai Wong
In this paper, we propose M4LE, a Multi-ability, Multi-range, Multi-task, Multi-domain benchmark for Long-context Evaluation.
1 code implementation • 24 Jul 2023 • YuFei Wang, Wanjun Zhong, Liangyou Li, Fei Mi, Xingshan Zeng, Wenyong Huang, Lifeng Shang, Xin Jiang, Qun Liu
(2) Training methodologies: a detailed review of the prevailing training methods employed for LLM alignment.
no code implementations • 3 May 2023 • Hao Cheng, Meng Zhang, Liangyou Li, Qun Liu, Zhihua Zhang
Utilizing pivot language effectively can significantly improve low-resource machine translation.
no code implementations • 3 May 2023 • Hao Cheng, Meng Zhang, Weixuan Wang, Liangyou Li, Qun Liu, Zhihua Zhang
We can use automatic summarization or machine translation evaluation metrics for length-controllable machine translation, but this is not necessarily suitable and accurate.
no code implementations • 17 Dec 2022 • Xingshan Zeng, Liangyou Li, Qun Liu
To alleviate the data scarcity problem in End-to-end speech translation (ST), pre-training on data for speech recognition and machine translation is considered as an important technique.
1 code implementation • 28 Nov 2022 • Yusen Sun, Liangyou Li, Qun Liu, Dit-yan Yeung
Although lyrics generation has achieved significant progress in recent years, it has limited practical applications because the generated lyrics cannot be performed without composing compatible melodies.
no code implementations • Findings (NAACL) 2022 • Yinpeng Guo, Liangyou Li, Xin Jiang, Qun Liu
However, labeled cross-lingual corpus is expensive or even inaccessible, especially in the fields where labels are private, such as diagnostic results of symptoms in medicine and user profiles in business.
no code implementations • ACL 2022 • Meng Zhang, Liangyou Li, Qun Liu
Triangular machine translation is a special case of low-resource machine translation where the language pair of interest has limited parallel data, but both languages have abundant parallel data with a pivot language.
1 code implementation • ACL 2022 • Pengfei Li, Liangyou Li, Meng Zhang, Minghao Wu, Qun Liu
To the best of our knowledge, this is the first work to pre-train a unified model for fine-tuning on both NMT tasks.
no code implementations • 7 Sep 2021 • Zhiyuan Zhang, Ruixuan Luo, Xuancheng Ren, Qi Su, Liangyou Li, Xu sun
To enhance neural networks, we propose the adversarial parameter defense algorithm that minimizes the average risk of multiple adversarial parameter corruptions.
no code implementations • EMNLP 2021 • Minghao Wu, Yitong Li, Meng Zhang, Liangyou Li, Gholamreza Haffari, Qun Liu
In this work, we propose an approach, MultiUAT, that dynamically adjusts the training data usage based on the model's uncertainty on a small set of trusted clean data for multi-corpus machine translation.
no code implementations • 9 Aug 2021 • Minghan Wang, Yuxia Wang, Chang Su, Jiaxin Guo, Yingtao Zhang, Yujia Liu, Min Zhang, Shimin Tao, Xingshan Zeng, Liangyou Li, Hao Yang, Ying Qin
This paper describes our work in participation of the IWSLT-2021 offline speech translation task.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +5
no code implementations • Findings (ACL) 2021 • Xingshan Zeng, Liangyou Li, Qun Liu
To bridge the modality gap between speech and text, RealTranS gradually downsamples the input speech with interleaved convolution and unidirectional Transformer layers for acoustic modeling, and then maps speech features into text space with a weighted-shrinking operation and a semantic encoder.
no code implementations • 9 Jun 2021 • Yinpeng Guo, Liangyou Li, Xin Jiang, Qun Liu
Recently, pre-training multilingual language models has shown great potential in learning multilingual representation, a crucial topic of natural language processing.
no code implementations • 1 Jun 2021 • Xingshan Zeng, Liangyou Li, Qun Liu
We use a unified transformer architecture for our MultiST model, so that the data from different modalities (i. e., speech and text) and different tasks (i. e., Speech Recognition, Machine Translation, and Speech Translation) can be exploited to enhance the model's ability.
no code implementations • 25 Mar 2021 • Tong Cui, Jinghui Xiao, Liangyou Li, Xin Jiang, Qun Liu
Speech-enabled systems typically first convert audio to text through an automatic speech recognition (ASR) model and then feed the text to downstream natural language processing (NLP) modules.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +5
no code implementations • 20 Mar 2021 • Liangyou Li, Andy Way, Qun Liu
We present graph-based translation models which translate source graphs into target strings.
no code implementations • 23 Dec 2020 • Shaolei Zhang, Yang Feng, Liangyou Li
Simultaneous translation (ST) starts translations synchronously while reading source sentences, and is used in many online scenarios.
no code implementations • EMNLP 2021 • Mingzhou Xu, Liangyou Li, Derek. F. Wong, Qun Liu, Lidia S. Chao
Previous works have shown that contextual information can improve the performance of neural machine translation (NMT).
1 code implementation • 10 Jun 2020 • Xu Sun, Zhiyuan Zhang, Xuancheng Ren, Ruixuan Luo, Liangyou Li
We argue that the vulnerability of model parameters is of crucial value to the study of model robustness and generalization but little research has been devoted to understanding this matter.
no code implementations • Asian Chapter of the Association for Computational Linguistics 2020 • Yun Chen, Liangyou Li, Xin Jiang, Xiao Chen, Qun Liu
Despite the success of neural machine translation (NMT), simultaneous neural machine translation (SNMT), the task of translating in real time before a full sentence has been observed, remains challenging due to the syntactic structure difference and simultaneity requirements.
no code implementations • 8 Nov 2019 • Liangyou Li, Xin Jiang, Qun Liu
Previous work on document-level NMT usually focuses on limited contexts because of degraded performance on larger contexts.
no code implementations • WS 2019 • Wei Peng, Jianfeng Liu, Liangyou Li, Qun Liu
This paper describes Huawei{'}s neural machine translation systems for the WMT 2019 biomedical translation shared task.
no code implementations • IJCNLP 2017 • Long-Yue Wang, Jinhua Du, Liangyou Li, Zhaopeng Tu, Andy Way, Qun Liu
We showcase TODAY, a semantics-enhanced task-oriented dialogue translation system, whose novelties are: (i) task-oriented named entity (NE) definition and a hybrid strategy for NE recognition and translation; and (ii) a novel grounded semantic method for dialogue understanding and task-order management.
no code implementations • EACL 2017 • Liangyou Li, Andy Way, Qun Liu
In this paper, we present an improved graph-based translation model which segments an input graph into node-induced subgraphs by taking source context into consideration.
no code implementations • COLING 2016 • Jian Zhang, Liangyou Li, Andy Way, Qun Liu
In recent years, neural machine translation (NMT) has demonstrated state-of-the-art machine translation (MT) performance.