no code implementations • WMT (EMNLP) 2020 • Qu Cui, Xiang Geng, ShuJian Huang, Jiajun Chen
This paper describes our system of the sentence-level and word-level Quality Estimation Shared Task of WMT20.
no code implementations • WMT (EMNLP) 2021 • Yimeng Chen, Chang Su, Yingtao Zhang, Yuxia Wang, Xiang Geng, Hao Yang, Shimin Tao, Guo Jiaxin, Wang Minghan, Min Zhang, Yujia Liu, ShuJian Huang
This paper presents our work in WMT 2021 Quality Estimation (QE) Shared Task.
no code implementations • 27 Feb 2025 • Xiang Geng, Zhejian Lai, Jiajun Chen, Hao Yang, ShuJian Huang
Quality Estimation (QE) models evaluate the quality of machine translations without reference translations, serving as the reward models for the translation task.
no code implementations • 26 Dec 2024 • Jiawei Yu, Xiang Geng, Yuang Li, Mengxin Ren, Wei Tang, Jiahuan Li, Zhibin Lan, Min Zhang, Hao Yang, ShuJian Huang, Jinsong Su
Spoken named entity recognition (NER) aims to identify named entities from speech, playing an important role in speech processing.
1 code implementation • 22 May 2024 • Xiang Geng, Ming Zhu, Jiahuan Li, Zhejian Lai, Wei Zou, Shuaijie She, Jiaxin Guo, Xiaofeng Zhao, Yinglu Li, Yuang Li, Chang Su, Yanqing Zhao, Xinglin Lyu, Min Zhang, Jiajun Chen, Hao Yang, ShuJian Huang
For the second issue, we propose a method comprising two synergistic components: low-rank adaptation for training to maintain the original LLM parameters, and recovery KD, which utilizes data generated by the chat LLM itself to recover the original knowledge from the frozen parameters.
1 code implementation • 13 Apr 2024 • Wei Zou, Ziyuan Zhuang, Xiang Geng, ShuJian Huang, Jia Liu, Jiajun Chen
Paraphrase generation strives to generate high-quality and diverse expressions of a given text, a domain where diffusion models excel.
no code implementations • 21 Mar 2024 • Haofei Zhao, Yilun Liu, Shimin Tao, Weibin Meng, Yimeng Chen, Xiang Geng, Chang Su, Min Zhang, Hao Yang
Machine Translation Quality Estimation (MTQE) is the task of estimating the quality of machine-translated text in real time without the need for reference translations, which is of great importance for the development of MT.
1 code implementation • 12 Jan 2024 • Xu Huang, Zhirui Zhang, Xiang Geng, Yichao Du, Jiajun Chen, ShuJian Huang
This study investigates how Large Language Models (LLMs) leverage source and reference data in machine translation evaluation task, aiming to better understand the mechanisms behind their remarkable performance in this task.
1 code implementation • 12 Jan 2024 • Shuaijie She, Wei Zou, ShuJian Huang, Wenhao Zhu, Xiang Liu, Xiang Geng, Jiajun Chen
To enhance reasoning abilities in non-dominant languages, we propose a Multilingual-Alignment-as-Preference Optimization framework (MAPO), aiming to align the reasoning processes in other languages with the dominant language.
1 code implementation • 23 Sep 2023 • Xiang Geng, Zhejian Lai, Yu Zhang, Shimin Tao, Hao Yang, Jiajun Chen, ShuJian Huang
We generate pseudo MQM data using parallel data from the WMT translation task.
1 code implementation • 3 Dec 2022 • Shuaijie She, Xiang Geng, ShuJian Huang, Jiajun Chen
To separate the preference for factual consistency, we propose an unsupervised framework named CoP by controlling the preference of the generation model with the help of prompt.
no code implementations • 15 May 2021 • Qu Cui, ShuJian Huang, Jiahuan Li, Xiang Geng, Zaixiang Zheng, Guoping Huang, Jiajun Chen
However, we argue that there are gaps between the predictor and the estimator in both data quality and training objectives, which preclude QE models from benefiting from a large number of parallel corpora more directly.
1 code implementation • 17 Feb 2021 • Bin Gu, Guodong Liu, yanfu Zhang, Xiang Geng, Heng Huang
Modern machine learning algorithms usually involve tuning multiple (from one to thousands) hyperparameters which play a pivotal role in terms of model generalizability.
no code implementations • 29 Jul 2019 • Wanli Shi, Bin Gu, Xiang Li, Xiang Geng, Heng Huang
To address this problem, in this paper, we propose a novel scalable quadruply stochastic gradient algorithm (QSG-S2AUC) for nonlinear semi-supervised AUC optimization.
no code implementations • 26 Jul 2019 • Xiang Geng, Bin Gu, Xiang Li, Wanli Shi, Guansheng Zheng, Heng Huang
Specifically, to handle two types of data instances involved in S$^3$VM, TSGS$^3$VM samples a labeled instance and an unlabeled instance as well with the random features in each iteration to compute a triply stochastic gradient.