Vega-MT: The JD Explore Academy Translation System for WMT22

We describe the JD Explore Academy's submission of the WMT 2022 shared general translation task. We participated in all high-resource tracks and one medium-resource track, including Chinese-English, German-English, Czech-English, Russian-English, and Japanese-English. We push the limit of our previous work -- bidirectional training for translation by scaling up two main factors, i.e. language pairs and model sizes, namely the \textbf{Vega-MT} system. As for language pairs, we scale the "bidirectional" up to the "multidirectional" settings, covering all participating languages, to exploit the common knowledge across languages, and transfer them to the downstream bilingual tasks. As for model sizes, we scale the Transformer-Big up to the extremely large model that owns nearly 4.7 Billion parameters, to fully enhance the model capacity for our Vega-MT. Also, we adopt the data augmentation strategies, e.g. cycle translation for monolingual data, and bidirectional self-training for bilingual and monolingual data, to comprehensively exploit the bilingual and monolingual data. To adapt our Vega-MT to the general domain test set, generalization tuning is designed. Based on the official automatic scores of constrained systems, in terms of the sacreBLEU shown in Figure-1, we got the 1st place on {Zh-En (33.5), En-Zh (49.7), De-En (33.7), En-De (37.8), Cs-En (54.9), En-Cs (41.4) and En-Ru (32.7)}, 2nd place on {Ru-En (45.1) and Ja-En (25.6)}, and 3rd place on {En-Ja(41.5)}, respectively; W.R.T the COMET, we got the 1st place on {Zh-En (45.1), En-Zh (61.7), De-En (58.0), En-De (63.2), Cs-En (74.7), Ru-En (64.9), En-Ru (69.6) and En-Ja (65.1)}, 2nd place on {En-Cs (95.3) and Ja-En (40.6)}, respectively.

PDF Abstract

Datasets


  Add Datasets introduced or used in this paper
Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Machine Translation WMT 2022 Chinese-English Vega-MT SacreBLEU 33.5 # 1
Machine Translation WMT 2022 Czech-English Vega-MT SacreBLEU 54.9 # 1
Machine Translation WMT 2022 English-Chinese Vega-MT SacreBLEU 49.7 # 1
Machine Translation WMT 2022 English-Czech Vega-MT SacreBLEU 41.4 # 1
Machine Translation WMT 2022 English-German Vega-MT SacreBLEU 37.8 # 1
Machine Translation WMT 2022 English-Japanese Vega-MT SacreBLEU 41.5 # 1
Machine Translation WMT 2022 English-Russian Vega-MT SacreBLEU 32.7 # 1
Machine Translation WMT 2022 German-English Vega-MT SacreBLEU 33.7 # 1
Machine Translation WMT 2022 Japanese-English Vega-MT SacreBLEU 25.6 # 1
Machine Translation WMT 2022 Russian-English Vega-MT SacreBLEU 45.1 # 1

Methods