Low-Resource Neural Machine Translation

23 papers with code • 1 benchmarks • 4 datasets

Low-resource machine translation is the task of machine translation on a low-resource language where large data may not be available.

Low-resource neural machine translation with morphological modeling

anzeyimana/kinmt_naacl2024 3 Apr 2024

An attention augmentation scheme to the transformer model is proposed in a generic form to allow integration of pre-trained language models and also facilitate modeling of word order relationships between the source and target languages.

0
03 Apr 2024

On Bilingual Lexicon Induction with Large Language Models

cambridgeltl/prompt4bli 21 Oct 2023

Bilingual Lexicon Induction (BLI) is a core task in multilingual NLP that still, to a large extent, relies on calculating cross-lingual word representations.

8
21 Oct 2023

ConsistTL: Modeling Consistency in Transfer Learning for Low-Resource Neural Machine Translation

nlp2ct/consisttl 8 Dec 2022

In this paper, we propose a novel transfer learning method for NMT, namely ConsistTL, which can continuously transfer knowledge from the parent model during the training of the child model.

17
08 Dec 2022

Low-resource Neural Machine Translation with Cross-modal Alignment

ictnlp/lnmt-ca 13 Oct 2022

How to achieve neural machine translation with limited parallel data?

14
13 Oct 2022

On the Complementarity between Pre-Training and Random-Initialization for Resource-Rich Machine Translation

zanchangtong/ptvsri COLING 2022

Pre-Training (PT) of text representations has been successfully applied to low-resource Neural Machine Translation (NMT).

8
07 Sep 2022

Geographical Distance Is The New Hyperparameter: A Case Study Of Finding The Optimal Pre-trained Language For English-isiZulu Machine Translation

umair-nasir14/ngdc 17 May 2022

Stemming from the limited availability of datasets and textual resources for low-resource languages such as isiZulu, there is a significant need to be able to harness knowledge from pre-trained models to improve low resource machine translation.

0
17 May 2022

Linguistically-driven Multi-task Pre-training for Low-resource Neural Machine Translation

Mao-KU/JASS 20 Jan 2022

In the present study, we propose novel sequence-to-sequence pre-training objectives for low-resource machine translation (NMT): Japanese-specific sequence to sequence (JASS) for language pairs involving Japanese as the source or target language, and English-specific sequence to sequence (ENSS) for language pairs involving English.

16
20 Jan 2022

Sicilian Translator: A Recipe for Low-Resource NMT

ewdowiak/Sicilian_Translator 5 Oct 2021

With 17, 000 pairs of Sicilian-English translated sentences, Arba Sicula developed the first neural machine translator for the Sicilian language.

4
05 Oct 2021

Rethinking Data Augmentation for Low-Resource Neural Machine Translation: A Multi-Task Learning Approach

transducens/mtl-da-emnlp EMNLP 2021

Many DA approaches aim at expanding the support of the empirical data distribution by generating new sentence pairs that contain infrequent words, thus making it closer to the true data distribution of parallel sentences.

5
08 Sep 2021

Meta-Curriculum Learning for Domain Adaptation in Neural Machine Translation

NLP2CT/Meta-Curriculum 3 Mar 2021

Meta-learning has been sufficiently validated to be beneficial for low-resource neural machine translation (NMT).

23
03 Mar 2021