Getting More Data for Low-resource Morphological Inflection: Language Models and Data Augmentation

LREC 2020  ·  Alexey Sorokin ·

We investigate how to improve quality of low-resource morphological inflection without annotating more data. We examine two methods, language models and data augmentation. We show that the model whose decoder that additionally uses the states of the langauge model improves the model quality by 1.5{\%} in combination with both baselines. We also demonstrate that the augmentation of data improves performance by 9{\%} in average when adding 1000 artificially generated word forms to the dataset.

PDF Abstract

Datasets


  Add Datasets introduced or used in this paper

Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods


No methods listed for this paper. Add relevant methods here