Getting More Data for Low-resource Morphological Inflection: Language Models and Data Augmentation
We investigate how to improve quality of low-resource morphological inflection without annotating more data. We examine two methods, language models and data augmentation. We show that the model whose decoder that additionally uses the states of the langauge model improves the model quality by 1.5{\%} in combination with both baselines. We also demonstrate that the augmentation of data improves performance by 9{\%} in average when adding 1000 artificially generated word forms to the dataset.
PDF AbstractDatasets
Add Datasets
introduced or used in this paper
Results from the Paper
Submit
results from this paper
to get state-of-the-art GitHub badges and help the
community compare results to other papers.
Methods
No methods listed for this paper. Add
relevant methods here