no code implementations • NoDaLiDa 2021 • Hasan Tanvir, Claudia Kittask, Sandra Eiche, Kairit Sirts
This paper presents EstBERT, a large pretrained transformer-based language-specific BERT model for Estonian.
no code implementations • 1 Oct 2020 • Claudia Kittask, Kirill Milintsevich, Kairit Sirts
Recently, large pre-trained language models, such as BERT, have reached state-of-the-art performance in many natural language processing tasks, but for many languages, including Estonian, BERT models are not yet available.
no code implementations • RANLP 2019 • Claudia Kittask, Eduard Barbu
Researchers in Computational Linguistics build models of similarity and test them against human judgments.