no code implementations • EACL (BEA) 2021 • Anisia Katinskaia, Roman Yangarber
We approach the problem with the methods for grammatical error detection (GED), since we hypothesize that models for detecting grammatical mistakes can assess the correctness of potential alternative answers in a learning setting.
no code implementations • LREC 2022 • Anisia Katinskaia, Maria Lebedeva, Jue Hou, Roman Yangarber
We present ReLCo— the Revita Learner Corpus—a new semi-automatically annotated learner corpus for Russian.
no code implementations • WS (NoDaLiDa) 2019 • Sardana Ivanova, Anisia Katinskaia, Roman Yangarber
Revita is a freely available online language learning platform for learners beyond the beginner level.
no code implementations • games (LREC) 2022 • Jue Hou, Ilmari Kylliäinen, Anisia Katinskaia, Giacomo Furlan, Roman Yangarber
Our goal is to keep the learner engaged in long practice sessions over many months—rather than for the short-term.
no code implementations • 24 Sep 2024 • Jue Hou, Anisia Katinskaia, Anh-Duc Vu, Roman Yangarber
Exhaustive testing across a wide range of skills can provide a detailed picture of proficiency, but may be undesirable for a number of reasons.
no code implementations • 4 Jun 2024 • Anisia Katinskaia, Roman Yangarber
We perform probing using BERT and RoBERTa on alternative and non-alternative contexts.
no code implementations • 14 May 2024 • Anisia Katinskaia, Roman Yangarber
This paper investigates the application of GPT-3. 5 for Grammatical Error Correction (GEC) in multiple languages in several settings: zero-shot GEC, fine-tuning for GEC, and using GPT-3. 5 to re-rank correction hypotheses generated by other GEC models.
Ranked #5 on
Grammatical Error Correction
on UA-GEC
1 code implementation • 22 Apr 2024 • Jue Hou, Anisia Katinskaia, Lari Kotilainen, Sathianpong Trangcasanchai, Anh-Duc Vu, Roman Yangarber
This paper investigates what insights about linguistic features and what knowledge about the structure of natural language can be obtained from the encodings in transformer language models. In particular, we explore how BERT encodes the government relation between constituents in a sentence.
no code implementations • 9 May 2023 • Jue Hou, Anisia Katinskaia, Anh-Duc Vu, Roman Yangarber
Lastly, we show 4. that LMs of smaller size using morphological segmentation can perform comparably to models of larger size trained with BPE -- both in terms of (1) perplexity and (3) scores on downstream tasks.
no code implementations • 3 Dec 2022 • Anisia Katinskaia, Jue Hou, Anh-Duc Vu, Roman Yangarber
This paper presents the development of an AI-based language learning platform Revita.
no code implementations • LREC 2020 • Anisia Katinskaia, Sardana Ivanova, Roman Yangarber
We present the first version of the longitudinal Revita Learner Corpus (ReLCo), for Russian.
no code implementations • LREC 2020 • Christos Rodosthenous, Verena Lyding, Federico Sangati, Alex K{\"o}nig, er, Umair ul Hassan, Lionel Nicolas, Jolita Horbacauskiene, Anisia Katinskaia, Lavinia Aparaschivei
In this work, we report on a crowdsourcing experiment conducted using the V-TREL vocabulary trainer which is accessed via a Telegram chatbot interface to gather knowledge on word relations suitable for expanding ConceptNet.
no code implementations • LREC 2020 • Lionel Nicolas, Verena Lyding, Claudia Borg, Corina Forascu, Kar{\"e}n Fort, Katerina Zdravkova, Iztok Kosem, Jaka {\v{C}}ibej, {\v{S}}pela Arhar Holdt, Alice Millour, Alex K{\"o}nig, er, Christos Rodosthenous, Federico Sangati, Umair ul Hassan, Anisia Katinskaia, Anabela Barreiro, Lavinia Aparaschivei, Yaakov HaCohen-Kerner
We introduce in this paper a generic approach to combine implicit crowdsourcing and language learning in order to mass-produce language resources (LRs) for any language for which a crowd of language learners can be involved.
no code implementations • RANLP 2019 • Verena Lyding, Christos Rodosthenous, Federico Sangati, Umair ul Hassan, Lionel Nicolas, Alex K{\"o}nig, er, Jolita Horbacauskiene, Anisia Katinskaia
In this paper, we present our work on developing a vocabulary trainer that uses exercises generated from language resources such as ConceptNet and crowdsources the responses of the learners to enrich the language resource.
no code implementations • WS 2019 • Anisia Katinskaia, Sardana Ivanova
We present our work on the problem of Multiple Admissibility (MA) in language learning.
no code implementations • EACL 2017 • Lloren{\c{c}} Escoter, Lidia Pivovarova, Mian Du, Anisia Katinskaia, Roman Yangarber
In news aggregation systems focused on broad news domains, certain stories may appear in multiple articles.