no code implementations • ACL 2020 • Ionut-Teodor Sorodoc, Kristina Gulordava, Gemma Boleda
Language models keep track of complex information about the preceding context {--} including, e. g., syntactic relations in a sentence.
no code implementations • 8 Apr 2020 • Kristina Gulordava, Thomas Brochhagen, Gemma Boleda
We find that constraints in both learning and selection can foster mutual exclusivity, as long as they put words in competition for lexical meaning.
1 code implementation • ACL 2019 • Alex Kabbach, re, Kristina Gulordava, Aur{\'e}lie Herbelot
In this paper, we investigate the task of learning word embeddings from very sparse data in an incremental, cognitively-plausible way.
1 code implementation • ACL 2019 • Laura Aina, Kristina Gulordava, Gemma Boleda
In neural network models of language, words are commonly represented using context-invariant representations (word embeddings) which are then put in context in the hidden layers.
no code implementations • EMNLP 2018 • Kristina Gulordava, Laura Aina, Gemma Boleda
Recent state-of-the-art neural language models share the representations of words given by the input and output mappings.
2 code implementations • NAACL 2018 • Kristina Gulordava, Piotr Bojanowski, Edouard Grave, Tal Linzen, Marco Baroni
Recurrent neural networks (RNNs) have achieved impressive results in a variety of linguistic processing tasks, suggesting that they can induce non-trivial properties of language.
no code implementations • LREC 2016 • Sharid Lo{\'a}iciga, Kristina Gulordava
In this paper, we focus on the verb-particle (V-Prt) split construction in English and German and its difficulty for parsing and Machine Translation (MT).
no code implementations • TACL 2016 • Kristina Gulordava, Paola Merlo
We propose a method to evaluate the effects of word order of a language on dependency parsing performance, while controlling for confounding treebank properties.