1 code implementation • 19 Dec 2022 • Lester James Miranda, Ákos Kádár, Adriane Boyd, Sofie Van Landeghem, Anders Søgaard, Matthew Honnibal
In this technical report we lay out a bit of history and introduce the embedding methods in spaCy in detail.
1 code implementation • LREC 2022 • Chris Emmery, Ákos Kádár, Grzegorz Chrupała, Walter Daelemans
The perturbed data, models, and code are available for reproduction at https://github. com/cmry/augtox
no code implementations • ACL 2021 • Peng Xu, Wenjie Zi, Hamidreza Shahidi, Ákos Kádár, Keyi Tang, Wei Yang, Jawad Ateeq, Harsh Barot, Meidan Alon, Yanshuai Cao
A natural language database interface (NLDB) can democratize data-driven insights for non-technical users.
1 code implementation • 22 Feb 2021 • Judit Ács, Ákos Kádár, András Kornai
For POS tagging both of these strategies perform poorly and the best choice is to use a small LSTM over the subwords.
1 code implementation • EACL 2021 • Chris Emmery, Ákos Kádár, Grzegorz Chrupała
Written language contains stylistic cues that can be exploited to automatically infer a variety of potentially sensitive author information.
no code implementations • 9 Nov 2019 • Ákos Kádár, Grzegorz Chrupała, Afra Alishahi, Desmond Elliott
However, we do find that using an external machine translation model to generate the synthetic data sets results in better performance.
2 code implementations • NAACL 2019 • Enrique Manjavacas, Ákos Kádár, Mike Kestemont
Lemmatization of standard languages is concerned with (i) abstracting over morphological differences and (ii) resolving token-lemma ambiguities of inflected words in order to map them to a dictionary headword.
1 code implementation • CONLL 2018 • Ákos Kádár, Desmond Elliott, Marc-Alexandre Côté, Grzegorz Chrupała, Afra Alishahi
Recent work has shown how to learn better visual-semantic embeddings by leveraging image descriptions in more than one language.
no code implementations • COLING 2018 • Ákos Kádár, Marc-Alexandre Côté, Grzegorz Chrupała, Afra Alishahi
Hierarchical Multiscale LSTM (Chung et al., 2016a) is a state-of-the-art language model that learns interpretable structure from character-level input.
1 code implementation • 29 Jun 2018 • Marc-Alexandre Côté, Ákos Kádár, Xingdi Yuan, Ben Kybartas, Tavian Barnes, Emery Fine, James Moore, Ruo Yu Tao, Matthew Hausknecht, Layla El Asri, Mahmoud Adada, Wendy Tay, Adam Trischler
We introduce TextWorld, a sandbox learning environment for the training and evaluation of RL agents on text-based games.
1 code implementation • ACL 2018 • Thiago Castro Ferreira, Diego Moussallem, Ákos Kádár, Sander Wubben, Emiel Krahmer
Traditionally, Referring Expression Generation (REG) models first decide on the form and then on the content of references to discourse entities in text, typically relying on features such as salience and grammatical function.
no code implementations • WS 2019 • Grzegorz Chrupała, Lieke Gelderloos, Ákos Kádár, Afra Alishahi
In the domain of unsupervised learning most work on speech has focused on discovering low-level constructs such as phoneme inventories or word-like units.
no code implementations • IJCNLP 2017 • Desmond Elliott, Ákos Kádár
We decompose multimodal translation into two sub-tasks: learning to translate and learning visually grounded representations.
1 code implementation • CL 2017 • Ákos Kádár, Grzegorz Chrupała, Afra Alishahi
We present novel methods for analyzing the activation patterns of RNNs from a linguistic point of view and explore the types of linguistic structure they learn.
1 code implementation • IJCNLP 2015 • Grzegorz Chrupała, Ákos Kádár, Afra Alishahi
We propose Imaginet, a model of learning visually grounded representations of language from coupled textual and visual input.