no code implementations • 16 Dec 2022 • Lucie Charlotte Magister, Jonathan Mallinson, Jakub Adamek, Eric Malmi, Aliaksei Severyn
Chain of thought prompting successfully improves the reasoning capabilities of large language models, achieving state of the art results on a range of datasets.
no code implementations • NAACL (ACL) 2022 • Eric Malmi, Yue Dong, Jonathan Mallinson, Aleksandr Chuklin, Jakub Adamek, Daniil Mirylenka, Felix Stahlberg, Sebastian Krause, Shankar Kumar, Aliaksei Severyn
Text-editing models have recently become a prominent alternative to seq2seq models for monolingual text-generation tasks such as grammatical error correction, simplification, and style transfer.
no code implementations • 24 May 2022 • Jonathan Mallinson, Jakub Adamek, Eric Malmi, Aliaksei Severyn
This is achieved by decomposing the generation process into three sub-tasks: (1) tagging to decide on the subset of input tokens to be preserved in the output, (2) re-ordering to define their order in the output text, and (3) insertion to infill the missing tokens that are not present in the input.
1 code implementation • NeurIPS 2021 • Sachin Kumar, Eric Malmi, Aliaksei Severyn, Yulia Tsvetkov
As large-scale language model pretraining pushes the state-of-the-art in text generation, recent work has turned to controlling attributes of the text such models generate.
1 code implementation • ACL 2021 • Sascha Rothe, Jonathan Mallinson, Eric Malmi, Sebastian Krause, Aliaksei Severyn
This paper presents a simple recipe to train state-of-the-art multilingual Grammatical Error Correction (GEC) models.
Ranked #1 on
Grammatical Error Correction
on Falko-MERLIN
(using extra training data)
no code implementations • EMNLP 2020 • Eric Malmi, Aliaksei Severyn, Sascha Rothe
This allows us to identify the source tokens to delete to transform the source text to match the style of the target domain.
2 code implementations • Findings of the Association for Computational Linguistics 2020 • Jonathan Mallinson, Aliaksei Severyn, Eric Malmi, Guillermo Garrido
We achieve this by decomposing the text-editing task into two sub-tasks: tagging to decide on the subset of input tokens and their order in the output text and insertion to in-fill the missing tokens in the output not present in the input.
5 code implementations • IJCNLP 2019 • Eric Malmi, Sebastian Krause, Sascha Rothe, Daniil Mirylenka, Aliaksei Severyn
We propose LaserTagger - a sequence tagging approach that casts text generation as a text editing task.
Ranked #1 on
Sentence Fusion
on DiscoFuse
6 code implementations • TACL 2020 • Sascha Rothe, Shashi Narayan, Aliaksei Severyn
Unsupervised pre-training of large neural models has recently revolutionized Natural Language Processing.
Ranked #1 on
Split and Rephrase
on WikiSplit
no code implementations • 21 Feb 2019 • Octavian-Eugen Ganea, Sylvain Gelly, Gary Bécigneul, Aliaksei Severyn
The Softmax function on top of a final linear layer is the de facto method to output probability distributions in neural networks.
no code implementations • 14 Aug 2018 • Heike Adel, Anton Bryl, David Weiss, Aliaksei Severyn
We study cross-lingual sequence tagging with little or no labeled data in the target language.
no code implementations • 13 Jun 2018 • Stanislau Semeniuta, Aliaksei Severyn, Sylvain Gelly
Generative Adversarial Networks (GANs) are a promising approach to language generation.
2 code implementations • 11 Jun 2018 • Aleksandr Chuklin, Aliaksei Severyn, Johanne Trippas, Enrique Alfonseca, Hanna Silen, Damiano Spina
Many popular form factors of digital assistants---such as Amazon Echo, Apple Homepod, or Google Home---enable the user to hold a conversation with these systems based only on the speech modality.
no code implementations • 21 Apr 2018 • Ondřej Cífka, Aliaksei Severyn, Enrique Alfonseca, Katja Filippova
In this paper, we study recent neural generative models for text generation related to variational autoencoders.
1 code implementation • 30 Nov 2017 • Mostafa Dehghani, Aliaksei Severyn, Sascha Rothe, Jaap Kamps
In this paper, we propose a method for training neural networks when we have a large set of data with weak labels and a small amount of data with true labels.
no code implementations • 1 Nov 2017 • Mostafa Dehghani, Aliaksei Severyn, Sascha Rothe, Jaap Kamps
Thus we avoid that the weight updates computed from noisy labels harm the quality of the target network model.
1 code implementation • 28 Apr 2017 • Mostafa Dehghani, Hamed Zamani, Aliaksei Severyn, Jaap Kamps, W. Bruce Croft
Our experiments indicate that employing proper objective functions and letting the networks to learn the input representation based on weakly supervised data leads to impressive performance, with over 13% and 35% MAP improvements over the BM25 model on the Robust and the ClueWeb collections.
Ranked #8 on
Ad-Hoc Information Retrieval
on TREC Robust04
(MAP metric)
1 code implementation • 7 Mar 2017 • Jan Deriu, Aurelien Lucchi, Valeria De Luca, Aliaksei Severyn, Simon Müller, Mark Cieliebak, Thomas Hofmann, Martin Jaggi
This paper presents a novel approach for multi-lingual sentiment classification in short texts.
3 code implementations • EMNLP 2017 • Stanislau Semeniuta, Aliaksei Severyn, Erhardt Barth
In this paper we explore the effect of architectural choices on learning a Variational Autoencoder (VAE) for text generation.
no code implementations • 5 Apr 2016 • Aliaksei Severyn, Alessandro Moschitti
In this paper, we propose convolutional neural networks for learning an optimal representation of question and answer sentences.
1 code implementation • ACL 2016 • Daniel Andor, Chris Alberti, David Weiss, Aliaksei Severyn, Alessandro Presta, Kuzman Ganchev, Slav Petrov, Michael Collins
Our model is a simple feed-forward neural network that operates on a task-specific transition system, yet achieves comparable or better accuracies than recurrent models.
Ranked #17 on
Dependency Parsing
on Penn Treebank
2 code implementations • COLING 2016 • Stanislau Semeniuta, Aliaksei Severyn, Erhardt Barth
This paper presents a novel approach to recurrent neural network (RNN) regularization.
no code implementations • LREC 2014 • Olga Uryupina, Barbara Plank, Aliaksei Severyn, Agata Rotondi, Aless Moschitti, ro
In this paper we present SenTube -- a dataset of user-generated comments on YouTube videos annotated for information content and sentiment polarity.