1 code implementation • Findings (EMNLP) 2021 • Nicole Peinelt, Marek Rei, Maria Liakata
Large pre-trained language models such as BERT have been the driving force behind recent improvements across many NLP tasks.
1 code implementation • 7 Nov 2024 • Lisa Alazraki, Marek Rei
We present Tool selECTion via meta-reasONing (TECTON), a two-phase system that first reasons over a task using a custom fine-tuned LM head and outputs candidate tools.
no code implementations • 28 Jun 2024 • Zhenhao Li, Marek Rei, Lucia Specia
Pretrained language models have significantly advanced performance across various natural language processing tasks.
1 code implementation • 25 Jun 2024 • Matthieu Meeus, Igor Shilov, Shubham Jain, Manuel Faysse, Marek Rei, Yves-Alexandre de Montjoye
In the first part, we review the literature on MIAs against LLMs.
1 code implementation • 19 May 2024 • Mireia Hernandez Caralt, Clarence Boon Liang Ng, Marek Rei
Electronic Health Records (EHR) serve as a valuable source of patient information, offering insights into medical histories, treatments, and outcomes.
no code implementations • 15 Jan 2024 • Christopher Davis, Andrew Caines, Øistein Andersen, Shiva Taslimipoor, Helen Yannakoudakis, Zheng Yuan, Christopher Bryant, Marek Rei, Paula Buttery
Thanks to recent advances in generative AI, we are able to prompt large language models (LLMs) to produce texts which are fluent and grammatical.
1 code implementation • 23 Oct 2023 • Matthieu Meeus, Shubham Jain, Marek Rei, Yves-Alexandre de Montjoye
First, we propose a procedure for the development and evaluation of document-level membership inference for LLMs by leveraging commonly used data sources for training and the model release date.
no code implementations • 17 Jul 2023 • Andrew Caines, Luca Benedetto, Shiva Taslimipoor, Christopher Davis, Yuan Gao, Oeistein Andersen, Zheng Yuan, Mark Elliott, Russell Moore, Christopher Bryant, Marek Rei, Helen Yannakoudakis, Andrew Mullooly, Diane Nicholls, Paula Buttery
The recent release of very large language models such as PaLM and GPT-4 has made an unprecedented impact in the popular media and public consciousness, giving rise to a mixture of excitement and fear as to their capabilities and potential uses, and shining a light on natural language processing research which had not previously received so much attention.
1 code implementation • 22 May 2023 • Joe Stacey, Pasquale Minervini, Haim Dubossarsky, Oana-Maria Camburu, Marek Rei
With recent advances, neural models can achieve human-level performance on various natural language tasks.
no code implementations • 22 May 2023 • Joe Stacey, Marek Rei
Knowledge distillation optimises a smaller student model to behave similarly to a larger teacher model, retaining some of the performance benefits.
no code implementations • 14 Mar 2023 • Kamil Bujel, Andrew Caines, Helen Yannakoudakis, Marek Rei
Long-sequence transformers are designed to improve the representation of longer texts by language models and their performance on downstream document-level tasks.
no code implementations • 24 Feb 2023 • Clarence Boon Liang Ng, Diogo Santos, Marek Rei
Past studies on the ICD coding problem focus on predicting clinical codes primarily based on the discharge summary.
2 code implementations • 12 Feb 2023 • Stuart Mesham, Christopher Bryant, Marek Rei, Zheng Yuan
We extend a current sequence-tagging approach to Grammatical Error Correction (GEC) by introducing specialised tags for spelling correction and morphological inflection using the SymSpell and LemmInflect algorithms.
1 code implementation • 28 Oct 2022 • Christopher Davis, Christopher Bryant, Andrew Caines, Marek Rei, Paula Buttery
Targeted studies testing knowledge of subject-verb agreement (SVA) indicate that pre-trained language models encode syntactic information.
1 code implementation • 23 May 2022 • Joe Stacey, Pasquale Minervini, Haim Dubossarsky, Marek Rei
We can further improve model performance and span-level decisions by using the e-SNLI explanations during training.
2 code implementations • 15 Oct 2021 • Jordan Clive, Kris Cao, Marek Rei
Prefix-tuning is a powerful lightweight technique for adapting a large pre-trained language model to a downstream application.
Ranked #1 on Data-to-Text Generation on WebNLG
no code implementations • NAACL 2022 • Nihir Vedd, Zixu Wang, Marek Rei, Yishu Miao, Lucia Specia
In traditional Visual Question Generation (VQG), most images have multiple concepts (e. g. objects and categories) for which a question could be generated, but models are trained to mimic an arbitrary choice of concept as given in their training data.
1 code implementation • 7 Oct 2021 • Dan Hirlea, Christopher Bryant, Maurizio Zollo, Marek Rei
We introduce the novel task of detecting sustainability initiatives in company reports.
1 code implementation • 16 Apr 2021 • Joe Stacey, Yonatan Belinkov, Marek Rei
Natural Language Inference (NLI) models are known to learn from biases and artefacts within their training data, impacting how well they generalise to other unseen datasets.
1 code implementation • ACL 2022 • Michael Tänzer, Sebastian Ruder, Marek Rei
State-of-the-art pre-trained language models have been shown to memorise facts and perform well with limited amounts of training data.
no code implementations • 8 Apr 2021 • Vinodkumar Prabhakaran, Marek Rei, Ekaterina Shutova
Metaphors are widely used in political rhetoric as an effective framing device.
1 code implementation • ACL (RepL4NLP) 2021 • Kamil Bujel, Helen Yannakoudakis, Marek Rei
We investigate how sentence-level transformers can be modified into effective sequence labelers at the token level without any direct supervision.
1 code implementation • Findings (EMNLP) 2021 • Zhenhao Li, Marek Rei, Lucia Specia
Neural Machine Translation models are sensitive to noise in the input texts, such as misspelled words and ungrammatical constructions.
no code implementations • COLING 2020 • Andrew Caines, Christian Bentz, Kate Knill, Marek Rei, Paula Buttery
We describe the collection of transcription corrections and grammatical error annotations for the CrowdED Corpus of spoken English monologues on business topics.
1 code implementation • COLING 2020 • Miruna Pislar, Marek Rei
In natural languages, words are used in association to construct sentences.
no code implementations • 23 Oct 2020 • Nicole Peinelt, Marek Rei, Maria Liakata
Large pre-trained language models such as BERT have been the driving force behind recent improvements across many NLP tasks.
no code implementations • EMNLP 2020 • Simon Flachs, Ophélie Lacroix, Helen Yannakoudakis, Marek Rei, Anders Søgaard
Evaluation of grammatical error correction (GEC) systems has primarily focused on essays written by non-native learners of English, which however is only part of the full spectrum of GEC applications.
no code implementations • ACL 2020 • Omid Rohanian, Marek Rei, Shiva Taslimipoor, Le An Ha
Metaphor is a linguistic device in which a concept is expressed by mentioning another.
1 code implementation • ACL 2020 • Daniela Gerz, Ivan Vulić, Marek Rei, Roi Reichart, Anna Korhonen
We present a neural framework for learning associations between interrelated groups of words such as the ones found in Subject-Verb-Object (SVO) structures.
no code implementations • 26 Nov 2019 • Bo-Hsiang Tseng, Marek Rei, Paweł Budzianowski, Richard E. Turner, Bill Byrne, Anna Korhonen
Dialogue systems benefit greatly from optimizing on detailed annotations, such as transcribed utterances, internal dialogue state representations and dialogue act labels.
no code implementations • IJCNLP 2019 • Verna Dankers, Marek Rei, Martha Lewis, Ekaterina Shutova
Metaphors allow us to convey emotion by connecting physical experiences and abstract concepts.
no code implementations • IJCNLP 2019 • Bo-Hsiang Tseng, Marek Rei, Pawe{\l} Budzianowski, Richard Turner, Bill Byrne, Anna Korhonen
Dialogue systems benefit greatly from optimizing on detailed annotations, such as transcribed utterances, internal dialogue state representations and dialogue act labels.
no code implementations • WS 2019 • Jeroen Van Hautte, Guy Emerson, Marek Rei
Word embeddings are an essential component in a wide range of natural language processing applications.
no code implementations • WS 2019 • Zheng Yuan, Felix Stahlberg, Marek Rei, Bill Byrne, Helen Yannakoudakis
In this paper, we describe our submission to the BEA 2019 shared task on grammatical error correction.
1 code implementation • WS 2019 • Samuel Bell, Helen Yannakoudakis, Marek Rei
Grammatical error detection (GED) in non-native writing requires systems to identify a wide range of errors in text written by language learners.
no code implementations • SEMEVAL 2019 • Guy Aglionby, Chris Davis, Pushkar Mishra, Andrew Caines, Helen Yannakoudakis, Marek Rei, Ekaterina Shutova, Paula Buttery
We describe the CAMsterdam team entry to the SemEval-2019 Shared Task 6 on offensive language identification in Twitter data.
no code implementations • NAACL 2019 • Simon Flachs, Oph{\'e}lie Lacroix, Marek Rei, Helen Yannakoudakis, Anders S{\o}gaard
While rule-based detection of subject-verb agreement (SVA) errors is sensitive to syntactic parsing errors and irregularities and exceptions to the main rules, neural sequential labelers have a tendency to overfit their training data.
no code implementations • 30 Nov 2018 • Marek Rei, Joshua Oppenheimer, Marek Sirendi
We describe a novel neural network architecture for the prediction of ventricular tachyarrhythmias.
2 code implementations • 14 Nov 2018 • Marek Rei, Anders Søgaard
Learning to construct text representations in end-to-end systems can be difficult, as natural languages are highly compositional and task-specific annotated datasets are often limited in size.
Ranked #1 on Grammatical Error Detection on JFLEG
1 code implementation • CONLL 2018 • Maria Barrett, Joachim Bingel, Nora Hollenstein, Marek Rei, Anders S{\o}gaard
Learning attention functions requires large volumes of data, but many NLP tasks simulate human behavior, and in this paper, we show that human attention really does provide a good inductive bias on many attention functions in NLP.
no code implementations • NAACL 2018 • Yiannos Stathopoulos, Simon Baker, Marek Rei, Simone Teufel
Our results show that the best performing MIR models make use of our typed index, compared to a formula index only containing raw symbols, thereby demonstrating the usefulness of variable typing.
1 code implementation • ACL 2018 • Marek Rei, Daniela Gerz, Ivan Vulić
Experiments show excellent performance on scoring graded lexical entailment, raising the state-of-the-art on the HyperLex dataset by approximately 25%.
no code implementations • NAACL 2018 • Marek Rei, Anders Søgaard
Can attention- or gradient-based visualization techniques be used to infer token-level labels for binary sequence tagging problems, using networks trained only on sentence-level labels?
no code implementations • 21 Jan 2018 • Ronan Cummins, Marek Rei
Grammatical error detection and automated essay scoring are two tasks in the area of automated assessment.
no code implementations • EMNLP 2017 • Marek Rei, Luana Bulat, Douwe Kiela, Ekaterina Shutova
The ubiquity of metaphor in our everyday communication makes it an important problem for natural language understanding.
no code implementations • EMNLP 2017 • Helen Yannakoudakis, Marek Rei, {\O}istein E. Andersen, Zheng Yuan
We propose an approach to N-best list reranking using neural sequence-labelling models.
no code implementations • WS 2017 • Youmna Farag, Marek Rei, Ted Briscoe
Additionally, extending the model with corrections provides further performance gains when data sparsity is an issue.
no code implementations • WS 2017 • Marek Rei, Helen Yannakoudakis
We investigate the utility of different auxiliary objectives and training strategies within a neural sequence labeling approach to error detection in learner writing.
Ranked #2 on Grammatical Error Detection on CoNLL-2014 A1
no code implementations • WS 2017 • Marek Rei
Automated methods for essay scoring have made great progress in recent years, achieving accuracies very close to human annotators.
no code implementations • WS 2017 • Marek Rei, Mariano Felice, Zheng Yuan, Ted Briscoe
Shortage of available training data is holding back progress in the area of automated error detection.
Ranked #3 on Grammatical Error Detection on FCE
3 code implementations • ACL 2017 • Marek Rei
We propose a sequence labeling framework with a secondary training objective, learning to predict surrounding words for every word in the dataset.
Ranked #4 on Grammatical Error Detection on FCE
no code implementations • COLING 2016 • Marek Rei, Gamal K. O. Crichton, Sampo Pyysalo
Sequence labeling architectures use word embeddings for capturing similarity, but suffer when handling previously unseen or rare words.
Ranked #7 on Grammatical Error Detection on FCE
no code implementations • ACL 2016 • Marek Rei, Helen Yannakoudakis
In this paper, we present the first experiments using neural network models for the task of error detection in learner writing.
Ranked #3 on Grammatical Error Detection on CoNLL-2014 A1
3 code implementations • ACL 2016 • Dimitrios Alikaniotis, Helen Yannakoudakis, Marek Rei
Automated Text Scoring (ATS) provides a cost-effective and consistent alternative to human marking.
no code implementations • WS 2016 • Marek Rei, Ronan Cummins
We investigate the task of assessing sentence-level prompt relevance in learner essays.
no code implementations • WS 2016 • Kris Cao, Marek Rei
This paper presents a joint model for performing unsupervised morphological analysis on words, and learning a character-level composition function from morphemes to word embeddings.
no code implementations • EMNLP 2015 • Marek Rei
We investigate an extension of continuous online learning in recurrent neural network language models.