no code implementations • Findings (EMNLP) 2021 • Juliette Faille, Albert Gatt, Claire Gardent
While powerful pre-trained language models have improved the fluency of text generation models, semantic adequacy -the ability to generate text that is semantically faithful to the input- remains an unsolved issue.
1 code implementation • GeBNLP (COLING) 2020 • Marion Bartl, Malvina Nissim, Albert Gatt
Contextualized word embeddings have been replacing standard embeddings as the representational knowledge source of choice in NLP systems.
no code implementations • INLG (ACL) 2020 • Emiel van Miltenburg, Wei-Ting Lu, Emiel Krahmer, Albert Gatt, Guanyi Chen, Lin Li, Kees Van Deemter
Because our manipulated descriptions form minimal pairs with the reference descriptions, we are able to assess the impact of different kinds of errors on the perceived quality of the descriptions.
no code implementations • ACL (NL4XAI, INLG) 2020 • Ettore Mariotti, Jose M. Alonso, Albert Gatt
The opaque nature of many machine learning techniques prevents the wide adoption of powerful information processing tools for high stakes scenarios.
no code implementations • ACL (NL4XAI, INLG) 2020 • Juliette Faille, Albert Gatt, Claire Gardent
End-to-end encoder-decoder approaches to data-to-text generation are often black boxes whose predictions are difficult to explain.
no code implementations • CMCL (ACL) 2022 • Inga Lang, Lonneke Plas, Malvina Nissim, Albert Gatt
We find that adding visual vectors increases classification performance on our dataset in many cases.
no code implementations • EAMT 2022 • Anabela Barreiro, José GC de Souza, Albert Gatt, Mehul Bhatt, Elena Lloret, Aykut Erdem, Dimitra Gkatzia, Helena Moniz, Irene Russo, Fabio Kepler, Iacer Calixto, Marcin Paprzycki, François Portet, Isabelle Augenstein, Mirela Alhasani
This paper presents the Multitask, Multilingual, Multimodal Language Generation COST Action – Multi3Generation (CA18231), an interdisciplinary network of research groups working on different aspects of language generation.
no code implementations • 2 Sep 2024 • Ivana Beňová, Michal Gregor, Albert Gatt
This study investigates the ability of various vision-language (VL) models to ground context-dependent and non-context-dependent verb phrases.
no code implementations • 19 Aug 2024 • Mika Sie, Ruby Beek, Michiel Bots, Sjaak Brinkkemper, Albert Gatt
In this paper, we show that the effectiveness of a two-step architecture for summarizing long regulatory texts varies significantly depending on the model used.
1 code implementation • 17 Aug 2024 • Patrícia Schmidtová, Saad Mahamood, Simone Balloccu, Ondřej Dušek, Albert Gatt, Dimitra Gkatzia, David M. Howcroft, Ondřej Plátek, Adarsa Sivaprasad
Automatic metrics are extensively used to evaluate natural language processing systems.
no code implementations • 12 Aug 2024 • Yingjin Song, Denis Paperno, Albert Gatt
Visual storytelling systems generate multi-sentence stories from image sequences.
no code implementations • 15 Jul 2024 • Vincent Quantmeyer, Pablo Mosteiro, Albert Gatt
Here we build on the existence task from the VALSE benchmark (Parcalabescu et al, 2022) which we use to test models' understanding of negation, a particularly interesting issue for multimodal models.
1 code implementation • 26 Jun 2024 • Anna Bavaresco, Raffaella Bernardi, Leonardo Bertolazzi, Desmond Elliott, Raquel Fernández, Albert Gatt, Esam Ghaleb, Mario Giulianelli, Michael Hanna, Alexander Koller, André F. T. Martins, Philipp Mondorf, Vera Neplenbroek, Sandro Pezzelle, Barbara Plank, David Schlangen, Alessandro Suglia, Aditya K Surikuchi, Ece Takmaz, Alberto Testoni
There is an increasing trend towards evaluating NLP models with LLM-generated judgments instead of human judgments.
no code implementations • 17 Jun 2024 • Leonardo Bertolazzi, Albert Gatt, Raffaella Bernardi
The reasoning abilities of Large Language Models (LLMs) are becoming a central focus of study in NLP.
no code implementations • 13 Nov 2023 • Ilker Kesen, Andrea Pedrotti, Mustafa Dogan, Michele Cafagna, Emre Can Acikgoz, Letitia Parcalabescu, Iacer Calixto, Anette Frank, Albert Gatt, Aykut Erdem, Erkut Erdem
With the ever-increasing popularity of pretrained Video-Language Models (VidLMs), there is a pressing need to develop robust evaluation methodologies that delve deeper into their visio-linguistic capabilities.
1 code implementation • 10 Oct 2023 • Yupei Du, Albert Gatt, Dong Nguyen
Dataset cartography is a simple yet effective dual-model approach that improves the robustness of fine-tuned PLMs.
no code implementations • 20 Sep 2023 • Claudia Tagliaferri, Sofia Axioti, Albert Gatt, Denis Paperno
Derivationally related words, such as "runner" and "running", exhibit semantic differences which also elicit different visual scenarios.
no code implementations • 6 Jul 2023 • Burak Kilic, Florix Bex, Albert Gatt
We focused on two finetuning objectives; SetFit (Sentence Transformer Finetuning), a contrastive learning setup, and a vanilla finetuning setup on a legal provision classification task.
no code implementations • 2 May 2023 • Anya Belz, Craig Thomson, Ehud Reiter, Gavin Abercrombie, Jose M. Alonso-Moral, Mohammad Arvan, Anouck Braggaar, Mark Cieliebak, Elizabeth Clark, Kees Van Deemter, Tanvi Dinkar, Ondřej Dušek, Steffen Eger, Qixiang Fang, Mingqi Gao, Albert Gatt, Dimitra Gkatzia, Javier González-Corbelle, Dirk Hovy, Manuela Hürlimann, Takumi Ito, John D. Kelleher, Filip Klubicka, Emiel Krahmer, Huiyuan Lai, Chris van der Lee, Yiru Li, Saad Mahamood, Margot Mieskes, Emiel van Miltenburg, Pablo Mosteiro, Malvina Nissim, Natalie Parde, Ondřej Plátek, Verena Rieser, Jie Ruan, Joel Tetreault, Antonio Toral, Xiaojun Wan, Leo Wanner, Lewis Watson, Diyi Yang
We report our efforts in identifying a set of previous human evaluations in NLP that would be suitable for a coordinated study examining what makes human evaluations in NLP more/less reproducible.
no code implementations • 28 Apr 2023 • Michele Cafagna, Lina M. Rojas-Barahona, Kees Van Deemter, Albert Gatt
When applied to Image-to-text models, interpretability methods often provide token-by-token explanations namely, they compute a visual explanation for each token of the generated sequence.
1 code implementation • 23 Feb 2023 • Michele Cafagna, Kees Van Deemter, Albert Gatt
We present the High-Level Dataset a dataset extending 14997 images from the COCO dataset, aligned with a new set of 134, 973 human-annotated (high-level) captions collected along three axes: scenes, actions, and rationales.
no code implementations • 9 Nov 2022 • Michele Cafagna, Kees Van Deemter, Albert Gatt
Image captioning models tend to describe images in an object-centric way, emphasising visible objects.
no code implementations • PVLAM (LREC) 2022 • Marc Tanti, Shaun Abdilla, Adrian Muscat, Claudia Borg, Reuben A. Farrugia, Albert Gatt
To encourage the development of more human-focused descriptions, we developed a new data set of facial descriptions based on the CelebA image data set.
1 code implementation • DeepLo 2022 • Kurt Micallef, Albert Gatt, Marc Tanti, Lonneke van der Plas, Claudia Borg
We also present a newly created corpus for Maltese, and determine the effect that the pre-training data size and domain have on the downstream performance.
1 code implementation • ACL 2022 • Letitia Parcalabescu, Michele Cafagna, Lilitta Muradjan, Anette Frank, Iacer Calixto, Albert Gatt
We propose VALSE (Vision And Language Structured Evaluation), a novel benchmark designed for testing general-purpose pretrained vision and language (V&L) models for their visio-linguistic grounding capabilities on specific linguistic phenomena.
no code implementations • 15 Nov 2021 • Andrea DeMarco, Carlos Mena, Albert Gatt, Claudia Borg, Aiden Williams, Lonneke van der Plas
Recent years have seen an increased interest in the computational speech processing of Maltese, but resources remain sparse.
no code implementations • 15 Sep 2021 • Michele Cafagna, Kees Van Deemter, Albert Gatt
Images can be described in terms of the objects they contain, or in terms of the types of scene or place that they instantiate.
1 code implementation • EMNLP (BlackboxNLP) 2021 • Marc Tanti, Lonneke van der Plas, Claudia Borg, Albert Gatt
Recent work has shown evidence that the knowledge acquired by multilingual BERT (mBERT) has two components: a language-specific and a language-neutral one.
1 code implementation • ACL (EvalNLGEval, INLG) 2020 • Lorenzo De Mattei, Michele Cafagna, Huiyuan Lai, Felice Dell'Orletta, Malvina Nissim, Albert Gatt
An ongoing debate in the NLG community concerns the best way to evaluate systems, with human evaluation often being considered the most reliable method, compared to corpus-based metrics.
no code implementations • ACL (mmsr, IWCS) 2021 • Letitia Parcalabescu, Albert Gatt, Anette Frank, Iacer Calixto
We investigate the reasoning ability of pretrained vision and language (V&L) models in two tasks that require multimodal integration: (1) discriminating a correct image-sentence pair from an incorrect one, and (2) counting entities in an image.
1 code implementation • 16 Nov 2020 • Gaetana Ruggiero, Albert Gatt, Malvina Nissim
Existing research on Authorship Attribution (AA) focuses on texts for which a lot of data is available (e. g novels), mainly in English.
1 code implementation • 27 Oct 2020 • Marion Bartl, Malvina Nissim, Albert Gatt
Contextualized word embeddings have been replacing standard embeddings as the representational knowledge source of choice in NLP systems.
no code implementations • LREC 2020 • Stavros Assimakopoulos, Rebecca Vella Muskat, Lonneke van der Plas, Albert Gatt
In view of this, we suggest a multi-layer annotation scheme, which is pilot-tested against a binary +/- hate speech classification and appears to yield higher inter-annotator agreement.
no code implementations • LREC 2020 • Carlos Mena, Albert Gatt, Andrea DeMarco, Claudia Borg, Lonneke van der Plas, Amanda Muscat, Ian Padovani
Maltese, the national language of Malta, is spoken by approximately 500, 000 people.
Automatic Speech Recognition Automatic Speech Recognition (ASR) +1
1 code implementation • 9 Nov 2019 • Marc Tanti, Albert Gatt, Kenneth P. Camilleri
We also observe that the merge architecture can have its recurrent neural network pre-trained in a text-only language model (transfer learning) rather than be initialised randomly as usual.
no code implementations • WS 2019 • Chris van der Lee, Albert Gatt, Emiel van Miltenburg, S Wubben, er, Emiel Krahmer
Currently, there is little agreement as to how Natural Language Generation (NLG) systems should be evaluated.
no code implementations • WS 2019 • Somayeh Jafaritazehjani, Albert Gatt, Marc Tanti
Natural Language Inference (NLI) is the task of determining the semantic relationship between a premise and a hypothesis.
no code implementations • 21 Sep 2019 • Somaye Jafaritazehjani, Albert Gatt, Marc Tanti
Natural Language Inference (NLI) is the task of determining the semantic relationship between a premise and a hypothesis.
no code implementations • ACL 2019 • Angelo Basile, Albert Gatt, Malvina Nissim
Inspired by Labov's seminal work on stylistic variation as a function of social stratification, we develop and compare neural models that predict a person's presumed socio-economic status, obtained through distant supervision, from their writing style on social media.
1 code implementation • 1 Jan 2019 • Marc Tanti, Albert Gatt, Kenneth P. Camilleri
When designing a neural caption generator, a convolutional neural network can be used to extract image features.
1 code implementation • 12 Oct 2018 • Marc Tanti, Albert Gatt, Adrian Muscat
Image caption generation systems are typically evaluated against reference outputs.
1 code implementation • 12 Oct 2018 • Marc Tanti, Albert Gatt, Kenneth P. Camilleri
This paper addresses the sensitivity of neural image caption generators to their visual input.
no code implementations • WS 2018 • Albert Gatt, Nicolás Marín, Gustavo Rivas-Gervilla, Daniel Sánchez
More specifically, we study the ability of several measures of referential success to predict the success of a user in choosing the right object, given a referring expression.
no code implementations • WS 2018 • Alejandro Ramos-Soto, Ehud Reiter, Kees Van Deemter, Jose M. Alonso, Albert Gatt
We present a data resource which can be useful for research purposes on language grounding tasks in the context of geographical referring expression generation.
no code implementations • 10 Aug 2018 • Steffen Pauws, Albert Gatt, Emiel Krahmer, Ehud Reiter
Financial reports are produced to assess healthcare organizations on some key performance indicators to steer their healthcare delivery.
1 code implementation • COLING 2018 • Hoa Trong Vu, Claudio Greco, Aliia Erofeeva, Somayeh Jafaritazehjan, Guido Linders, Marc Tanti, Alberto Testoni, Raffaella Bernardi, Albert Gatt
Capturing semantic relations between sentences, such as entailment, is a long-standing challenge for computational semantics.
Ranked #2 on Natural Language Inference on V-SNLI
1 code implementation • LREC 2018 • Albert Gatt, Marc Tanti, Adrian Muscat, Patrizia Paggio, Reuben A. Farrugia, Claudia Borg, Kenneth P. Camilleri, Mike Rosner, Lonneke van der Plas
To gain a better understanding of the variation we find in face description and the possible issues that this may raise, we also conducted an annotation study on a subset of the corpus.
no code implementations • WS 2017 • Hoa Trong Vu, Thuong-Hai Pham, Xiaoyu Bai, Marc Tanti, Lonneke van der Plas, Albert Gatt
System using BiLSTM and max pooling.
4 code implementations • WS 2017 • Marc Tanti, Albert Gatt, Kenneth P. Camilleri
This view suggests that the RNN should only be used to encode linguistic features and that only the final representation should be `merged' with the image features at a later stage.
no code implementations • 30 Mar 2017 • Alejandro Ramos-Soto, Jose M. Alonso, Ehud Reiter, Kees Van Deemter, Albert Gatt
We present a novel heuristic approach that defines fuzzy geographical descriptors using data gathered from a survey with human subjects.
no code implementations • 29 Mar 2017 • Albert Gatt, Emiel Krahmer
This paper surveys the current state of the art in Natural Language Generation (NLG), defined as the task of generating text or speech from non-linguistic input.
12 code implementations • 27 Mar 2017 • Marc Tanti, Albert Gatt, Kenneth P. Camilleri
When a recurrent neural network language model is used for caption generation, the image information can be fed to the neural network either by directly incorporating it in the RNN -- conditioning the language model by `injecting' image features -- or in a layer following the RNN -- conditioning the language model by `merging' image features.
no code implementations • WS 2017 • Claudia Borg, Albert Gatt
In particular, we analyse a dataset of morphologically related word clusters to evaluate the difference in results for concatenative and nonconcatenative clusters.
no code implementations • LREC 2014 • Claudia Borg, Albert Gatt
The automatic discovery and clustering of morphologically related words is an important problem with several practical applications.
no code implementations • LREC 2012 • Michael Rosner, Albert Gatt, Andrew Attard, Jan Joachimsen
This paper discusses the ongoing development of a new Maltese spell checker, highlighting the methodologies which would best suit such a language.
no code implementations • LREC 2012 • Anja Belz, Albert Gatt
Starting in 2007, the field of natural language generation (NLG) has organised shared-task evaluation events every year, under the Generation Challenges umbrella.