Search Results for author: Albert Gatt

Found 51 papers, 17 papers with code

Incorporating an Error Corpus into a Spellchecker for Maltese

no code implementations LREC 2012 Michael Rosner, Albert Gatt, Andrew Attard, Jan Joachimsen

This paper discusses the ongoing development of a new Maltese spell checker, highlighting the methodologies which would best suit such a language.

A Repository of Data and Evaluation Resources for Natural Language Generation

no code implementations LREC 2012 Anja Belz, Albert Gatt

Starting in 2007, the field of natural language generation (NLG) has organised shared-task evaluation events every year, under the Generation Challenges umbrella.

Data-to-Text Generation Machine Translation

Crowd-sourcing evaluation of automatically acquired, morphologically related word groupings

no code implementations LREC 2014 Claudia Borg, Albert Gatt

The automatic discovery and clustering of morphologically related words is an important problem with several practical applications.

Clustering

Morphological Analysis for the Maltese Language: The Challenges of a Hybrid System

no code implementations WS 2017 Claudia Borg, Albert Gatt

In particular, we analyse a dataset of morphologically related word clusters to evaluate the difference in results for concatenative and nonconcatenative clusters.

Clustering Morphological Analysis

Where to put the Image in an Image Caption Generator

12 code implementations27 Mar 2017 Marc Tanti, Albert Gatt, Kenneth P. Camilleri

When a recurrent neural network language model is used for caption generation, the image information can be fed to the neural network either by directly incorporating it in the RNN -- conditioning the language model by `injecting' image features -- or in a layer following the RNN -- conditioning the language model by `merging' image features.

Caption Generation Language Modelling

Survey of the State of the Art in Natural Language Generation: Core tasks, applications and evaluation

no code implementations29 Mar 2017 Albert Gatt, Emiel Krahmer

This paper surveys the current state of the art in Natural Language Generation (NLG), defined as the task of generating text or speech from non-linguistic input.

nlg evaluation Text Generation

An Empirical Approach for Modeling Fuzzy Geographical Descriptors

no code implementations30 Mar 2017 Alejandro Ramos-Soto, Jose M. Alonso, Ehud Reiter, Kees Van Deemter, Albert Gatt

We present a novel heuristic approach that defines fuzzy geographical descriptors using data gathered from a survey with human subjects.

Referring Expression Referring expression generation +1

What is the Role of Recurrent Neural Networks (RNNs) in an Image Caption Generator?

4 code implementations WS 2017 Marc Tanti, Albert Gatt, Kenneth P. Camilleri

This view suggests that the RNN should only be used to encode linguistic features and that only the final representation should be `merged' with the image features at a later stage.

Image Captioning

Face2Text: Collecting an Annotated Image Description Corpus for the Generation of Rich Face Descriptions

1 code implementation LREC 2018 Albert Gatt, Marc Tanti, Adrian Muscat, Patrizia Paggio, Reuben A. Farrugia, Claudia Borg, Kenneth P. Camilleri, Mike Rosner, Lonneke van der Plas

To gain a better understanding of the variation we find in face description and the possible issues that this may raise, we also conducted an annotation study on a subset of the corpus.

Making effective use of healthcare data using data-to-text technology

no code implementations10 Aug 2018 Steffen Pauws, Albert Gatt, Emiel Krahmer, Ehud Reiter

Financial reports are produced to assess healthcare organizations on some key performance indicators to steer their healthcare delivery.

Meteorologists and Students: A resource for language grounding of geographical descriptors

no code implementations WS 2018 Alejandro Ramos-Soto, Ehud Reiter, Kees Van Deemter, Jose M. Alonso, Albert Gatt

We present a data resource which can be useful for research purposes on language grounding tasks in the context of geographical referring expression generation.

Referring Expression Referring expression generation

Specificity measures and reference

no code implementations WS 2018 Albert Gatt, Nicolás Marín, Gustavo Rivas-Gervilla, Daniel Sánchez

More specifically, we study the ability of several measures of referential success to predict the success of a user in choosing the right object, given a referring expression.

Referring Expression Specificity

You Write Like You Eat: Stylistic variation as a predictor of social stratification

no code implementations ACL 2019 Angelo Basile, Albert Gatt, Malvina Nissim

Inspired by Labov's seminal work on stylistic variation as a function of social stratification, we develop and compare neural models that predict a person's presumed socio-economic status, obtained through distant supervision, from their writing style on social media.

Visuallly Grounded Generation of Entailments from Premises

no code implementations21 Sep 2019 Somaye Jafaritazehjani, Albert Gatt, Marc Tanti

Natural Language Inference (NLI) is the task of determining the semantic relationship between a premise and a hypothesis.

Natural Language Inference Sentence

Visually grounded generation of entailments from premises

no code implementations WS 2019 Somayeh Jafaritazehjani, Albert Gatt, Marc Tanti

Natural Language Inference (NLI) is the task of determining the semantic relationship between a premise and a hypothesis.

Natural Language Inference Sentence

On Architectures for Including Visual Information in Neural Language Models for Image Description

1 code implementation9 Nov 2019 Marc Tanti, Albert Gatt, Kenneth P. Camilleri

We also observe that the merge architecture can have its recurrent neural network pre-trained in a text-only language model (transfer learning) rather than be initialised randomly as usual.

Language Modelling Sentence +1

Annotating for Hate Speech: The MaNeCo Corpus and Some Input from Critical Discourse Analysis

no code implementations LREC 2020 Stavros Assimakopoulos, Rebecca Vella Muskat, Lonneke van der Plas, Albert Gatt

In view of this, we suggest a multi-layer annotation scheme, which is pilot-tested against a binary +/- hate speech classification and appears to yield higher inter-annotator agreement.

Unmasking Contextual Stereotypes: Measuring and Mitigating BERT's Gender Bias

1 code implementation27 Oct 2020 Marion Bartl, Malvina Nissim, Albert Gatt

Contextualized word embeddings have been replacing standard embeddings as the representational knowledge source of choice in NLP systems.

counterfactual Word Embeddings

Datasets and Models for Authorship Attribution on Italian Personal Writings

1 code implementation16 Nov 2020 Gaetana Ruggiero, Albert Gatt, Malvina Nissim

Existing research on Authorship Attribution (AA) focuses on texts for which a lot of data is available (e. g novels), mainly in English.

Authorship Attribution Authorship Verification

Seeing past words: Testing the cross-modal capabilities of pretrained V&L models on counting tasks

no code implementations ACL (mmsr, IWCS) 2021 Letitia Parcalabescu, Albert Gatt, Anette Frank, Iacer Calixto

We investigate the reasoning ability of pretrained vision and language (V&L) models in two tasks that require multimodal integration: (1) discriminating a correct image-sentence pair from an incorrect one, and (2) counting entities in an image.

Sentence Task 2

On the interaction of automatic evaluation and task framing in headline style transfer

1 code implementation ACL (EvalNLGEval, INLG) 2020 Lorenzo De Mattei, Michele Cafagna, Huiyuan Lai, Felice Dell'Orletta, Malvina Nissim, Albert Gatt

An ongoing debate in the NLG community concerns the best way to evaluate systems, with human evaluation often being considered the most reliable method, compared to corpus-based metrics.

Style Transfer

On the Language-specificity of Multilingual BERT and the Impact of Fine-tuning

1 code implementation EMNLP (BlackboxNLP) 2021 Marc Tanti, Lonneke van der Plas, Claudia Borg, Albert Gatt

Recent work has shown evidence that the knowledge acquired by multilingual BERT (mBERT) has two components: a language-specific and a language-neutral one.

Language Identification Natural Language Inference +3

What Vision-Language Models `See' when they See Scenes

no code implementations15 Sep 2021 Michele Cafagna, Kees Van Deemter, Albert Gatt

Images can be described in terms of the objects they contain, or in terms of the types of scene or place that they instantiate.

Object

Analysis of Data Augmentation Methods for Low-Resource Maltese ASR

no code implementations15 Nov 2021 Andrea DeMarco, Carlos Mena, Albert Gatt, Claudia Borg, Aiden Williams, Lonneke van der Plas

Recent years have seen an increased interest in the computational speech processing of Maltese, but resources remain sparse.

Data Augmentation Language Modelling +2

VALSE: A Task-Independent Benchmark for Vision and Language Models Centered on Linguistic Phenomena

1 code implementation ACL 2022 Letitia Parcalabescu, Michele Cafagna, Lilitta Muradjan, Anette Frank, Iacer Calixto, Albert Gatt

We propose VALSE (Vision And Language Structured Evaluation), a novel benchmark designed for testing general-purpose pretrained vision and language (V&L) models for their visio-linguistic grounding capabilities on specific linguistic phenomena.

image-sentence alignment valid

Face2Text revisited: Improved data set and baseline results

no code implementations PVLAM (LREC) 2022 Marc Tanti, Shaun Abdilla, Adrian Muscat, Claudia Borg, Reuben A. Farrugia, Albert Gatt

To encourage the development of more human-focused descriptions, we developed a new data set of facial descriptions based on the CelebA image data set.

Transfer Learning

HL Dataset: Visually-grounded Description of Scenes, Actions and Rationales

1 code implementation23 Feb 2023 Michele Cafagna, Kees Van Deemter, Albert Gatt

We present the High-Level Dataset a dataset extending 14997 images from the COCO dataset, aligned with a new set of 134, 973 human-annotated (high-level) captions collected along three axes: scenes, actions, and rationales.

Common Sense Reasoning Vocal Bursts Intensity Prediction

Interpreting Vision and Language Generative Models with Semantic Visual Priors

no code implementations28 Apr 2023 Michele Cafagna, Lina M. Rojas-Barahona, Kees Van Deemter, Albert Gatt

When applied to Image-to-text models, interpretability methods often provide token-by-token explanations namely, they compute a visual explanation for each token of the generated sequence.

Contrast Is All You Need

no code implementations6 Jul 2023 Burak Kilic, Florix Bex, Albert Gatt

We focused on two finetuning objectives; SetFit (Sentence Transformer Finetuning), a contrastive learning setup, and a vanilla finetuning setup on a legal provision classification task.

Classification Contrastive Learning +1

The Scenario Refiner: Grounding subjects in images at the morphological level

no code implementations20 Sep 2023 Claudia Tagliaferri, Sofia Axioti, Albert Gatt, Denis Paperno

Derivationally related words, such as "runner" and "running", exhibit semantic differences which also elicit different visual scenarios.

FTFT: Efficient and Robust Fine-Tuning by Transferring Training Dynamics

1 code implementation10 Oct 2023 Yupei Du, Albert Gatt, Dong Nguyen

Dataset cartography is a simple yet effective dual-model approach that improves the robustness of fine-tuned PLMs.

ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models

no code implementations13 Nov 2023 Ilker Kesen, Andrea Pedrotti, Mustafa Dogan, Michele Cafagna, Emre Can Acikgoz, Letitia Parcalabescu, Iacer Calixto, Anette Frank, Albert Gatt, Aykut Erdem, Erkut Erdem

With the ever-increasing popularity of pretrained Video-Language Models (VidLMs), there is a pressing need to develop robust evaluation methodologies that delve deeper into their visio-linguistic capabilities.

counterfactual Language Modelling

Multi3Generation: Multitask, Multilingual, Multimodal Language Generation

no code implementations EAMT 2022 Anabela Barreiro, José GC de Souza, Albert Gatt, Mehul Bhatt, Elena Lloret, Aykut Erdem, Dimitra Gkatzia, Helena Moniz, Irene Russo, Fabio Kepler, Iacer Calixto, Marcin Paprzycki, François Portet, Isabelle Augenstein, Mirela Alhasani

This paper presents the Multitask, Multilingual, Multimodal Language Generation COST Action – Multi3Generation (CA18231), an interdisciplinary network of research groups working on different aspects of language generation.

Text Generation

Entity-Based Semantic Adequacy for Data-to-Text Generation

no code implementations Findings (EMNLP) 2021 Juliette Faille, Albert Gatt, Claire Gardent

While powerful pre-trained language models have improved the fluency of text generation models, semantic adequacy -the ability to generate text that is semantically faithful to the input- remains an unsolved issue.

Data-to-Text Generation

Gradations of Error Severity in Automatic Image Descriptions

no code implementations INLG (ACL) 2020 Emiel van Miltenburg, Wei-Ting Lu, Emiel Krahmer, Albert Gatt, Guanyi Chen, Lin Li, Kees Van Deemter

Because our manipulated descriptions form minimal pairs with the reference descriptions, we are able to assess the impact of different kinds of errors on the perceived quality of the descriptions.

Towards Harnessing Natural Language Generation to Explain Black-box Models

no code implementations ACL (NL4XAI, INLG) 2020 Ettore Mariotti, Jose M. Alonso, Albert Gatt

The opaque nature of many machine learning techniques prevents the wide adoption of powerful information processing tools for high stakes scenarios.

Decision Making Explainable artificial intelligence +2

The Natural Language Pipeline, Neural Text Generation and Explainability

no code implementations ACL (NL4XAI, INLG) 2020 Juliette Faille, Albert Gatt, Claire Gardent

End-to-end encoder-decoder approaches to data-to-text generation are often black boxes whose predictions are difficult to explain.

Data-to-Text Generation

Unmasking Contextual Stereotypes: Measuring and Mitigating BERT’s Gender Bias

1 code implementation GeBNLP (COLING) 2020 Marion Bartl, Malvina Nissim, Albert Gatt

Contextualized word embeddings have been replacing standard embeddings as the representational knowledge source of choice in NLP systems.

counterfactual Word Embeddings

Cannot find the paper you are looking for? You can Submit a new open access paper.