Search Results for author: Michele Cafagna

Found 10 papers, 4 papers with code

ViLMA: A Zero-Shot Benchmark for Linguistic and Temporal Grounding in Video-Language Models

no code implementations13 Nov 2023 Ilker Kesen, Andrea Pedrotti, Mustafa Dogan, Michele Cafagna, Emre Can Acikgoz, Letitia Parcalabescu, Iacer Calixto, Anette Frank, Albert Gatt, Aykut Erdem, Erkut Erdem

With the ever-increasing popularity of pretrained Video-Language Models (VidLMs), there is a pressing need to develop robust evaluation methodologies that delve deeper into their visio-linguistic capabilities.

counterfactual Language Modelling

Interpreting Vision and Language Generative Models with Semantic Visual Priors

no code implementations28 Apr 2023 Michele Cafagna, Lina M. Rojas-Barahona, Kees Van Deemter, Albert Gatt

When applied to Image-to-text models, interpretability methods often provide token-by-token explanations namely, they compute a visual explanation for each token of the generated sequence.

HL Dataset: Visually-grounded Description of Scenes, Actions and Rationales

1 code implementation23 Feb 2023 Michele Cafagna, Kees Van Deemter, Albert Gatt

We present the High-Level Dataset a dataset extending 14997 images from the COCO dataset, aligned with a new set of 134, 973 human-annotated (high-level) captions collected along three axes: scenes, actions, and rationales.

Common Sense Reasoning Vocal Bursts Intensity Prediction

VALSE: A Task-Independent Benchmark for Vision and Language Models Centered on Linguistic Phenomena

1 code implementation ACL 2022 Letitia Parcalabescu, Michele Cafagna, Lilitta Muradjan, Anette Frank, Iacer Calixto, Albert Gatt

We propose VALSE (Vision And Language Structured Evaluation), a novel benchmark designed for testing general-purpose pretrained vision and language (V&L) models for their visio-linguistic grounding capabilities on specific linguistic phenomena.

image-sentence alignment valid

What Vision-Language Models `See' when they See Scenes

no code implementations15 Sep 2021 Michele Cafagna, Kees Van Deemter, Albert Gatt

Images can be described in terms of the objects they contain, or in terms of the types of scene or place that they instantiate.

Object

On the interaction of automatic evaluation and task framing in headline style transfer

1 code implementation ACL (EvalNLGEval, INLG) 2020 Lorenzo De Mattei, Michele Cafagna, Huiyuan Lai, Felice Dell'Orletta, Malvina Nissim, Albert Gatt

An ongoing debate in the NLG community concerns the best way to evaluate systems, with human evaluation often being considered the most reliable method, compared to corpus-based metrics.

Style Transfer

Norm It! Lexical Normalization for Italian and Its Downstream Effects for Dependency Parsing

no code implementations LREC 2020 Rob van der Goot, Alan Ramponi, Tommaso Caselli, Michele Cafagna, Lorenzo De Mattei

However, for Italian, there is no benchmark available for lexical normalization, despite the presence of many benchmarks for other tasks involving social media data.

Dependency Parsing Lexical Normalization

GePpeTto Carves Italian into a Language Model

1 code implementation29 Apr 2020 Lorenzo De Mattei, Michele Cafagna, Felice Dell'Orletta, Malvina Nissim, Marco Guerini

We provide a thorough analysis of GePpeTto's quality by means of both an automatic and a human-based evaluation.

Language Modelling Sentence +1

Cannot find the paper you are looking for? You can Submit a new open access paper.