Search Results for author: Pietro Astolfi

Found 16 papers, 7 papers with code

Object-centric Binding in Contrastive Language-Image Pretraining

no code implementations19 Feb 2025 Rim Assouel, Pietro Astolfi, Florian Bordes, Michal Drozdzal, Adriana Romero-Soriano

However, these models have limitations in understanding complex compositional scenes involving multiple objects and their spatial relationships.

Image-text matching Object +1

Boosting Latent Diffusion with Perceptual Objectives

no code implementations6 Nov 2024 Tariq Berrada, Pietro Astolfi, Melissa Hall, Marton Havasi, Yohann Benchetrit, Adriana Romero-Soriano, Karteek Alahari, Michal Drozdzal, Jakob Verbeek

LDMs learn the data distribution in the latent space of an autoencoder (AE) and produce images by mapping the generated latents into RGB image space using the AE decoder.

Decoder

Consistency-diversity-realism Pareto fronts of conditional image generative models

1 code implementation14 Jun 2024 Pietro Astolfi, Marlene Careil, Melissa Hall, Oscar Mañas, Matthew Muckley, Jakob Verbeek, Adriana Romero Soriano, Michal Drozdzal

Building world models that accurately and comprehensively represent the real world is the utmost aspiration for conditional image generative models as it would enable their use as world simulators.

Diversity

Improving Text-to-Image Consistency via Automatic Prompt Optimization

no code implementations26 Mar 2024 Oscar Mañas, Pietro Astolfi, Melissa Hall, Candace Ross, Jack Urbanek, Adina Williams, Aishwarya Agrawal, Adriana Romero-Soriano, Michal Drozdzal

In this paper, we address these challenges and introduce a T2I optimization-by-prompting framework, OPT2I, which leverages a large language model (LLM) to improve prompt-image consistency in T2I models.

Language Modelling Large Language Model

DP-RDM: Adapting Diffusion Models to Private Domains Without Fine-Tuning

1 code implementation21 Mar 2024 Jonathan Lebensold, Maziar Sanjabi, Pietro Astolfi, Adriana Romero-Soriano, Kamalika Chaudhuri, Mike Rabbat, Chuan Guo

Text-to-image diffusion models have been shown to suffer from sample-level memorization, possibly reproducing near-perfect replica of images that they are trained on, which may be undesirable.

Memorization Retrieval

Improved baselines for vision-language pre-training

1 code implementation15 May 2023 Enrico Fini, Pietro Astolfi, Adriana Romero-Soriano, Jakob Verbeek, Michal Drozdzal

Indeed, we find that a simple CLIP baseline can also be improved substantially, up to a 25% relative improvement on downstream zero-shot tasks, by using well-known training techniques that are popular in other subfields.

Contrastive Learning Data Augmentation +1

Instance-Conditioned GAN Data Augmentation for Representation Learning

no code implementations16 Mar 2023 Pietro Astolfi, Arantxa Casanova, Jakob Verbeek, Pascal Vincent, Adriana Romero-Soriano, Michal Drozdzal

We showcase the benefits of DA_IC-GAN by plugging it out-of-the-box into the supervised training of ResNets and DeiT models on the ImageNet dataset, and achieving accuracy boosts up to between 1%p and 2%p with the highest capacity models.

Data Augmentation Few-Shot Learning +1

Tractogram filtering of anatomically non-plausible fibers with geometric deep learning

no code implementations24 Mar 2020 Pietro Astolfi, Ruben Verhagen, Laurent Petit, Emanuele Olivetti, Jonathan Masci, Davide Boscaini, Paolo Avesani

The intuitive idea is to model a fiber as a point cloud and the goal is to investigate whether and how a geometric deep learning model might capture its anatomical properties.

Anatomy Deep Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.