Search Results for author: Pietro Barbiero

Found 37 papers, 25 papers with code

Logic Explanation of AI Classifiers by Categorical Explaining Functors

no code implementations20 Mar 2025 Stefano Fioravanti, Francesco Giannini, Paolo Frazzetto, Fabio Zanasi, Pietro Barbiero

The most common methods in explainable artificial intelligence are post-hoc techniques which identify the most relevant features used by pretrained opaque models.

Explainable artificial intelligence

Neural Interpretable Reasoning

no code implementations17 Feb 2025 Pietro Barbiero, Giuseppe Marra, Gabriele Ciravegna, David Debot, Francesco De Santis, Michelangelo Diligenti, Mateo Espinosa Zarlenga, Francesco Giannini

We formalize a novel modeling framework for achieving interpretability in deep learning, anchored in the principle of inference equivariance.

A Survey on Federated Learning in Human Sensing

no code implementations7 Jan 2025 Mohan Li, Martin Gjoreski, Pietro Barbiero, Gašper Slapničar, Mitja Luštrek, Nicholas D. Lane, Marc Langheinrich

However, its reliance on detailed and often privacy-sensitive data as the basis for its machine learning (ML) models raises significant legal and ethical concerns.

Federated Learning Survey

Counterfactual Explanations for Clustering Models

no code implementations19 Sep 2024 Aurora Spagnol, Kacper Sokol, Pietro Barbiero, Marc Langheinrich, Martin Gjoreski

While many explainable artificial intelligence techniques exist for supervised machine learning, unsupervised learning -- and clustering in particular -- has been largely neglected.

Clustering counterfactual +1

Interpretable Concept-Based Memory Reasoning

1 code implementation22 Jul 2024 David Debot, Pietro Barbiero, Francesco Giannini, Gabriele Ciravegna, Michelangelo Diligenti, Giuseppe Marra

The presence of an explicit memory and the symbolic evaluation allow domain experts to inspect and formally verify the validity of certain global properties of interest for the task prediction process.

Decision Making

Self-supervised Interpretable Concept-based Models for Text Classification

no code implementations20 Jun 2024 Francesco De Santis, Philippe Bich, Gabriele Ciravegna, Pietro Barbiero, Danilo Giordano, Tania Cerquitelli

Additionally, we show that our models are (i) interpretable, offering meaningful logical explanations for their predictions; (ii) interactable, allowing humans to modify intermediate predictions through concept interventions; and (iii) controllable, guiding the LLMs' decoding process to follow a required decision-making path.

Decision Making text-classification +1

Causal Concept Graph Models: Beyond Causal Opacity in Deep Learning

1 code implementation26 May 2024 Gabriele Dominici, Pietro Barbiero, Mateo Espinosa Zarlenga, Alberto Termine, Martin Gjoreski, Giuseppe Marra, Marc Langheinrich

Causal opacity denotes the difficulty in understanding the "hidden" causal structure underlying the decisions of deep neural network (DNN) models.

counterfactual Decision Making +2

AnyCBMs: How to Turn Any Black Box into a Concept Bottleneck Model

no code implementations26 May 2024 Gabriele Dominici, Pietro Barbiero, Francesco Giannini, Martin Gjoreski, Marc Langhenirich

Interpretable deep learning aims at developing neural architectures whose decision-making processes could be understood by their users.

Decision Making

Federated Behavioural Planes: Explaining the Evolution of Client Behaviour in Federated Learning

1 code implementation24 May 2024 Dario Fenoglio, Gabriele Dominici, Pietro Barbiero, Alberto Tonda, Martin Gjoreski, Marc Langheinrich

Federated Learning (FL), a privacy-aware approach in distributed deep learning environments, enables many clients to collaboratively train a model without sharing sensitive data, thereby reducing privacy risks.

counterfactual Decision Making +1

Counterfactual Concept Bottleneck Models

1 code implementation2 Feb 2024 Gabriele Dominici, Pietro Barbiero, Francesco Giannini, Martin Gjoreski, Giuseppe Marra, Marc Langheinrich

Current deep learning models are not designed to simultaneously address three fundamental questions: predict class labels to solve a given classification task (the "What?

counterfactual Decision Making

Digital Histopathology with Graph Neural Networks: Concepts and Explanations for Clinicians

no code implementations4 Dec 2023 Alessandro Farace di Villaforesta, Lucie Charlotte Magister, Pietro Barbiero, Pietro Liò

To address the challenge of the ``black-box" nature of deep learning in medical settings, we combine GCExplainer - an automated concept discovery solution - along with Logic Explained Networks to provide global explanations for Graph Neural Networks.

graph construction Panoptic Segmentation

Everybody Needs a Little HELP: Explaining Graphs via Hierarchical Concepts

1 code implementation25 Nov 2023 Jonas Jürß, Lucie Charlotte Magister, Pietro Barbiero, Pietro Liò, Nikola Simidjievski

A line of interpretable methods approach this by discovering a small set of relevant concepts as subgraphs in the last GNN layer that together explain the prediction.

Drug Discovery Travel Time Estimation

From Charts to Atlas: Merging Latent Spaces into One

1 code implementation11 Nov 2023 Donato Crisostomi, Irene Cannistraci, Luca Moschella, Pietro Barbiero, Marco Ciccone, Pietro Liò, Emanuele Rodolà

Models trained on semantically related datasets and tasks exhibit comparable inter-sample relations within their latent spaces.

Relational Concept Bottleneck Models

1 code implementation23 Aug 2023 Pietro Barbiero, Francesco Giannini, Gabriele Ciravegna, Michelangelo Diligenti, Giuseppe Marra

The design of interpretable deep learning models working in relational domains poses an open challenge: interpretable deep learning methods, such as Concept Bottleneck Models (CBMs), are not designed to solve relational problems, while relational deep learning models, such as Graph Neural Networks (GNNs), are not as interpretable as CBMs.

Deep Learning Image Classification +2

SHARCS: Shared Concept Space for Explainable Multimodal Learning

1 code implementation1 Jul 2023 Gabriele Dominici, Pietro Barbiero, Lucie Charlotte Magister, Pietro Liò, Nikola Simidjievski

Multimodal learning is an essential paradigm for addressing complex real-world problems, where individual data modalities are typically insufficient to accurately solve a given modelling task.

Retrieval

GCI: A (G)raph (C)oncept (I)nterpretation Framework

1 code implementation9 Feb 2023 Dmitry Kazhdan, Botty Dimanov, Lucie Charlotte Magister, Pietro Barbiero, Mateja Jamnik, Pietro Lio

Explainable AI (XAI) underwent a recent surge in research on concept extraction, focusing on extracting human-interpretable concepts from Deep Neural Networks.

Explainable Artificial Intelligence (XAI) Molecular Property Prediction +2

Towards Robust Metrics for Concept Representation Evaluation

1 code implementation25 Jan 2023 Mateo Espinosa Zarlenga, Pietro Barbiero, Zohreh Shams, Dmitry Kazhdan, Umang Bhatt, Adrian Weller, Mateja Jamnik

In this paper, we show that such metrics are not appropriate for concept learning and propose novel metrics for evaluating the purity of concept representations in both approaches.

Benchmarking Disentanglement

Extending Logic Explained Networks to Text Classification

2 code implementations4 Nov 2022 Rishabh Jain, Gabriele Ciravegna, Pietro Barbiero, Francesco Giannini, Davide Buffelli, Pietro Lio

Recently, Logic Explained Networks (LENs) have been proposed as explainable-by-design neural models providing logic explanations for their predictions.

text-classification Text Classification

Global Explainability of GNNs via Logic Combination of Learned Concepts

1 code implementation13 Oct 2022 Steve Azzolin, Antonio Longa, Pietro Barbiero, Pietro Liò, Andrea Passerini

While instance-level explanation of GNN is a well-studied problem with plenty of approaches being developed, providing a global explanation for the behaviour of a GNN is much less explored, despite its potential in interpretability and debugging.

Diagnostic

Global Concept-Based Interpretability for Graph Neural Networks via Neuron Analysis

1 code implementation22 Aug 2022 Han Xuanyuan, Pietro Barbiero, Dobrik Georgiev, Lucie Charlotte Magister, Pietro Lió

We propose a novel approach for producing global explanations for GNNs using neuron-level concepts to enable practitioners to have a high-level view of the model.

On The Quality Assurance Of Concept-Based Representations

no code implementations29 Sep 2021 Mateo Espinosa Zarlenga, Pietro Barbiero, Zohreh Shams, Dmitry Kazhdan, Umang Bhatt, Mateja Jamnik

Recent work on Explainable AI has focused on concept-based explanations, where deep learning models are explained in terms of high-level units of information, referred to as concepts.

Disentanglement

Logic Explained Networks

2 code implementations11 Aug 2021 Gabriele Ciravegna, Pietro Barbiero, Francesco Giannini, Marco Gori, Pietro Lió, Marco Maggini, Stefano Melacci

The language used to communicate the explanations must be formal enough to be implementable in a machine and friendly enough to be understandable by a wide audience.

Explainable artificial intelligence

Algorithmic Concept-based Explainable Reasoning

1 code implementation15 Jul 2021 Dobrik Georgiev, Pietro Barbiero, Dmitry Kazhdan, Petar Veličković, Pietro Liò

Recent research on graph neural network (GNN) models successfully applied GNNs to classical graph algorithms and combinatorial optimisation problems.

Graph Neural Network

Entropy-based Logic Explanations of Neural Networks

3 code implementations12 Jun 2021 Pietro Barbiero, Gabriele Ciravegna, Francesco Giannini, Pietro Lió, Marco Gori, Stefano Melacci

Explainable artificial intelligence has rapidly emerged since lawmakers have started requiring interpretable models for safety-critical domains.

Explainable artificial intelligence Image Classification

Graph representation forecasting of patient's medical conditions: towards a digital twin

1 code implementation17 Sep 2020 Pietro Barbiero, Ramon Viñas Torné, Pietro Lió

Objective: Modern medicine needs to shift from a wait and react, curative discipline to a preventative, interdisciplinary science aiming at providing personalised, systemic and precise treatment plans to patients.

Generative Adversarial Network Graph Neural Network

Gradient-based Competitive Learning: Theory

no code implementations6 Sep 2020 Giansalvo Cirrincione, Pietro Barbiero, Gabriele Ciravegna, Vincenzo Randazzo

The former is just an adaptation of a standard competitive layer for deep clustering, while the latter is trained on the transposed matrix.

Clustering Deep Clustering +1

Topological Gradient-based Competitive Learning

1 code implementation21 Aug 2020 Pietro Barbiero, Gabriele Ciravegna, Vincenzo Randazzo, Giansalvo Cirrincione

The aim of this work is to present a novel comprehensive theory aspiring at bridging competitive learning with gradient-based learning, thus allowing the use of extremely powerful deep neural networks for feature extraction and projection combined with the remarkable flexibility and expressiveness of competitive learning.

Clustering Deep Clustering

Modeling Generalization in Machine Learning: A Methodological and Computational Study

1 code implementation28 Jun 2020 Pietro Barbiero, Giovanni Squillero, Alberto Tonda

As machine learning becomes more and more available to the general public, theoretical questions are turning into pressing practical issues.

BIG-bench Machine Learning

A Novel Outlook on Feature Selection as a Multi-objective Problem

1 code implementation Artificial Evolution 2020 Pietro Barbiero, Evelyne Lutton, Giovanni Squillero, Alberto Tonda

We thus propose a multi-objective optimization approach to feature selection, EvoFS, with the objectives to i. minimize feature subset size, ii.

feature selection

Uncovering Coresets for Classification With Multi-Objective Evolutionary Algorithms

1 code implementation20 Feb 2020 Pietro Barbiero, Giovanni Squillero, Alberto Tonda

A coreset is a subset of the training set, using which a machine learning algorithm obtains performances similar to what it would deliver if trained over the whole original data.

Classification Core set discovery +2

The GH-EXIN neural network for hierarchical clustering

1 code implementation Neural Networks 2020 Giansalvo Cirrincione, Gabriele Ciravegna, Pietro Barbiero, Vincenzo Randazzo, Eros Pasero

Furthermore, an important and very promising application of GH-EXIN in two-way hierarchical clustering, for the analysis of gene expression data in the study of the colorectal cancer is described.

Clustering Self-Organized Clustering

Cannot find the paper you are looking for? You can Submit a new open access paper.