no code implementations • 17 Jul 2024 • Biagio La Rosa
This thesis addresses this issue by contributing to the field of eXplainable AI, focusing on enhancing the interpretability of deep neural networks.
1 code implementation • NeurIPS 2023 • Biagio La Rosa, Leilani H. Gilpin, Roberto Capobianco
Compositional Explanations is a method for identifying logical formulas of concepts that approximate the neurons' behavior.
no code implementations • Computer Graphics Forum 2023 • Biagio La Rosa, Graziano Blasilli, Romain Bourqui, David Auber, Giuseppe Santucci, Roberto Capobianco, Enrico Bertini, Romain Giot, Marco Angelini
The survey concludes by identifying future research challenges and bridging activities that are helpful to strengthen the role of Visual Analytics as effective support for eXplainable Deep Learning and to foster the adoption of Visual Analytics solutions in the eXplainable Deep Learning community.
1 code implementation • IEEE Transactions on Artificial Intelligence 2022 • Alessio Ragno, Biagio La Rosa, Roberto Capobianco
We then evaluate the explanations of the interpretable models by comparing them with post-hoc approaches and self-explainable models.
1 code implementation • Applied Intelligence 2022 • Biagio La Rosa, Roberto Capobianco, Daniele Nardi
This paper presents Memory Wrap, a module (i. e, a set of layers) that can be added to deep learning models to improve their performance and interpretability in settings where few data are available.
1 code implementation • 16 Sep 2021 • Sayo M. Makinwa, Biagio La Rosa, Roberto Capobianco
The recent success of deep learning models in solving complex problems and in different domains has increased interest in understanding what they learn.
1 code implementation • 1 Jun 2021 • Biagio La Rosa, Roberto Capobianco, Daniele Nardi
Due to their black-box and data-hungry nature, deep learning techniques are not yet widely adopted for real-world applications in critical domains, like healthcare and justice.
1 code implementation • 11 Jul 2020 • Biagio La Rosa, Roberto Capobianco, Daniele Nardi
Our results show that we are able to explain agent’s decisions in (1) and to reconstruct the most relevant sentences used by the network to select the story ending in (2).