1 code implementation • 15 Feb 2023 • Francesco Lomuscio, Paolo Bajardi, Alan Perotti, Elvio G. Amparore
Several explainable AI methods allow a Machine Learning user to get insights on the classification process of a black-box model in the form of local linear explanations.
no code implementations • 31 Aug 2022 • Gabriele D'Acunto, Gianmarco De Francisci Morales, Paolo Bajardi, Francesco Bonchi
Our model allows sampling an MN-DAG according to user-specified priors on the time-dependence and multiscale properties of the causal graph.
no code implementations • 17 Feb 2022 • Alan Perotti, Paolo Bajardi, Francesco Bonchi, André Panisson
Decoupling the feature space (edges) from a desired high-level explanation language (such as motifs) is thus a major challenge towards developing actionable explanations for graph classification tasks.
no code implementations • 9 Nov 2021 • Gabriele D'Acunto, Paolo Bajardi, Francesco Bonchi, Gianmarco De Francisci Morales
They link the evolution of the causal structure of equity risk factors with market volatility and a worsening macroeconomic environment, and show that, in times of financial crisis, exposure to different factors boils down to exposure to the market risk factor.
1 code implementation • 1 Jun 2021 • Elvio G. Amparore, Alan Perotti, Paolo Bajardi
This highlights the need to have standard and unbiased evaluation procedures for Local Linear Explanations in the XAI field.
Explainable artificial intelligence Explainable Artificial Intelligence (XAI)
no code implementations • 22 Feb 2021 • Duilio Balsamo, Paolo Bajardi, Alberto Salomone, Rossano Schifanella
We aimed to find a large cohort of Reddit users interested in discussing the use of opioids, trace the temporal evolution of their interest, and extensively characterize patterns of the nonmedical consumption of opioids, with a focus on routes of administration and drug tampering.
Information Retrieval Computers and Society
no code implementations • 8 Nov 2020 • Cecilia Panigutti, Alan Perotti, Andrè Panisson, Paolo Bajardi, Dino Pedreschi
The pervasive application of algorithmic decision-making is raising concerns on the risk of unintended bias in AI systems deployed in critical settings such as healthcare.