1 code implementation • 1 Feb 2024 • Aditya Bhattacharya, Simone Stumpf, Lucija Gosak, Gregor Stiglic, Katrien Verbert
Explanations in interactive machine-learning systems facilitate debugging and improving prediction models.
no code implementations • 3 Oct 2023 • Aditya Bhattacharya, Simone Stumpf, Lucija Gosak, Gregor Stiglic, Katrien Verbert
Our research involved a comprehensive examination of the impact of global explanations rooted in both data-centric and model-centric perspectives within systems designed to support healthcare experts in optimising machine learning models through both automated and manual data configurations.
no code implementations • 31 Jul 2023 • Ivania Donoso-Guzmán, Jeroen Ooge, Denis Parra, Katrien Verbert
While research on explainable AI (XAI) is booming and explanation techniques have proven promising in many application domains, standardised human-centred evaluation procedures are still missing.
no code implementations • 21 Feb 2023 • Aditya Bhattacharya, Jeroen Ooge, Gregor Stiglic, Katrien Verbert
Results indicate that our participants preferred our representation of data-centric explanations that provide local explanations with a global overview over other methods.
no code implementations • 28 Jan 2021 • Diego Rojo, Nyi Nyi Htun, Denis Parra, Robin De Croon, Katrien Verbert
To validate AHMoSe, we describe a use case scenario in the viticulture domain, grape quality prediction, where the system enables users to diagnose and select prediction models that perform better.
no code implementations • 20 Feb 2020 • Gregor Stiglic, Primoz Kocbek, Nino Fijacko, Marinka Zitnik, Katrien Verbert, Leona Cilar
There is a need of ensuring machine learning models that are interpretable.