no code implementations • 16 Mar 2023 • Fabian Hinder, Valerie Vaquet, Johannes Brinkrolf, Barbara Hammer
To do so, we propose a methodology to reduce the explanation of concept drift to an explanation of models that are trained in a suitable way extracting relevant information regarding the drift.
no code implementations • 8 Feb 2023 • Valerie Vaquet, Fabian Hinder, Johannes Brinkrolf, Barbara Hammer
Learning from non-stationary data streams is a research direction that gains increasing interest as more data in form of streams becomes available, for example from social media, smartphones, or industrial process monitoring.
no code implementations • 2 Dec 2022 • Fabian Hinder, Valerie Vaquet, Johannes Brinkrolf, Barbara Hammer
More precisely, we relate a change of the ITTE to the presence of real drift, i. e., a changed posterior, and to a change of the training result under the assumption of optimality.
1 code implementation • 15 Feb 2022 • André Artelt, Johannes Brinkrolf, Roel Visser, Barbara Hammer
While machine learning models are usually assumed to always output a prediction, there also exist extensions in the form of reject options which allow the model to reject inputs where only a prediction with an unacceptably low certainty would be possible.
1 code implementation • 3 Mar 2021 • André Artelt, Valerie Vaquet, Riza Velioglu, Fabian Hinder, Johannes Brinkrolf, Malte Schilling, Barbara Hammer
Counterfactual explanations explain a behavior to the user by proposing actions -- as changes to the input -- that would cause a different (specified) behavior of the system.