no code implementations • 12 Dec 2024 • Fabian Hinder, Valerie Vaquet, David Komnick, Barbara Hammer
Besides the classical offline setup of machine learning, stream learning constitutes a well-established setup where data arrives over time in potentially non-stationary environments.
no code implementations • 25 Nov 2024 • Fabian Hinder, Valerie Vaquet, Barbara Hammer
Concept drift refers to the change of data distributions over time.
1 code implementation • 16 Oct 2024 • Felix Störck, Fabian Hinder, Johannes Brinkrolf, Benjamin Paassen, Valerie Vaquet, Barbara Hammer
The contribution of this work is twofold: 1) we develop a general framework for fair machine learning of partition-based models that does not depend on a specific fairness definition, and 2) we derive a fair version of learning vector quantization (LVQ) as a specific instantiation.
no code implementations • 16 Oct 2024 • Valerie Vaquet, Fabian Hinder, André Artelt, Inaam Ashraf, Janine Strotherm, Jonas Vaquet, Johannes Brinkrolf, Barbara Hammer
Research on methods for planning and controlling water distribution networks gains increasing relevance as the availability of drinking water will decrease as a consequence of climate change.
1 code implementation • 3 Jan 2024 • Valerie Vaquet, Fabian Hinder, Barbara Hammer
In this work, we explore the potential of model-loss-based and distribution-based drift detection methods to tackle leakage detection.
1 code implementation • 15 Dec 2023 • Fabian Hinder, Valerie Vaquet, Barbara Hammer
Concept drift, i. e., the change of the data generating distribution, can render machine learning models inaccurate.
1 code implementation • 24 Oct 2023 • Valerie Vaquet, Fabian Hinder, Jonas Vaquet, Kathrin Lammers, Lars Quakernack, Barbara Hammer
Facing climate change, the already limited availability of drinking water will decrease in the future rendering drinking water an increasingly scarce resource.
no code implementations • 24 Oct 2023 • Fabian Hinder, Valerie Vaquet, Barbara Hammer
In addition to providing a systematic literature review, this work provides precise mathematical definitions of the considered problems and contains standardized experiments on parametric artificial datasets allowing for a direct comparison of different strategies for detection and localization.
no code implementations • 16 Mar 2023 • Fabian Hinder, Valerie Vaquet, Johannes Brinkrolf, Barbara Hammer
To do so, we propose a methodology to reduce the explanation of concept drift to an explanation of models that are trained in a suitable way extracting relevant information regarding the drift.
no code implementations • 8 Feb 2023 • Valerie Vaquet, Fabian Hinder, Johannes Brinkrolf, Barbara Hammer
Learning from non-stationary data streams is a research direction that gains increasing interest as more data in form of streams becomes available, for example from social media, smartphones, or industrial process monitoring.
no code implementations • 2 Dec 2022 • Fabian Hinder, Valerie Vaquet, Johannes Brinkrolf, Barbara Hammer
More precisely, we relate a change of the ITTE to the presence of real drift, i. e., a changed posterior, and to a change of the training result under the assumption of optimality.
no code implementations • 13 May 2022 • Fabian Hinder, André Artelt, Valerie Vaquet, Barbara Hammer
The notion of concept drift refers to the phenomenon that the data generating distribution changes over time; as a consequence machine learning models may become inaccurate and need adjustment.
no code implementations • 19 Feb 2022 • Fabian Hinder, Valerie Vaquet, Barbara Hammer
In this paper, we analyze structural properties of the drift induced signals in the context of different metrics.
1 code implementation • 6 Apr 2021 • André Artelt, Fabian Hinder, Valerie Vaquet, Robert Feldhans, Barbara Hammer
We also propose a method for automatically finding regions in data space that are affected by a given model adaptation and thus should be explained.
1 code implementation • 3 Mar 2021 • André Artelt, Valerie Vaquet, Riza Velioglu, Fabian Hinder, Johannes Brinkrolf, Malte Schilling, Barbara Hammer
Counterfactual explanations explain a behavior to the user by proposing actions -- as changes to the input -- that would cause a different (specified) behavior of the system.