Search Results for author: Johannes Haug

Found 8 papers, 8 papers with code

Change Detection for Local Explainability in Evolving Data Streams

1 code implementation6 Sep 2022 Johannes Haug, Alexander Braun, Stefan Zürn, Gjergji Kasneci

In particular, we show that local attributions can become obsolete each time the predictive model is updated or concept drift alters the data generating distribution.

Change Detection

Standardized Evaluation of Machine Learning Methods for Evolving Data Streams

1 code implementation28 Apr 2022 Johannes Haug, Effi Tramountani, Gjergji Kasneci

In this sense, we hope that our work will contribute to more standardized, reliable and realistic testing and comparison of online machine learning methods.

BIG-bench Machine Learning feature selection

Dynamic Model Tree for Interpretable Data Stream Learning

1 code implementation30 Mar 2022 Johannes Haug, Klaus Broelemann, Gjergji Kasneci

Dynamic Model Trees are thus a powerful online learning framework that contributes to more lightweight and interpretable machine learning in data streams.

BIG-bench Machine Learning Interpretable Machine Learning

Deep Neural Networks and Tabular Data: A Survey

2 code implementations5 Oct 2021 Vadim Borisov, Tobias Leemann, Kathrin Seßler, Johannes Haug, Martin Pawelczyk, Gjergji Kasneci

Moreover, we discuss deep learning approaches for generating tabular data, and we also provide an overview over strategies for explaining deep models on tabular data.

On Baselines for Local Feature Attributions

1 code implementation4 Jan 2021 Johannes Haug, Stefan Zürn, Peter El-Jiz, Gjergji Kasneci

Our experimental study illustrates the sensitivity of popular attribution models to the baseline, thus laying the foundation for a more in-depth discussion on sensible baseline methods for tabular data.

Fairness

Learning Parameter Distributions to Detect Concept Drift in Data Streams

2 code implementations19 Oct 2020 Johannes Haug, Gjergji Kasneci

By treating the parameters of a predictive model as random variables, we show that concept drift corresponds to a change in the distribution of optimal parameters.

Learning Model-Agnostic Counterfactual Explanations for Tabular Data

3 code implementations21 Oct 2019 Martin Pawelczyk, Johannes Haug, Klaus Broelemann, Gjergji Kasneci

On one hand, we suggest to complement the catalogue of counterfactual quality measures [1] using a criterion to quantify the degree of difficulty for a certain counterfactual suggestion.

counterfactual

Cannot find the paper you are looking for? You can Submit a new open access paper.