no code implementations • 30 Aug 2024 • Thomas Schnake, Farnoush Rezaei Jafari, Jonas Lederer, Ping Xiong, Shinichi Nakajima, Stefan Gugler, Grégoire Montavon, Klaus-Robert Müller
Explainable Artificial Intelligence (XAI) plays a crucial role in fostering transparency and trust in AI systems, where traditional XAI approaches typically offer one level of abstraction for explanations, often in the form of heatmaps highlighting single or multiple input features.
no code implementations • 15 Aug 2024 • Jacob Kauffmann, Jonas Dippel, Lukas Ruff, Wojciech Samek, Klaus-Robert Müller, Grégoire Montavon
Unsupervised learning has become an essential building block of AI systems.
1 code implementation • 11 Jun 2024 • Farnoush Rezaei Jafari, Grégoire Montavon, Klaus-Robert Müller, Oliver Eberle
Recent sequence modeling approaches using selective state space sequence models, referred to as Mamba models, have seen a surge of interest.
no code implementations • 12 Mar 2024 • Simon Letzgus, Klaus-Robert Müller, Grégoire Montavon
In regression, explanations need to be precisely formulated to address specific user queries (e. g.\ distinguishing between `Why is the output above 0?'
no code implementations • 30 Jan 2024 • Florian Bley, Sebastian Lapuschkin, Wojciech Samek, Grégoire Montavon
So far, the question of explaining predictive uncertainty, i. e. why a model 'doubts', has been scarcely studied.
no code implementations • 13 Oct 2023 • Oliver Eberle, Jochen Büttner, Hassan El-Hajj, Grégoire Montavon, Klaus-Robert Müller, Matteo Valleriani
An ML based analysis of these tables helps to unveil important facets of the spatio-temporal evolution of knowledge and innovation in the field of mathematical astronomy in the period, as taught at European universities.
no code implementations • 2 Oct 2023 • Sidney Bender, Christopher J. Anders, Pattarawatt Chormai, Heike Marxfeld, Jan Herrmann, Grégoire Montavon
This paper introduces a novel technique called counterfactual knowledge distillation (CFKD) to detect and remove reliance on confounders in deep learning models with the help of human expert feedback.
no code implementations • 12 Apr 2023 • Lorenz Linhardt, Klaus-Robert Müller, Grégoire Montavon
In this paper, we demonstrate that acceptance of explanations by the user is not a guarantee for a machine learning model to be robust against Clever Hans effects, which may remain undetected.
no code implementations • 11 Mar 2023 • Johanna Vielhaben, Sebastian Lapuschkin, Grégoire Montavon, Wojciech Samek
In this way, we extend the applicability of a family of XAI methods to domains (e. g. speech) where the input is only interpretable after a transformation.
Explainable artificial intelligence Explainable Artificial Intelligence (XAI) +3
no code implementations • 30 Dec 2022 • Pattarawat Chormai, Jan Herrmann, Klaus-Robert Müller, Grégoire Montavon
Explanations often take the form of a heatmap identifying input features (e. g. pixels) that are relevant to the model's decision.
no code implementations • CVPR 2023 • Alexander Binder, Leander Weber, Sebastian Lapuschkin, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
To address shortcomings of this test, we start by observing an experimental gap in the ranking of explanation methods between randomization-based sanity checks [1] and model output faithfulness measures (e. g. [25]).
1 code implementation • 15 Feb 2022 • Ameen Ali, Thomas Schnake, Oliver Eberle, Grégoire Montavon, Klaus-Robert Müller, Lior Wolf
Transformers have become an important workhorse of machine learning, with numerous applications.
no code implementations • 23 Aug 2021 • Siddhant Agarwal, Nicola Tosi, Pan Kessel, Doris Breuer, Grégoire Montavon
Using a dataset of 10, 525 two-dimensional simulations of the thermal evolution of the mantle of a Mars-like planet, we show that deep learning techniques can produce reliable parameterized surrogates (i. e. surrogates that predict state variables such as temperature based only on parameters) of the underlying partial differential equations.
1 code implementation • 9 Jun 2021 • Léo Andeol, Yusei Kawakami, Yuichiro Wada, Takafumi Kanamori, Klaus-Robert Müller, Grégoire Montavon
However, common ML losses do not give strong guarantees on how consistently the ML model performs for different domains, in particular, whether the model performs well on a domain at the expense of its performance on another domain.
no code implementations • 24 Sep 2020 • Lukas Ruff, Jacob R. Kauffmann, Robert A. Vandermeulen, Grégoire Montavon, Wojciech Samek, Marius Kloft, Thomas G. Dietterich, Klaus-Robert Müller
Deep learning approaches to anomaly detection have recently improved the state of the art in detection performance on complex datasets such as large collections of images or text.
1 code implementation • 12 Aug 2020 • Kateryna Melnyk, Stefan Klus, Grégoire Montavon, Tim Conrad
We demonstrate that our method can capture temporary changes in the time-evolving graph on both created synthetic data and real-world data.
no code implementations • 18 Jun 2020 • Jacob Kauffmann, Lukas Ruff, Grégoire Montavon, Klaus-Robert Müller
The 'Clever Hans' effect occurs when the learned model produces correct predictions based on the 'wrong' features.
Anomaly Detection Explainable Artificial Intelligence (XAI) +1
no code implementations • 5 Jun 2020 • Thomas Schnake, Oliver Eberle, Jonas Lederer, Shinichi Nakajima, Kristof T. Schütt, Klaus-Robert Müller, Grégoire Montavon
In this paper, we show that GNNs can in fact be naturally explained using higher-order expansions, i. e. by identifying groups of edges that jointly contribute to the prediction.
no code implementations • 17 Mar 2020 • Wojciech Samek, Grégoire Montavon, Sebastian Lapuschkin, Christopher J. Anders, Klaus-Robert Müller
With the broader and highly successful usage of machine learning in industry and the sciences, there has been a growing demand for Explainable AI.
1 code implementation • 11 Mar 2020 • Oliver Eberle, Jochen Büttner, Florian Kräutli, Klaus-Robert Müller, Matteo Valleriani, Grégoire Montavon
Many learning algorithms such as kernel machines, nearest neighbors, clustering, or anomaly detection, are based on the concept of 'distance' or 'similarity'.
no code implementations • 25 Sep 2019 • Leila Arras, Jose A. Arjona-Medina, Michael Widrich, Grégoire Montavon, Michael Gillhofer, Klaus-Robert Müller, Sepp Hochreiter, Wojciech Samek
While neural networks have acted as a strong unifying force in the design of modern AI systems, the neural network architectures themselves remain highly heterogeneous due to the variety of tasks to be solved.
no code implementations • 18 Jun 2019 • Jacob Kauffmann, Malte Esders, Lukas Ruff, Grégoire Montavon, Wojciech Samek, Klaus-Robert Müller
Cluster predictions of the obtained networks can then be quickly and accurately attributed to the input features.
1 code implementation • 26 Feb 2019 • Sebastian Lapuschkin, Stephan Wäldchen, Alexander Binder, Grégoire Montavon, Wojciech Samek, Klaus-Robert Müller
Current learning machines have successfully solved hard application problems, reaching high accuracy and displaying seemingly "intelligent" behavior.
1 code implementation • 13 Aug 2018 • Maximilian Alber, Sebastian Lapuschkin, Philipp Seegerer, Miriam Hägele, Kristof T. Schütt, Grégoire Montavon, Wojciech Samek, Klaus-Robert Müller, Sven Dähne, Pieter-Jan Kindermans
The presented library iNNvestigate addresses this by providing a common interface and out-of-the- box implementation for many analysis methods, including the reference implementation for PatternNet and PatternAttribution as well as for LRP-methods.
no code implementations • 29 Jun 2018 • Jacob Kauffmann, Grégoire Montavon, Luiz Alberto Lima, Shinichi Nakajima, Klaus-Robert Müller, Nico Görnitz
Detecting and explaining anomalies is a challenging effort.
no code implementations • 11 Jun 2018 • Christopher Anders, Grégoire Montavon, Wojciech Samek, Klaus-Robert Müller
We apply the deep Taylor / LRP technique to understand the deep network's classification decisions, and identify a "border effect": a tendency of the classifier to look mainly at the bordering frames of the input.
no code implementations • 16 May 2018 • Jacob Kauffmann, Klaus-Robert Müller, Grégoire Montavon
The proposed One-Class DTD is applicable to a number of common distance-based SVM kernels and is able to reliably explain a wide set of data anomalies.
1 code implementation • 18 Jul 2017 • Franziska Horn, Leila Arras, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
When dealing with large collections of documents, it is imperative to quickly get an overview of the texts' contents.
2 code implementations • 17 Jul 2017 • Franziska Horn, Leila Arras, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
When working with a new dataset, it is important to first explore and familiarize oneself with it, before applying any advanced machine learning algorithms.
no code implementations • 24 Jun 2017 • Grégoire Montavon, Wojciech Samek, Klaus-Robert Müller
This paper provides an entry point to the problem of interpreting a deep neural network model and explaining its predictions.
1 code implementation • WS 2017 • Leila Arras, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
Recently, a technique called Layer-wise Relevance Propagation (LRP) was shown to deliver insightful explanations in the form of input space relevances for understanding feed-forward neural network classification decisions.
1 code implementation • 23 Dec 2016 • Leila Arras, Franziska Horn, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
Text documents can be described by a number of abstract concepts such as semantic category, writing style, or sentiment.
no code implementations • NeurIPS 2016 • Grégoire Montavon, Klaus-Robert Müller, Marco Cuturi
This metric between observations can then be used to define the Wasserstein distance between the distribution induced by the Boltzmann machine on the one hand, and that given by the training sample on the other hand.
no code implementations • 24 Nov 2016 • Wojciech Samek, Grégoire Montavon, Alexander Binder, Sebastian Lapuschkin, Klaus-Robert Müller
Complex nonlinear models such as deep neural network (DNNs) have become an important tool for image classification, speech recognition, natural language processing, and many other fields of application.
no code implementations • 23 Jun 2016 • Farhad Arbabzadah, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
We further observe that the explanation method provides important insights into the nature of features of the base model, which allow one to assess the aptitude of the base model for a given transfer learning task.
1 code implementation • WS 2016 • Leila Arras, Franziska Horn, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
Layer-wise relevance propagation (LRP) is a recently proposed technique for explaining predictions of complex non-linear classifiers in terms of input variables.
1 code implementation • 4 Apr 2016 • Alexander Binder, Grégoire Montavon, Sebastian Bach, Klaus-Robert Müller, Wojciech Samek
Layer-wise relevance propagation is a framework which allows to decompose the prediction of a deep neural network computed over a sample, e. g. an image, down to relevance scores for the single input dimensions of the sample such as subpixels of an image.
4 code implementations • 8 Dec 2015 • Grégoire Montavon, Sebastian Bach, Alexander Binder, Wojciech Samek, Klaus-Robert Müller
Although our focus is on image classification, the method is applicable to a broad set of input data, learning tasks and network architectures.
no code implementations • CVPR 2016 • Sebastian Bach, Alexander Binder, Grégoire Montavon, Klaus-Robert Müller, Wojciech Samek
Fisher Vector classifiers and Deep Neural Networks (DNNs) are popular and successful algorithms for solving image classification problems.
1 code implementation • 21 Sep 2015 • Wojciech Samek, Alexander Binder, Grégoire Montavon, Sebastian Bach, Klaus-Robert Müller
Our main result is that the recently proposed Layer-wise Relevance Propagation (LRP) algorithm qualitatively and quantitatively provides a better explanation of what made a DNN arrive at a particular classification decision than the sensitivity-based approach or the deconvolution method.
no code implementations • 7 Jul 2015 • Grégoire Montavon, Klaus-Robert Müller, Marco Cuturi
The Boltzmann machine provides a useful framework to learn highly complex, multimodal and multiscale data distributions that occur in the real world.
no code implementations • NeurIPS 2012 • Grégoire Montavon, Katja Hansen, Siamac Fazli, Matthias Rupp, Franziska Biegler, Andreas Ziehe, Alexandre Tkatchenko, Anatole V. Lilienfeld, Klaus-Robert Müller
The accurate prediction of molecular energetics in chemical compound space is a crucial ingredient for rational compound design.
no code implementations • NeurIPS 2010 • Grégoire Montavon, Klaus-Robert Müller, Mikio L. Braun
Deep networks can potentially express a learning problem more efficiently than local learning machines.