1 code implementation • 1 Dec 2022 • Riza Velioglu, Jan Philip Göpfert, André Artelt, Barbara Hammer
On the other hand, Machine Learning (ML) benefits from the vast amount of data available and can deal with high-dimensional sources, yet it has rarely been applied to being used in processes.
Explainable artificial intelligence Explainable Artificial Intelligence (XAI)
no code implementations • 25 Dec 2020 • Jan Philip Göpfert, Ulrike Kuhl, Lukas Hindemith, Heiko Wersing, Barbara Hammer
After developing a theoretical framework of intuitiveness as a property of algorithms, we introduce an active teaching paradigm involving a prototypical two-dimensional spatial learning task as a method to judge the efficacy of human-machine interactions.
no code implementations • 8 Nov 2020 • Jan Philip Göpfert, Heiko Wersing, Barbara Hammer
When training automated systems, it has been shown to be beneficial to adapt the representation of data by learning a problem-specific metric.
2 code implementations • 22 Apr 2020 • Niklas Risse, Christina Göpfert, Jan Philip Göpfert
Adversarial robustness of machine learning models has attracted considerable attention over recent years.
no code implementations • 21 Oct 2019 • Jan Philip Göpfert, Heiko Wersing, Barbara Hammer
In this contribution, we focus on the capabilities of explainers for convolutional deep neural networks in an extreme situation: a setting in which humans and networks fundamentally disagree.
no code implementations • 31 Jul 2019 • Christina Göpfert, Jan Philip Göpfert, Barbara Hammer
The existence of adversarial examples has led to considerable uncertainty regarding the trust one can justifiably put in predictions produced by automated systems.
1 code implementation • 25 Feb 2019 • Jan Philip Göpfert, André Artelt, Heiko Wersing, Barbara Hammer
Convolutional neural networks have been used to achieve a string of successes during recent years, but their lack of interpretability remains a serious issue.