no code implementations • 8 May 2024 • Lucas Kook, Chris Kolb, Philipp Schiele, Daniel Dold, Marcel Arpogaus, Cornelius Fritz, Philipp F. Baumann, Philipp Kopper, Tobias Pielok, Emilio Dorigatti, David Rügamer
Neural network representations of simple models, such as linear regression, are being studied increasingly to better understand the underlying principles of deep learning algorithms.
no code implementations • 3 May 2024 • David Rügamer, Chris Kolb, Tobias Weber, Lucas Kook, Thomas Nagler
The complexity of black-box algorithms can lead to various challenges, including the introduction of biases.
no code implementations • 7 Jul 2023 • Chris Kolb, Christian L. Müller, Bernd Bischl, David Rügamer
Additionally, our theory establishes results of independent interest regarding matching local minima for arbitrary, potentially unregularized, objectives.
2 code implementations • 6 Apr 2021 • David Rügamer, Chris Kolb, Cornelius Fritz, Florian Pfisterer, Philipp Kopper, Bernd Bischl, Ruolin Shen, Christina Bukas, Lisa Barros de Andrade e Sousa, Dominik Thalmeier, Philipp Baumann, Lucas Kook, Nadja Klein, Christian L. Müller
In this paper we describe the implementation of semi-structured deep distributional regression, a flexible framework to learn conditional distributions based on the combination of additive regression models and deep networks.
2 code implementations • 13 Feb 2020 • David Rügamer, Chris Kolb, Nadja Klein
We propose a general framework to combine structured regression models and deep neural networks into a unifying network architecture.