no code implementations • 22 Apr 2024 • Paulo Yanez Sarmiento, Simon Witzke, Nadja Klein, Bernhard Y. Renard
Our approach enforces sparsity directly by pruning the relevance propagation for the different layers.
no code implementations • 18 Mar 2024 • Christian Schlauch, Christian Wirth, Nadja Klein
Previous work has shown that using such informative priors to regularize probabilistic deep learning (DL) models increases their performance and data-efficiency.
no code implementations • 12 Jan 2024 • Maximilian Kertel, Nadja Klein
We present a boosting-based method to learn additive Structural Equation Models (SEMs) from observational data, with a focus on the theoretical aspects of determining the causal order among variables.
no code implementations • 1 Jun 2023 • Till Baldenius, Nicolas Koch, Hannah Klauber, Nadja Klein
We use large-scale foot traffic data for millions of places in 315 US cities between 2018 and 2020 to estimate an index of experienced isolation in daily visits between whites and other ethnic groups.
1 code implementation • 19 May 2023 • Victor Medina-Olivares, Stefan Lessmann, Nadja Klein
We propose a novel method for predicting time-to-event in the presence of cure fractions based on flexible survivals models integrated into a deep neural network framework.
no code implementations • 11 May 2023 • Benedikt Lütke Schwienhorst, Lucas Kock, David J. Nott, Nadja Klein
A theoretical analysis shows that dropout regularization prefers rare but important features in both the mean and dispersion, generalizing an earlier result for conventional generalized linear models.
no code implementations • 13 Jan 2023 • Nikolaus Umlauf, Johannes Seiler, Mattias Wetscher, Thorsten Simon, Stefan Lang, Nadja Klein
Recently, fitting probabilistic models have gained importance in many areas but estimation of such distributional models with very large data sets is a difficult task.
no code implementations • 1 Nov 2022 • Christian Schlauch, Nadja Klein, Christian Wirth
We show that our method outperforms both non-informed and informed learning methods, that are often used in the literature.
1 code implementation • 19 Oct 2022 • Nikolaus Umlauf, Nadja Klein
Recently regression forests have been incorporated into the framework of distributional regression, a nowadays popular regression approach aiming at estimating complete conditional distributions rather than relating the mean of an output variable to input features only - as done classically.
1 code implementation • 3 Oct 2021 • Clara Hoffmann, Nadja Klein
To address this shortcoming we investigate efficient and scalable approximate inference for the implicit copula neural linear model of Klein, Nott and Smith (2021) in order to quantify uncertainty for the predictions of end-to-end learners.
2 code implementations • 6 Apr 2021 • David Rügamer, Chris Kolb, Cornelius Fritz, Florian Pfisterer, Philipp Kopper, Bernd Bischl, Ruolin Shen, Christina Bukas, Lisa Barros de Andrade e Sousa, Dominik Thalmeier, Philipp Baumann, Lucas Kook, Nadja Klein, Christian L. Müller
In this paper we describe the implementation of semi-structured deep distributional regression, a flexible framework to learn conditional distributions based on the combination of additive regression models and deep networks.
1 code implementation • 20 Dec 2020 • Manuel Carlan, Thomas Kneib, Nadja Klein
A simulation study demonstrates the competitiveness of our approach against its likelihood-based counterpart but also Bayesian additive models of location, scale and shape and Bayesian quantile regression.
Methodology
no code implementations • 5 Oct 2020 • Nadja Klein, Michael Stanley Smith, David J. Nott
Using data from the Australian National Electricity Market, we show that our deep time series models provide accurate short term probabilistic price forecasts, with the copula model dominating.
2 code implementations • 13 Feb 2020 • David Rügamer, Chris Kolb, Nadja Klein
We propose a general framework to combine structured regression models and deep neural networks into a unifying network architecture.
no code implementations • 25 Sep 2019 • Nikolaus Umlauf, Nadja Klein, Thorsten Simon, Achim Zeileis
At the core of the package are algorithms for highly-efficient Bayesian estimation and inference that can be applied to generalized additive models (GAMs) or generalized additive models for location, scale, and shape (GAMLSS), also known as distributional regression.
no code implementations • 26 Aug 2019 • Nadja Klein, David J. Nott, Michael Stanley Smith
The end result is a scalable distributional DNN regression method with marginally calibrated predictions, and our work complements existing methods for probability calibration.
1 code implementation • 9 Sep 2016 • Elisabeth Waldmann, David Taylor-Robinson, Nadja Klein, Thomas Kneib, Tania Pressler, Matthias Schmid, Andreas Mayr
Joint Models for longitudinal and time-to-event data have gained a lot of attention in the last few years as they are a helpful technique to approach common a data structure in clinical studies where longitudinal outcomes are recorded alongside event times.