no code implementations • 7 Mar 2024 • Jonas Weidner, Ivan Ezhov, Michal Balcerak, Marie-Christin Metz, Sergey Litvinov, Sebastian Kaltenbach, Leonhard Feiner, Laurin Lux, Florian Kofler, Jana Lipkova, Jonas Latz, Daniel Rueckert, Bjoern Menze, Benedikt Wiestler
Biophysical modeling, particularly involving partial differential equations (PDEs), offers significant potential for tailoring disease treatment protocols to individual patients.
no code implementations • 23 May 2023 • Kexin Jin, ChenGuang Liu, Jonas Latz
Indeed, we introduce and study the Stochastic Gradient Langevin Diffusion (SGLDiff), a continuous-time Markov process that follows the Langevin diffusion corresponding to a data subset and switches this data subset after exponential waiting times.
1 code implementation • 8 Feb 2023 • Tamara G. Grossmann, Urszula Julia Komorowska, Jonas Latz, Carola-Bibiane Schönlieb
In terms of solution time and accuracy, physics-informed neural networks have not been able to outperform the finite element method in our study.
no code implementations • 8 Sep 2022 • Kexin Jin, Jonas Latz, ChenGuang Liu, Alessandro Scagliotti
This model is a piecewise-deterministic Markov process that represents the particle movement by an underdamped dynamical system and the data subsampling through a stochastic switching of the dynamical system.
no code implementations • 11 Aug 2022 • Jeremy Budd, Yves van Gennip, Jonas Latz, Simone Parisotto, Carola-Bibiane Schönlieb
Practical image segmentation tasks concern images which must be reconstructed from noisy, distorted, and/or incomplete observations.
no code implementations • 22 Mar 2022 • Jonas Latz
Sparse inversion and classification problems are ubiquitous in modern data science and imaging.
no code implementations • 7 Dec 2021 • Kexin Jin, Jonas Latz, ChenGuang Liu, Carola-Bibiane Schönlieb
Optimization problems with continuous data appear in, e. g., robust machine learning, functional data analysis, and variational inference.
no code implementations • 15 Apr 2020 • Jonas Latz
After introducing it, we study theoretical properties of the stochastic gradient process: We show that it converges weakly to the gradient flow with respect to the full target function, as the learning rate approaches zero.
no code implementations • 24 Jan 2020 • Daniel Kressner, Jonas Latz, Stefano Massei, Elisabeth Ullmann
Many techniques for data science and uncertainty quantification demand efficient tools to handle Gaussian random fields, which are defined in terms of their mean functions and covariance operators.
no code implementations • 26 Feb 2019 • Jonas Latz
The subject of this article is the introduction of a new concept of well-posedness of Bayesian inverse problems.
1 code implementation • 23 Jul 2018 • Matthieu Bulté, Jonas Latz, Elisabeth Ullmann
Importantly, the particle filters enable the adaptive updating of the estimate for $g$ as new observations become available.
Computation Numerical Analysis