no code implementations • 28 Feb 2024 • Niclas Göring, Florian Hess, Manuel Brenner, Zahra Monfared, Daniel Durstewitz
We explain why and how out-of-domain (OOD) generalization (OODG) in DSR profoundly differs from OODG considered elsewhere in machine learning.
1 code implementation • 7 Jun 2023 • Florian Hess, Zahra Monfared, Manuel Brenner, Daniel Durstewitz
Here we report that a surprisingly simple modification of teacher forcing leads to provably strictly all-time bounded gradients in training on chaotic systems, and, when paired with a simple architectural rearrangement of a tractable RNN design, piecewise-linear RNNs (PLRNNs), allows for faithful reconstruction in spaces of at most the dimensionality of the observed system.
no code implementations • 15 Dec 2022 • Manuel Brenner, Florian Hess, Georgia Koppe, Daniel Durstewitz
Many, if not most, systems of interest in science are naturally described as nonlinear dynamical systems.
1 code implementation • 6 Jul 2022 • Manuel Brenner, Florian Hess, Jonas M. Mikhaeil, Leonard Bereska, Zahra Monfared, Po-Chen Kuo, Daniel Durstewitz
In many scientific disciplines, we are interested in inferring the nonlinear dynamical system underlying a set of observed time series, a challenging task in the face of chaotic behavior and noise.
no code implementations • 29 Sep 2021 • Manuel Brenner, Leonard Bereska, Jonas Magdy Mikhaeil, Florian Hess, Zahra Monfared, Po-Chen Kuo, Daniel Durstewitz
In many scientific disciplines, we are interested in inferring the nonlinear dynamical system underlying a set of observed time series, a challenging task in the face of chaotic behavior and noise.