Search Results for author: Frederik Benzing

Found 4 papers, 4 papers with code

Random initialisations performing above chance and how to find them

1 code implementation15 Sep 2022 Frederik Benzing, Simon Schug, Robert Meier, Johannes von Oswald, Yassir Akram, Nicolas Zucchet, Laurence Aitchison, Angelika Steger

Neural networks trained with stochastic gradient descent (SGD) starting from different random initialisations typically find functionally very similar solutions, raising the question of whether there are meaningful differences between different SGD solutions.

Gradient Descent on Neurons and its Link to Approximate Second-Order Optimization

1 code implementation28 Jan 2022 Frederik Benzing

Second-order optimizers are thought to hold the potential to speed up neural network training, but due to the enormous size of the curvature matrix, they typically require approximations to be computationally tractable.

Unifying Regularisation Methods for Continual Learning

2 code implementations11 Jun 2020 Frederik Benzing

Moreover, we show that for SI the relation to the Fisher -- and in fact its performance -- is due to a previously unknown bias.

Continual Learning

Optimal Kronecker-Sum Approximation of Real Time Recurrent Learning

1 code implementation11 Feb 2019 Frederik Benzing, Marcelo Matheus Gauy, Asier Mujika, Anders Martinsson, Angelika Steger

In contrast, the online training algorithm Real Time Recurrent Learning (RTRL) provides untruncated gradients, with the disadvantage of impractically large computational costs.

Memorization

Cannot find the paper you are looking for? You can Submit a new open access paper.