1 code implementation • NeurIPS 2023 • Maximilian Mueller, Tiffany Vlaar, David Rolnick, Matthias Hein
Sharpness-aware minimization (SAM) was proposed to reduce sharpness of minima and has been shown to enhance generalization performance in various settings.
no code implementations • 30 Jun 2021 • Tiffany Vlaar, Jonathan Frankle
In this paper, we put inferences of this kind to the test, systematically evaluating how linear interpolation and final performance vary when altering the data, choice of initialization, and other optimizer and architecture design choices.
1 code implementation • 20 Jun 2021 • Tiffany Vlaar, Benedict Leimkuhler
We also discuss splitting choices for the neural network parameters which could enhance generalization performance when neural networks are trained from scratch.
1 code implementation • 20 Jun 2021 • Benedict Leimkuhler, Tiffany Vlaar, Timothée Pouchon, Amos Storkey
We employ constraints to control the parameter space of deep neural networks throughout training.
no code implementations • 17 Jun 2020 • Benedict Leimkuhler, Timothée Pouchon, Tiffany Vlaar, Amos Storkey
We propose a method for efficiently incorporating constraints into a stochastic gradient Langevin framework for the training of deep neural networks.
1 code implementation • 30 Aug 2019 • Benedict Leimkuhler, Charles Matthews, Tiffany Vlaar
We describe easy-to-implement hybrid partitioned numerical algorithms, based on discretized stochastic differential equations, which are adapted to feed-forward neural networks, including a multi-layer Langevin algorithm, AdLaLa (combining the adaptive Langevin and Langevin algorithms) and LOL (combining Langevin and Overdamped Langevin); we examine the convergence of these methods using numerical studies and compare their performance among themselves and in relation to standard alternatives such as stochastic gradient descent and ADAM.