Search Results for author: Tiffany Vlaar

Found 6 papers, 4 papers with code

Normalization Layers Are All That Sharpness-Aware Minimization Needs

1 code implementation NeurIPS 2023 Maximilian Mueller, Tiffany Vlaar, David Rolnick, Matthias Hein

Sharpness-aware minimization (SAM) was proposed to reduce sharpness of minima and has been shown to enhance generalization performance in various settings.

What can linear interpolation of neural network loss landscapes tell us?

no code implementations30 Jun 2021 Tiffany Vlaar, Jonathan Frankle

In this paper, we put inferences of this kind to the test, systematically evaluating how linear interpolation and final performance vary when altering the data, choice of initialization, and other optimizer and architecture design choices.

Multirate Training of Neural Networks

1 code implementation20 Jun 2021 Tiffany Vlaar, Benedict Leimkuhler

We also discuss splitting choices for the neural network parameters which could enhance generalization performance when neural networks are trained from scratch.

Transfer Learning

Better Training using Weight-Constrained Stochastic Dynamics

1 code implementation20 Jun 2021 Benedict Leimkuhler, Tiffany Vlaar, Timothée Pouchon, Amos Storkey

We employ constraints to control the parameter space of deep neural networks throughout training.

Constraint-Based Regularization of Neural Networks

no code implementations17 Jun 2020 Benedict Leimkuhler, Timothée Pouchon, Tiffany Vlaar, Amos Storkey

We propose a method for efficiently incorporating constraints into a stochastic gradient Langevin framework for the training of deep neural networks.

Image Classification

Partitioned integrators for thermodynamic parameterization of neural networks

1 code implementation30 Aug 2019 Benedict Leimkuhler, Charles Matthews, Tiffany Vlaar

We describe easy-to-implement hybrid partitioned numerical algorithms, based on discretized stochastic differential equations, which are adapted to feed-forward neural networks, including a multi-layer Langevin algorithm, AdLaLa (combining the adaptive Langevin and Langevin algorithms) and LOL (combining Langevin and Overdamped Langevin); we examine the convergence of these methods using numerical studies and compare their performance among themselves and in relation to standard alternatives such as stochastic gradient descent and ADAM.

Cannot find the paper you are looking for? You can Submit a new open access paper.