Search Results for author: David G. T. Barrett

Found 13 papers, 6 papers with code

Why neural networks find simple solutions: the many regularizers of geometric complexity

no code implementations27 Sep 2022 Benoit Dherin, Michael Munn, Mihaela Rosca, David G. T. Barrett

Using a combination of theoretical arguments and empirical results, we show that many common training heuristics such as parameter norm regularization, spectral norm regularization, flatness regularization, implicit gradient regularization, noise regularization and the choice of parameter initialization all act to control geometric complexity, providing a unifying framework in which to characterize the behavior of deep learning models.

The Geometric Occam's Razor Implicit in Deep Learning

no code implementations30 Nov 2021 Benoit Dherin, Michael Munn, David G. T. Barrett

We argue that over-parameterized neural networks trained with stochastic gradient descent are subject to a Geometric Occam's Razor; that is, these networks are implicitly regularized by the geometric model complexity.

Discretization Drift in Two-Player Games

3 code implementations28 May 2021 Mihaela Rosca, Yan Wu, Benoit Dherin, David G. T. Barrett

Gradient-based methods for two-player games produce rich dynamics that can solve challenging problems, yet can be difficult to stabilize and understand.

Vocal Bursts Valence Prediction

On the Origin of Implicit Regularization in Stochastic Gradient Descent

no code implementations ICLR 2021 Samuel L. Smith, Benoit Dherin, David G. T. Barrett, Soham De

To interpret this phenomenon we prove that for SGD with random shuffling, the mean SGD iterate also stays close to the path of gradient flow if the learning rate is small and finite, but on a modified loss.

Implicit Gradient Regularization

no code implementations ICLR 2021 David G. T. Barrett, Benoit Dherin

We call this Implicit Gradient Regularization (IGR) and we use backward error analysis to calculate the size of this regularization.

Learning to Make Analogies by Contrasting Abstract Relational Structure

2 code implementations ICLR 2019 Felix Hill, Adam Santoro, David G. T. Barrett, Ari S. Morcos, Timothy Lillicrap

Here, we study how analogical reasoning can be induced in neural networks that learn to perceive and reason about raw visual data.

Analyzing biological and artificial neural networks: challenges with opportunities for synergy?

no code implementations31 Oct 2018 David G. T. Barrett, Ari S. Morcos, Jakob H. Macke

We explore opportunities for synergy between the two fields, such as the use of DNNs as in-silico model systems for neuroscience, and how this synergy can lead to new hypotheses about the operating principles of biological neural networks.

Object Recognition

Measuring abstract reasoning in neural networks

2 code implementations ICML 2018 David G. T. Barrett, Felix Hill, Adam Santoro, Ari S. Morcos, Timothy Lillicrap

To succeed at this challenge, models must cope with various generalisation `regimes' in which the training and test data differ in clearly-defined ways.

On the importance of single directions for generalization

1 code implementation ICLR 2018 Ari S. Morcos, David G. T. Barrett, Neil C. Rabinowitz, Matthew Botvinick

Finally, we find that class selectivity is a poor predictor of task importance, suggesting not only that networks which generalize well minimize their dependence on individual units by reducing their selectivity, but also that individually selective units may not be necessary for strong network performance.

Cognitive Psychology for Deep Neural Networks: A Shape Bias Case Study

no code implementations ICML 2017 Samuel Ritter, David G. T. Barrett, Adam Santoro, Matt M. Botvinick

To explore the potential value of these tools, we chose a well-established analysis from developmental psychology that explains how children learn word labels for objects, and applied that analysis to DNNs.

One-Shot Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.