Search Results for author: Ard A. Louis

Found 9 papers, 1 papers with code

Do deep neural networks have an inbuilt Occam's razor?

no code implementations13 Apr 2023 Chris Mingard, Henry Rees, Guillermo Valle-Pérez, Ard A. Louis

The remarkable performance of overparameterized deep neural networks (DNNs) must arise from an interplay between network architecture, training algorithms, and structure in the data.

Inductive Bias

Double-descent curves in neural networks: a new perspective using Gaussian processes

no code implementations14 Feb 2021 Ouns El Harzli, Bernardo Cuenca Grau, Guillermo Valle-Pérez, Ard A. Louis

Double-descent curves in neural networks describe the phenomenon that the generalisation error initially descends with increasing parameters, then grows after reaching an optimal number of parameters which is less than the number of data points, but then descends again in the overparameterized regime.

Gaussian Processes Learning Theory

Generalization bounds for deep learning

no code implementations7 Dec 2020 Guillermo Valle-Pérez, Ard A. Louis

Here we introduce desiderata for techniques that predict generalization errors for deep learning models in supervised learning.

Generalization Bounds

Is SGD a Bayesian sampler? Well, almost

no code implementations26 Jun 2020 Chris Mingard, Guillermo Valle-Pérez, Joar Skalse, Ard A. Louis

Our main findings are that $P_{SGD}(f\mid S)$ correlates remarkably well with $P_B(f\mid S)$ and that $P_B(f\mid S)$ is strongly biased towards low-error and low complexity functions.

Gaussian Processes Inductive Bias

Simplicity bias in the parameter-function map of deep neural networks

no code implementations28 May 2019 Guillermo Valle-Pérez, Chico Q. Camargo, Ard A. Louis

Deep neural networks can be viewed as a mapping from the space of parameters (the weights) to the space of functions (how inputs get transformed to outputs by the network).

Deep learning generalizes because the parameter-function map is biased towards simple functions

no code implementations ICLR 2019 Guillermo Valle-Pérez, Chico Q. Camargo, Ard A. Louis

We then provide clear evidence for this strong simplicity bias in a model DNN for Boolean functions, as well as in much larger fully connected and convolutional networks applied to CIFAR10 and MNIST.

Gaussian Processes Learning Theory

Modelling the Self-Assembly of Virus Capsids

1 code implementation10 Oct 2009 Iain G. Johnston, Ard A. Louis, Jonathan P. K. Doye

We use computer simulations to study a model, first proposed by Wales [1], for the reversible and monodisperse self-assembly of simple icosahedral virus capsid structures.

Biomolecules

Cannot find the paper you are looking for? You can Submit a new open access paper.