no code implementations • 9 Dec 2024 • William T. Redman, Zhangyang Wang, Alessandro Ingrosso, Sebastian Goldt
We develop a new method for measuring the effect of individual weights on the statistics of the FCN representations ("cavity method"), which allows us to find evidence in support of this hypothesis.
no code implementations • 25 Oct 2024 • Riccardo Rende, Federica Gerace, Alessandro Laio, Sebastian Goldt
The remarkable capability of over-parameterised neural networks to generalise effectively has been explained by invoking a ``simplicity bias'': neural networks prevent overfitting by initially learning simple classifiers before progressing to more complex, non-linear functions.
no code implementations • 12 Apr 2024 • Lorenzo Bardone, Sebastian Goldt
In particular, higher-order input cumulants (HOCs) are crucial for their performance.
1 code implementation • 22 Dec 2023 • Eszter Székely, Lorenzo Bardone, Federica Gerace, Sebastian Goldt
Our results show that neural networks extract information from higher-ordercorrelations in the spiked cumulant model efficiently, and reveal a large gap in the amount of data required by neural networks and random features to learn from higher-order cumulants.
1 code implementation • 21 Jun 2023 • Viplove Arora, Daniele Irto, Sebastian Goldt, Guido Sanguinetti
Pruning deep neural networks is a widely used strategy to alleviate the computational burden in machine learning.
no code implementations • 17 Jun 2023 • Nishil Patel, Sebastian Lee, Stefano Sarao Mannelli, Sebastian Goldt, Andrew Saxe
Reinforcement learning (RL) algorithms have proven transformative in a range of domains.
1 code implementation • NeurIPS 2023 • Riccardo Giuseppe Margiotta, Sebastian Goldt, Guido Sanguinetti
Machine learning models are famously vulnerable to adversarial attacks: small ad-hoc perturbations of the data that can catastrophically alter the model predictions.
no code implementations • 14 Apr 2023 • Riccardo Rende, Federica Gerace, Alessandro Laio, Sebastian Goldt
In MLM, a word is randomly masked in an input sequence, and the network is trained to predict the missing word.
1 code implementation • 21 Nov 2022 • Maria Refinetti, Alessandro Ingrosso, Sebastian Goldt
The ability of deep neural networks to generalise well even when they interpolate their training data has been explained using various "simplicity biases".
1 code implementation • 29 May 2022 • Alireza Seif, Sarah A. M. Loos, Gennaro Tucci, Édgar Roldán, Sebastian Goldt
Here, we propose a simple model for a seq2seq task that has the advantage of providing explicit control over the degree of memory, or non-Markovianity, in the sequences -- the stochastic switching-Ornstein-Uhlenbeck (SSOU) model.
1 code implementation • 18 May 2022 • Sebastian Lee, Stefano Sarao Mannelli, Claudia Clopath, Sebastian Goldt, Andrew Saxe
Continual learning - learning new tasks in sequence while maintaining performance on old tasks - remains particularly challenging for artificial neural networks.
no code implementations • 1 Feb 2022 • Alessandro Ingrosso, Sebastian Goldt
Here, we show how initially fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs, resulting in localised, space-tiling receptive fields.
1 code implementation • 6 Jan 2022 • Maria Refinetti, Sebastian Goldt
We derive a set of asymptotically exact equations that describe the generalisation dynamics of autoencoders trained with stochastic gradient descent (SGD) in the limit of high-dimensional inputs.
1 code implementation • 9 Jul 2021 • Sebastian Lee, Sebastian Goldt, Andrew Saxe
Using each teacher to represent a different task, we investigate how the relationship between teachers affects the amount of forgetting and transfer exhibited by the student when the task switches.
1 code implementation • 7 Jun 2021 • Diego Doimo, Aldo Glielmo, Sebastian Goldt, Alessandro Laio
Deep neural networks (DNNs) defy the classical bias-variance trade-off: adding parameters to a DNN that interpolates its training data will typically improve its generalization performance.
1 code implementation • 16 May 2021 • Sebastian Goldt, Florent Krzakala, Lenka Zdeborová, Nicolas Brunel
The advent of comprehensive synaptic wiring diagrams of large neural circuits has created the field of connectomics and given rise to a number of open research questions.
1 code implementation • 23 Feb 2021 • Maria Refinetti, Sebastian Goldt, Florent Krzakala, Lenka Zdeborová
Here, we show theoretically that two-layer neural networks (2LNN) with only a few hidden neurons can beat the performance of kernel learning on a simple Gaussian mixture classification task.
1 code implementation • NeurIPS 2021 • Bruno Loureiro, Cédric Gerbelot, Hugo Cui, Sebastian Goldt, Florent Krzakala, Marc Mézard, Lenka Zdeborová
While still solvable in a closed form, this generalization is able to capture the learning curves for a broad range of realistic data sets, thus redeeming the potential of the teacher-student framework.
1 code implementation • 24 Nov 2020 • Maria Refinetti, Stéphane d'Ascoli, Ruben Ohana, Sebastian Goldt
Direct Feedback Alignment (DFA) is emerging as an efficient and biologically plausible alternative to the ubiquitous backpropagation algorithm for training deep neural networks.
1 code implementation • 25 Jun 2020 • Sebastian Goldt, Bruno Loureiro, Galen Reeves, Florent Krzakala, Marc Mézard, Lenka Zdeborová
Here, we go beyond this simple paradigm by studying the performance of neural networks trained on data drawn from pre-trained generative models.
1 code implementation • 25 Sep 2019 • Sebastian Goldt, Marc Mézard, Florent Krzakala, Lenka Zdeborová
We demonstrate that learning of the hidden manifold model is amenable to an analytical treatment by proving a "Gaussian Equivalence Property" (GEP), and we use the GEP to show how the dynamics of two-layer neural networks trained using one-pass stochastic gradient descent is captured by a set of integro-differential equations that track the performance of the network at all times.
3 code implementations • NeurIPS 2019 • Sebastian Goldt, Madhu S. Advani, Andrew M. Saxe, Florent Krzakala, Lenka Zdeborová
Deep neural networks achieve stellar generalisation even when they have enough parameters to easily fit all their training data.
no code implementations • 25 Jan 2019 • Sebastian Goldt, Madhu S. Advani, Andrew M. Saxe, Florent Krzakala, Lenka Zdeborová
Deep neural networks achieve stellar generalisation on a variety of problems, despite often being large enough to easily fit all their training data.