Search Results for author: Sebastian Goldt

Found 21 papers, 16 papers with code

Sliding down the stairs: how correlated latent variables accelerate learning with neural networks

no code implementations12 Apr 2024 Lorenzo Bardone, Sebastian Goldt

In particular, higher-order input cumulants (HOCs) are crucial for their performance.

Learning from higher-order statistics, efficiently: hypothesis tests, random features, and neural networks

1 code implementation22 Dec 2023 Eszter Székely, Lorenzo Bardone, Federica Gerace, Sebastian Goldt

Our results show that neural networks extract information from higher-order correlations in the spiked cumulant model efficiently, and reveal a large gap in the amount of data required by neural networks and random features to learn from higher-order cumulants.

Quantifying lottery tickets under label noise: accuracy, calibration, and complexity

1 code implementation21 Jun 2023 Viplove Arora, Daniele Irto, Sebastian Goldt, Guido Sanguinetti

Pruning deep neural networks is a widely used strategy to alleviate the computational burden in machine learning.

Attacks on Online Learners: a Teacher-Student Analysis

1 code implementation NeurIPS 2023 Riccardo Giuseppe Margiotta, Sebastian Goldt, Guido Sanguinetti

Machine learning models are famously vulnerable to adversarial attacks: small ad-hoc perturbations of the data that can catastrophically alter the model predictions.

Mapping of attention mechanisms to a generalized Potts model

no code implementations14 Apr 2023 Riccardo Rende, Federica Gerace, Alessandro Laio, Sebastian Goldt

In MLM, a word is randomly masked in an input sequence, and the network is trained to predict the missing word.

Language Modelling Masked Language Modeling

Neural networks trained with SGD learn distributions of increasing complexity

1 code implementation21 Nov 2022 Maria Refinetti, Alessandro Ingrosso, Sebastian Goldt

The ability of deep neural networks to generalise well even when they interpolate their training data has been explained using various "simplicity biases".

The impact of memory on learning sequence-to-sequence tasks

1 code implementation29 May 2022 Alireza Seif, Sarah A. M. Loos, Gennaro Tucci, Édgar Roldán, Sebastian Goldt

Here, we propose a simple model for a seq2seq task that has the advantage of providing explicit control over the degree of memory, or non-Markovianity, in the sequences -- the stochastic switching-Ornstein-Uhlenbeck (SSOU) model.

Machine Translation

Maslow's Hammer for Catastrophic Forgetting: Node Re-Use vs Node Activation

1 code implementation18 May 2022 Sebastian Lee, Stefano Sarao Mannelli, Claudia Clopath, Sebastian Goldt, Andrew Saxe

Continual learning - learning new tasks in sequence while maintaining performance on old tasks - remains particularly challenging for artificial neural networks.

Continual Learning

Data-driven emergence of convolutional structure in neural networks

no code implementations1 Feb 2022 Alessandro Ingrosso, Sebastian Goldt

Here, we show how initially fully-connected neural networks solving a discrimination task can learn a convolutional structure directly from their inputs, resulting in localised, space-tiling receptive fields.

Tensor Decomposition Translation

The dynamics of representation learning in shallow, non-linear autoencoders

1 code implementation6 Jan 2022 Maria Refinetti, Sebastian Goldt

We derive a set of asymptotically exact equations that describe the generalisation dynamics of autoencoders trained with stochastic gradient descent (SGD) in the limit of high-dimensional inputs.

Representation Learning

Continual Learning in the Teacher-Student Setup: Impact of Task Similarity

1 code implementation9 Jul 2021 Sebastian Lee, Sebastian Goldt, Andrew Saxe

Using each teacher to represent a different task, we investigate how the relationship between teachers affects the amount of forgetting and transfer exhibited by the student when the task switches.

Continual Learning

Redundant representations help generalization in wide neural networks

1 code implementation7 Jun 2021 Diego Doimo, Aldo Glielmo, Sebastian Goldt, Alessandro Laio

Deep neural networks (DNNs) defy the classical bias-variance trade-off: adding parameters to a DNN that interpolates its training data will typically improve its generalization performance.

Image Classification Learning Theory

Bayesian reconstruction of memories stored in neural networks from their connectivity

1 code implementation16 May 2021 Sebastian Goldt, Florent Krzakala, Lenka Zdeborová, Nicolas Brunel

The advent of comprehensive synaptic wiring diagrams of large neural circuits has created the field of connectomics and given rise to a number of open research questions.

Bayesian Inference

Classifying high-dimensional Gaussian mixtures: Where kernel methods fail and neural networks succeed

1 code implementation23 Feb 2021 Maria Refinetti, Sebastian Goldt, Florent Krzakala, Lenka Zdeborová

Here, we show theoretically that two-layer neural networks (2LNN) with only a few hidden neurons can beat the performance of kernel learning on a simple Gaussian mixture classification task.

Image Classification

Learning curves of generic features maps for realistic datasets with a teacher-student model

1 code implementation NeurIPS 2021 Bruno Loureiro, Cédric Gerbelot, Hugo Cui, Sebastian Goldt, Florent Krzakala, Marc Mézard, Lenka Zdeborová

While still solvable in a closed form, this generalization is able to capture the learning curves for a broad range of realistic data sets, thus redeeming the potential of the teacher-student framework.

Align, then memorise: the dynamics of learning with feedback alignment

1 code implementation24 Nov 2020 Maria Refinetti, Stéphane d'Ascoli, Ruben Ohana, Sebastian Goldt

Direct Feedback Alignment (DFA) is emerging as an efficient and biologically plausible alternative to the ubiquitous backpropagation algorithm for training deep neural networks.

The Gaussian equivalence of generative models for learning with shallow neural networks

1 code implementation25 Jun 2020 Sebastian Goldt, Bruno Loureiro, Galen Reeves, Florent Krzakala, Marc Mézard, Lenka Zdeborová

Here, we go beyond this simple paradigm by studying the performance of neural networks trained on data drawn from pre-trained generative models.

BIG-bench Machine Learning

Modelling the influence of data structure on learning in neural networks: the hidden manifold model

1 code implementation25 Sep 2019 Sebastian Goldt, Marc Mézard, Florent Krzakala, Lenka Zdeborová

We demonstrate that learning of the hidden manifold model is amenable to an analytical treatment by proving a "Gaussian Equivalence Property" (GEP), and we use the GEP to show how the dynamics of two-layer neural networks trained using one-pass stochastic gradient descent is captured by a set of integro-differential equations that track the performance of the network at all times.

Generative Adversarial Network

Generalisation dynamics of online learning in over-parameterised neural networks

no code implementations25 Jan 2019 Sebastian Goldt, Madhu S. Advani, Andrew M. Saxe, Florent Krzakala, Lenka Zdeborová

Deep neural networks achieve stellar generalisation on a variety of problems, despite often being large enough to easily fit all their training data.

Cannot find the paper you are looking for? You can Submit a new open access paper.