Search Results for author: Enrico M. Malatesta

Found 9 papers, 0 papers with code

The twin peaks of learning neural networks

no code implementations23 Jan 2024 Elizaveta Demyanenko, Christoph Feinauer, Enrico M. Malatesta, Luca Saglietti

Recent works demonstrated the existence of a double-descent phenomenon for the generalization error of neural networks, where highly overparameterized models escape overfitting and achieve good test performance, at odds with the standard bias-variance trade-off described by statistical learning theory.

Learning Theory

High-dimensional manifold of solutions in neural networks: insights from statistical physics

no code implementations17 Sep 2023 Enrico M. Malatesta

In these pedagogic notes I review the statistical mechanics approach to neural networks, focusing on the paradigmatic example of the perceptron architecture with binary an continuous weights, in the classification setting.

Linear Mode Connectivity

The star-shaped space of solutions of the spherical negative perceptron

no code implementations18 May 2023 Brandon Livio Annesi, Clarissa Lauditi, Carlo Lucibello, Enrico M. Malatesta, Gabriele Perugini, Fabrizio Pittorino, Luca Saglietti

Empirical studies on the landscape of neural networks have shown that low-energy configurations are often found in complex connected structures, where zero-energy paths between pairs of distant solutions can be constructed.

Typical and atypical solutions in non-convex neural networks with discrete and continuous weights

no code implementations26 Apr 2023 Carlo Baldassi, Enrico M. Malatesta, Gabriele Perugini, Riccardo Zecchina

We analyze the geometry of the landscape of solutions in both models and find important similarities and differences.

Unveiling the structure of wide flat minima in neural networks

no code implementations2 Jul 2021 Carlo Baldassi, Clarissa Lauditi, Enrico M. Malatesta, Gabriele Perugini, Riccardo Zecchina

The success of deep learning has revealed the application potential of neural networks across the sciences and opened up fundamental theoretical problems.

Wide flat minima and optimal generalization in classifying high-dimensional Gaussian mixtures

no code implementations27 Oct 2020 Carlo Baldassi, Enrico M. Malatesta, Matteo Negri, Riccardo Zecchina

We analyze the connection between minimizers with good generalizing properties and high local entropy regions of a threshold-linear classifier in Gaussian mixtures with the mean squared error loss function.

Random Combinatorial Optimization Problems: Mean Field and Finite-Dimensional Results

no code implementations1 Feb 2019 Enrico M. Malatesta

The second part of the thesis deals with mean-field combinatorial optimization problems.

Disordered Systems and Neural Networks

Cannot find the paper you are looking for? You can Submit a new open access paper.