no code implementations • 23 Jan 2024 • Elizaveta Demyanenko, Christoph Feinauer, Enrico M. Malatesta, Luca Saglietti
Recent works demonstrated the existence of a double-descent phenomenon for the generalization error of neural networks, where highly overparameterized models escape overfitting and achieve good test performance, at odds with the standard bias-variance trade-off described by statistical learning theory.
no code implementations • 17 Sep 2023 • Enrico M. Malatesta
In these pedagogic notes I review the statistical mechanics approach to neural networks, focusing on the paradigmatic example of the perceptron architecture with binary an continuous weights, in the classification setting.
no code implementations • 18 May 2023 • Brandon Livio Annesi, Clarissa Lauditi, Carlo Lucibello, Enrico M. Malatesta, Gabriele Perugini, Fabrizio Pittorino, Luca Saglietti
Empirical studies on the landscape of neural networks have shown that low-energy configurations are often found in complex connected structures, where zero-energy paths between pairs of distant solutions can be constructed.
no code implementations • 26 Apr 2023 • Carlo Baldassi, Enrico M. Malatesta, Gabriele Perugini, Riccardo Zecchina
We analyze the geometry of the landscape of solutions in both models and find important similarities and differences.
no code implementations • 1 Oct 2021 • Carlo Baldassi, Clarissa Lauditi, Enrico M. Malatesta, Rosalba Pacelli, Gabriele Perugini, Riccardo Zecchina
Current deep neural networks are highly overparameterized (up to billions of connection weights) and nonlinear.
no code implementations • 2 Jul 2021 • Carlo Baldassi, Clarissa Lauditi, Enrico M. Malatesta, Gabriele Perugini, Riccardo Zecchina
The success of deep learning has revealed the application potential of neural networks across the sciences and opened up fundamental theoretical problems.
no code implementations • 27 Oct 2020 • Carlo Baldassi, Enrico M. Malatesta, Matteo Negri, Riccardo Zecchina
We analyze the connection between minimizers with good generalizing properties and high local entropy regions of a threshold-linear classifier in Gaussian mixtures with the mean squared error loss function.
no code implementations • 17 Jul 2019 • Carlo Baldassi, Enrico M. Malatesta, Riccardo Zecchina
Rectified Linear Units (ReLU) have become the main model for the neural units in current deep learning systems.
no code implementations • 1 Feb 2019 • Enrico M. Malatesta
The second part of the thesis deals with mean-field combinatorial optimization problems.
Disordered Systems and Neural Networks