no code implementations • 7 Feb 2024 • Kevin Kögler, Alexander Shevchenko, Hamed Hassani, Marco Mondelli
For the prototypical case of the 1-bit compression of sparse Gaussian data, we prove that gradient descent converges to a solution that completely disregards the sparse structure of the input.
no code implementations • 27 Dec 2022 • Alexander Shevchenko, Kevin Kögler, Hamed Hassani, Marco Mondelli
Autoencoders are a popular model in many branches of machine learning and lossy data compression.
no code implementations • 3 Nov 2021 • Alexander Shevchenko, Vyacheslav Kungurtsev, Marco Mondelli
Understanding the properties of neural networks trained via stochastic gradient descent (SGD) is at the heart of the theory of deep learning.
no code implementations • ICML 2020 • Alexander Shevchenko, Marco Mondelli
In this paper, we shed light on this phenomenon: we show that the combination of stochastic gradient descent (SGD) and over-parameterization makes the landscape of multilayer neural networks approximately connected and thus more favorable to optimization.