no code implementations • 8 Mar 2024 • Hristo Papazov, Scott Pesme, Nicolas Flammarion
In this work, we investigate the effect of momentum on the optimisation trajectory of gradient descent.
no code implementations • 17 Feb 2023 • Mathieu Even, Scott Pesme, Suriya Gunasekar, Nicolas Flammarion
In this paper, we investigate the impact of stochasticity and large stepsizes on the implicit regularisation of gradient descent (GD) and stochastic gradient descent (SGD) over diagonal linear networks.
no code implementations • NeurIPS 2021 • Scott Pesme, Loucas Pillaud-Vivien, Nicolas Flammarion
Understanding the implicit bias of training algorithms is of crucial importance in order to explain the success of overparametrised neural networks.
no code implementations • ICML 2020 • Scott Pesme, Aymeric Dieuleveut, Nicolas Flammarion
Constant step-size Stochastic Gradient Descent exhibits two phases: a transient phase during which iterates make fast progress towards the optimum, followed by a stationary phase during which iterates oscillate around the optimal point.
no code implementations • NeurIPS 2020 • Scott Pesme, Nicolas Flammarion
We consider the robust linear regression problem in the online setting where we have access to the data in a streaming manner, one data point after the other.