Search Results for author: Courtney Paquette

Found 12 papers, 0 papers with code

Mirror Descent Algorithms with Nearly Dimension-Independent Rates for Differentially-Private Stochastic Saddle-Point Problems

no code implementations5 Mar 2024 Tomás González, Cristóbal Guzmán, Courtney Paquette

For convex-concave and first-order-smooth stochastic objectives, our algorithms attain a rate of $\sqrt{\log(d)/n} + (\log(d)^{3/2}/[n\varepsilon])^{1/3}$, where $d$ is the dimension of the problem and $n$ the dataset size.

LEMMA

Hitting the High-Dimensional Notes: An ODE for SGD learning dynamics on GLMs and multi-index models

no code implementations17 Aug 2023 Elizabeth Collins-Woodfin, Courtney Paquette, Elliot Paquette, Inbar Seroussi

In addition to the deterministic equivalent, we introduce an SDE with a simplified diffusion coefficient (homogenized SGD) which allows us to analyze the dynamics of general statistics of SGD iterates.

Retrieval

Only Tails Matter: Average-Case Universality and Robustness in the Convex Regime

no code implementations20 Jun 2022 Leonardo Cunha, Gauthier Gidel, Fabian Pedregosa, Damien Scieur, Courtney Paquette

The recently developed average-case analysis of optimization methods allows a more fine-grained and representative convergence analysis than usual worst-case results.

Implicit Regularization or Implicit Conditioning? Exact Risk Trajectories of SGD in High Dimensions

no code implementations15 Jun 2022 Courtney Paquette, Elliot Paquette, Ben Adlam, Jeffrey Pennington

Stochastic gradient descent (SGD) is a pillar of modern machine learning, serving as the go-to optimization algorithm for a diverse array of problems.

Computational Efficiency

Trajectory of Mini-Batch Momentum: Batch Size Saturation and Convergence in High Dimensions

no code implementations2 Jun 2022 Kiwon Lee, Andrew N. Cheng, Courtney Paquette, Elliot Paquette

We analyze the dynamics of large batch stochastic gradient descent with momentum (SGD+M) on the least squares problem when both the number of samples and dimensions are large.

Homogenization of SGD in high-dimensions: Exact dynamics and generalization properties

no code implementations14 May 2022 Courtney Paquette, Elliot Paquette, Ben Adlam, Jeffrey Pennington

By analyzing homogenized SGD, we provide exact non-asymptotic high-dimensional expressions for the generalization performance of SGD in terms of a solution of a Volterra integral equation.

Vocal Bursts Intensity Prediction

Dynamics of Stochastic Momentum Methods on Large-scale, Quadratic Models

no code implementations NeurIPS 2021 Courtney Paquette, Elliot Paquette

We analyze a class of stochastic gradient algorithms with momentum on a high-dimensional random least squares problem.

SGD in the Large: Average-case Analysis, Asymptotics, and Stepsize Criticality

no code implementations8 Feb 2021 Courtney Paquette, Kiwon Lee, Fabian Pedregosa, Elliot Paquette

We propose a new framework, inspired by random matrix theory, for analyzing the dynamics of stochastic gradient descent (SGD) when both number of samples and dimensions are large.

Halting Time is Predictable for Large Models: A Universality Property and Average-case Analysis

no code implementations8 Jun 2020 Courtney Paquette, Bart van Merriënboer, Elliot Paquette, Fabian Pedregosa

In fact, the halting time exhibits a universality property: it is independent of the probability distribution.

A termination criterion for stochastic gradient descent for binary classification

no code implementations23 Mar 2020 Sina Baghal, Courtney Paquette, Stephen A. Vavasis

We propose a new, simple, and computationally inexpensive termination test for constant step-size stochastic gradient descent (SGD) applied to binary classification on the logistic and hinge loss with homogeneous linear predictors.

Binary Classification Classification +1

Catalyst Acceleration for Gradient-Based Non-Convex Optimization

no code implementations31 Mar 2017 Courtney Paquette, Hongzhou Lin, Dmitriy Drusvyatskiy, Julien Mairal, Zaid Harchaoui

We introduce a generic scheme to solve nonconvex optimization problems using gradient-based algorithms originally designed for minimizing convex functions.

Cannot find the paper you are looking for? You can Submit a new open access paper.