Search Results for author: Dmitriy Drusvyatskiy

Found 22 papers, 8 papers with code

Linear Recursive Feature Machines provably recover low-rank matrices

1 code implementation9 Jan 2024 Adityanarayanan Radhakrishnan, Mikhail Belkin, Dmitriy Drusvyatskiy

A possible explanation is that common training algorithms for neural networks implicitly perform dimensionality reduction - a process called feature learning.

Dimensionality Reduction Low-Rank Matrix Completion +1

Asymptotic normality and optimality in nonsmooth stochastic approximation

no code implementations16 Jan 2023 Damek Davis, Dmitriy Drusvyatskiy, Liwei Jiang

In their seminal work, Polyak and Juditsky showed that stochastic approximation algorithms for solving smooth equations enjoy a central limit theorem.

Open-Ended Question Answering

Stochastic Approximation with Decision-Dependent Distributions: Asymptotic Normality and Optimality

1 code implementation9 Jul 2022 Joshua Cutler, Mateo Díaz, Dmitriy Drusvyatskiy

We show that under mild assumptions, the deviation between the average iterate of the algorithm and the solution is asymptotically normal, with a covariance that clearly decouples the effects of the gradient noise and the distributional shift.

Decision-Dependent Risk Minimization in Geometrically Decaying Dynamic Environments

no code implementations8 Apr 2022 Mitas Ray, Dmitriy Drusvyatskiy, Maryam Fazel, Lillian J. Ratliff

This paper studies the problem of expected loss minimization given a data distribution that is dependent on the decision-maker's action and evolves dynamically in time according to a geometric decay process.

Flat minima generalize for low-rank matrix recovery

no code implementations7 Mar 2022 Lijun Ding, Dmitriy Drusvyatskiy, Maryam Fazel, Zaid Harchaoui

Empirical evidence suggests that for a variety of overparameterized nonlinear models, most notably in neural network training, the growth of the loss around a minimizer strongly impacts its performance.

Matrix Completion

Multiplayer Performative Prediction: Learning in Decision-Dependent Games

no code implementations10 Jan 2022 Adhyyan Narang, Evan Faulkner, Dmitriy Drusvyatskiy, Maryam Fazel, Lillian J. Ratliff

We show that under mild assumptions, the performatively stable equilibria can be found efficiently by a variety of algorithms, including repeated retraining and the repeated (stochastic) gradient method.

Active manifolds, stratifications, and convergence to local minima in nonsmooth optimization

no code implementations26 Aug 2021 Damek Davis, Dmitriy Drusvyatskiy, Liwei Jiang

We show that the subgradient method converges only to local minimizers when applied to generic Lipschitz continuous and subdifferentially regular functions that are definable in an o-minimal structure.

Stochastic Optimization under Distributional Drift

1 code implementation NeurIPS 2021 Joshua Cutler, Dmitriy Drusvyatskiy, Zaid Harchaoui

We consider the problem of minimizing a convex function that is evolving according to unknown and possibly stochastic dynamics, which may depend jointly on time and on the decision variable itself.

Stochastic Optimization valid

Escaping strict saddle points of the Moreau envelope in nonsmooth optimization

no code implementations17 Jun 2021 Damek Davis, Mateo Díaz, Dmitriy Drusvyatskiy

The main conclusion is that a variety of algorithms for nonsmooth optimization can escape strict saddle points of the Moreau envelope at a controlled rate.

Proximal methods avoid active strict saddles of weakly convex functions

no code implementations16 Dec 2019 Damek Davis, Dmitriy Drusvyatskiy

We introduce a geometrically transparent strict saddle property for nonsmooth functions.

From low probability to high confidence in stochastic convex optimization

no code implementations31 Jul 2019 Damek Davis, Dmitriy Drusvyatskiy, Lin Xiao, Junyu Zhang

Standard results in stochastic convex optimization bound the number of samples that an algorithm needs to generate a point with small function value in expectation.

Stochastic Optimization Vocal Bursts Intensity Prediction

Stochastic algorithms with geometric step decay converge linearly on sharp functions

1 code implementation22 Jul 2019 Damek Davis, Dmitriy Drusvyatskiy, Vasileios Charisopoulos

In this work, we ask whether geometric step decay similarly improves stochastic algorithms for the class of sharp nonconvex problems.

Retrieval

Composite optimization for robust blind deconvolution

1 code implementation6 Jan 2019 Vasileios Charisopoulos, Damek Davis, Mateo Díaz, Dmitriy Drusvyatskiy

The blind deconvolution problem seeks to recover a pair of vectors from a set of rank one bilinear measurements.

Graphical Convergence of Subgradients in Nonconvex Optimization and Learning

no code implementations17 Oct 2018 Damek Davis, Dmitriy Drusvyatskiy

We investigate the stochastic optimization problem of minimizing population risk, where the loss defining the risk is assumed to be weakly convex.

regression Stochastic Optimization

Stochastic model-based minimization under high-order growth

no code implementations1 Jul 2018 Damek Davis, Dmitriy Drusvyatskiy, Kellie J. MacPhee

Given a nonsmooth, nonconvex minimization problem, we consider algorithms that iteratively sample and minimize stochastic convex models of the objective function.

Vocal Bursts Intensity Prediction

Stochastic subgradient method converges on tame functions

1 code implementation20 Apr 2018 Damek Davis, Dmitriy Drusvyatskiy, Sham Kakade, Jason D. Lee

This work considers the question: what convergence guarantees does the stochastic subgradient method have in the absence of smoothness and convexity?

Stochastic model-based minimization of weakly convex functions

no code implementations17 Mar 2018 Damek Davis, Dmitriy Drusvyatskiy

We consider a family of algorithms that successively sample and minimize simple stochastic models of the objective function.

Stochastic subgradient method converges at the rate $O(k^{-1/4})$ on weakly convex functions

2 code implementations8 Feb 2018 Damek Davis, Dmitriy Drusvyatskiy

We prove that the proximal stochastic subgradient method, applied to a weakly convex problem, drives the gradient of the Moreau envelope to zero at the rate $O(k^{-1/4})$.

Open-Ended Question Answering

The many faces of degeneracy in conic optimization

2 code implementations12 Jun 2017 Dmitriy Drusvyatskiy, Henry Wolkowicz

Slater's condition -- existence of a "strictly feasible solution" -- is a common assumption in conic optimization.

Optimization and Control

Catalyst Acceleration for Gradient-Based Non-Convex Optimization

no code implementations31 Mar 2017 Courtney Paquette, Hongzhou Lin, Dmitriy Drusvyatskiy, Julien Mairal, Zaid Harchaoui

We introduce a generic scheme to solve nonconvex optimization problems using gradient-based algorithms originally designed for minimizing convex functions.

Cannot find the paper you are looking for? You can Submit a new open access paper.