Search Results for author: Pavel Dvurechensky

Found 24 papers, 7 papers with code

Analysis of Kernel Mirror Prox for Measure Optimization

no code implementations29 Feb 2024 Pavel Dvurechensky, Jia-Jie Zhu

By choosing a suitable function space as the dual to the non-negative measure cone, we study in a unified framework a class of functional saddle-point optimization problems, which we term the Mixed Functional Nash Equilibrium (MFNE), that underlies several existing machine learning algorithms, such as implicit generative models, distributionally robust optimization (DRO), and Wasserstein barycenters.

A conditional gradient homotopy method with applications to Semidefinite Programming

no code implementations7 Jul 2022 Pavel Dvurechensky, Shimrit Shtern, Mathias Staudigl

We propose a new homotopy-based conditional gradient method for solving convex optimization problems with a large number of simple conic constraints.

Combinatorial Optimization

Clipped Stochastic Methods for Variational Inequalities with Heavy-Tailed Noise

1 code implementation2 Jun 2022 Eduard Gorbunov, Marina Danilova, David Dobre, Pavel Dvurechensky, Alexander Gasnikov, Gauthier Gidel

In this work, we prove the first high-probability complexity results with logarithmic dependence on the confidence level for stochastic methods for solving monotone and structured non-monotone VIPs with non-sub-Gaussian (heavy-tailed) noise and unbounded domains.

Decentralized Local Stochastic Extra-Gradient for Variational Inequalities

no code implementations15 Jun 2021 Aleksandr Beznosikov, Pavel Dvurechensky, Anastasia Koloskova, Valentin Samokhin, Sebastian U Stich, Alexander Gasnikov

We extend the stochastic extragradient method to this very general setting and theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone (when a Minty solution exists) settings.

Federated Learning

Near-Optimal High Probability Complexity Bounds for Non-Smooth Stochastic Optimization with Heavy-Tailed Noise

1 code implementation10 Jun 2021 Eduard Gorbunov, Marina Danilova, Innokentiy Shibaev, Pavel Dvurechensky, Alexander Gasnikov

In our paper, we resolve this issue and derive the first high-probability convergence results with logarithmic dependence on the confidence level for non-smooth convex stochastic optimization problems with non-sub-Gaussian (heavy-tailed) noise.

Stochastic Optimization

Gradient Clipping Helps in Non-Smooth Stochastic Optimization with Heavy-Tailed Noise

no code implementations NeurIPS 2021 Eduard Gorbunov, Marina Danilova, Innokentiy Andreevich Shibaev, Pavel Dvurechensky, Alexander Gasnikov

In our paper, we resolve this issue and derive the first high-probability convergence results with logarithmical dependence on the confidence level for non-smooth convex stochastic optimization problems with non-sub-Gaussian (heavy-tailed) noise.

Stochastic Optimization

Decentralized Distributed Optimization for Saddle Point Problems

no code implementations15 Feb 2021 Alexander Rogozin, Alexander Beznosikov, Darina Dvinskikh, Dmitry Kovalev, Pavel Dvurechensky, Alexander Gasnikov

We consider distributed convex-concave saddle point problems over arbitrary connected undirected networks and propose a decentralized distributed algorithm for their solution.

Distributed Optimization Optimization and Control Distributed, Parallel, and Cluster Computing

First-Order Methods for Convex Optimization

no code implementations4 Jan 2021 Pavel Dvurechensky, Mathias Staudigl, Shimrit Shtern

In this survey we cover a number of key developments in gradient-based optimization methods.

Tensor methods for strongly convex strongly concave saddle point problems and strongly monotone variational inequalities

no code implementations31 Dec 2020 Petr Ostroukhov, Rinat Kamalov, Pavel Dvurechensky, Alexander Gasnikov

The first method is based on the assumption of $p$-th order smoothness of the objective and it achieves a convergence rate of $O \left( \left( \frac{L_p R^{p - 1}}{\mu} \right)^\frac{2}{p + 1} \log \frac{\mu R^2}{\varepsilon_G} \right)$, where $R$ is an estimate of the initial distance to the solution, and $\varepsilon_G$ is the error in terms of duality gap.

Optimization and Control

Inexact Tensor Methods and Their Application to Stochastic Convex Optimization

no code implementations31 Dec 2020 Artem Agafonov, Dmitry Kamzolov, Pavel Dvurechensky, Alexander Gasnikov

We propose general non-accelerated and accelerated tensor methods under inexact information on the derivatives of the objective, analyze their convergence rate.

Optimization and Control

Recent Theoretical Advances in Non-Convex Optimization

no code implementations11 Dec 2020 Marina Danilova, Pavel Dvurechensky, Alexander Gasnikov, Eduard Gorbunov, Sergey Guminov, Dmitry Kamzolov, Innokentiy Shibaev

For this setting, we first present known results for the convergence rates of deterministic first-order methods, which are then followed by a general theoretical analysis of optimal stochastic and randomized gradient schemes, and an overview of the stochastic first-order methods.

Zeroth-Order Algorithms for Smooth Saddle-Point Problems

no code implementations21 Sep 2020 Abdurakhmon Sadiev, Aleksandr Beznosikov, Pavel Dvurechensky, Alexander Gasnikov

In particular, our analysis shows that in the case when the feasible set is a direct product of two simplices, our convergence rate for the stochastic term is only by a $\log n$ factor worse than for the first-order methods.

Stochastic Saddle-Point Optimization for Wasserstein Barycenters

no code implementations11 Jun 2020 Daniil Tiapkin, Alexander Gasnikov, Pavel Dvurechensky

This leads to a complicated stochastic optimization problem where the objective is given as an expectation of a function given as a solution to a random optimization problem.

Stochastic Optimization

Accelerated meta-algorithm for convex optimization

2 code implementations18 Apr 2020 Darina Dvinskikh, Dmitry Kamzolov, Alexander Gasnikov, Pavel Dvurechensky, Dmitry Pasechnyk, Vladislav Matykhin, Alexei Chernov

We propose an accelerated meta-algorithm, which allows to obtain accelerated methods for convex unconstrained minimization in different settings.

Optimization and Control

Adaptive Gradient Descent for Convex and Non-Convex Stochastic Optimization

1 code implementation19 Nov 2019 Aleksandr Ogaltsov, Darina Dvinskikh, Pavel Dvurechensky, Alexander Gasnikov, Vladimir Spokoiny

In this paper we propose several adaptive gradient methods for stochastic optimization.

Optimization and Control

Generalized Self-concordant Hessian-barrier algorithms

no code implementations4 Nov 2019 Pavel Dvurechensky, Mathias Staudigl, César A. Uribe

Many problems in statistical learning, imaging, and computer vision involve the optimization of a non-convex objective function with singularities at the boundary of the feasible set.

On a Combination of Alternating Minimization and Nesterov's Momentum

no code implementations9 Jun 2019 Sergey Guminov, Pavel Dvurechensky, Nazarii Tupitsa, Alexander Gasnikov

In this paper we combine AM and Nesterov's acceleration to propose an accelerated alternating minimization algorithm.

Distributed Computation of Wasserstein Barycenters over Networks

no code implementations8 Mar 2018 César A. Uribe, Darina Dvinskikh, Pavel Dvurechensky, Alexander Gasnikov, Angelia Nedić

We propose a new \cu{class-optimal} algorithm for the distributed computation of Wasserstein Barycenters over networks.

An Accelerated Method for Derivative-Free Smooth Stochastic Convex Optimization

1 code implementation25 Feb 2018 Eduard Gorbunov, Pavel Dvurechensky, Alexander Gasnikov

In the two-point feedback setting, i. e. when pairs of function values are available, we propose an accelerated derivative-free algorithm together with its complexity analysis.

Optimization and Control Computational Complexity

Computational Optimal Transport: Complexity by Accelerated Gradient Descent Is Better Than by Sinkhorn's Algorithm

1 code implementation ICML 2018 Pavel Dvurechensky, Alexander Gasnikov, Alexey Kroshnin

We analyze two algorithms for approximating the general optimal transport (OT) distance between two discrete distributions of size $n$, up to accuracy $\varepsilon$.

Data Structures and Algorithms Optimization and Control

Cannot find the paper you are looking for? You can Submit a new open access paper.