no code implementations • 29 Feb 2024 • Pavel Dvurechensky, Jia-Jie Zhu
By choosing a suitable function space as the dual to the non-negative measure cone, we study in a unified framework a class of functional saddle-point optimization problems, which we term the Mixed Functional Nash Equilibrium (MFNE), that underlies several existing machine learning algorithms, such as implicit generative models, distributionally robust optimization (DRO), and Wasserstein barycenters.
no code implementations • 3 Oct 2023 • Eduard Gorbunov, Abdurakhmon Sadiev, Marina Danilova, Samuel Horváth, Gauthier Gidel, Pavel Dvurechensky, Alexander Gasnikov, Peter Richtárik
High-probability analysis of stochastic first-order optimization methods under mild assumptions on the noise has been gaining a lot of attention in recent years.
no code implementations • 2 Feb 2023 • Abdurakhmon Sadiev, Marina Danilova, Eduard Gorbunov, Samuel Horváth, Gauthier Gidel, Pavel Dvurechensky, Alexander Gasnikov, Peter Richtárik
During recent years the interest of optimization and machine learning communities in high-probability convergence of stochastic optimization methods has been growing.
no code implementations • 7 Jul 2022 • Pavel Dvurechensky, Shimrit Shtern, Mathias Staudigl
We propose a new homotopy-based conditional gradient method for solving convex optimization problems with a large number of simple conic constraints.
1 code implementation • 2 Jun 2022 • Eduard Gorbunov, Marina Danilova, David Dobre, Pavel Dvurechensky, Alexander Gasnikov, Gauthier Gidel
In this work, we prove the first high-probability complexity results with logarithmic dependence on the confidence level for stochastic methods for solving monotone and structured non-monotone VIPs with non-sub-Gaussian (heavy-tailed) noise and unbounded domains.
no code implementations • 15 Jun 2021 • Aleksandr Beznosikov, Pavel Dvurechensky, Anastasia Koloskova, Valentin Samokhin, Sebastian U Stich, Alexander Gasnikov
We extend the stochastic extragradient method to this very general setting and theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone (when a Minty solution exists) settings.
1 code implementation • 10 Jun 2021 • Eduard Gorbunov, Marina Danilova, Innokentiy Shibaev, Pavel Dvurechensky, Alexander Gasnikov
In our paper, we resolve this issue and derive the first high-probability convergence results with logarithmic dependence on the confidence level for non-smooth convex stochastic optimization problems with non-sub-Gaussian (heavy-tailed) noise.
no code implementations • NeurIPS 2021 • Eduard Gorbunov, Marina Danilova, Innokentiy Andreevich Shibaev, Pavel Dvurechensky, Alexander Gasnikov
In our paper, we resolve this issue and derive the first high-probability convergence results with logarithmical dependence on the confidence level for non-smooth convex stochastic optimization problems with non-sub-Gaussian (heavy-tailed) noise.
no code implementations • 16 Feb 2021 • Pavel Dvurechensky, Dmitry Kamzolov, Aleksandr Lukashevich, Soomin Lee, Erik Ordentlich, César A. Uribe, Alexander Gasnikov
Statistical preconditioning enables fast methods for distributed large-scale empirical risk minimization problems.
Distributed Optimization Optimization and Control
no code implementations • 15 Feb 2021 • Alexander Rogozin, Alexander Beznosikov, Darina Dvinskikh, Dmitry Kovalev, Pavel Dvurechensky, Alexander Gasnikov
We consider distributed convex-concave saddle point problems over arbitrary connected undirected networks and propose a decentralized distributed algorithm for their solution.
Distributed Optimization Optimization and Control Distributed, Parallel, and Cluster Computing
no code implementations • 4 Jan 2021 • Pavel Dvurechensky, Mathias Staudigl, Shimrit Shtern
In this survey we cover a number of key developments in gradient-based optimization methods.
no code implementations • 31 Dec 2020 • Petr Ostroukhov, Rinat Kamalov, Pavel Dvurechensky, Alexander Gasnikov
The first method is based on the assumption of $p$-th order smoothness of the objective and it achieves a convergence rate of $O \left( \left( \frac{L_p R^{p - 1}}{\mu} \right)^\frac{2}{p + 1} \log \frac{\mu R^2}{\varepsilon_G} \right)$, where $R$ is an estimate of the initial distance to the solution, and $\varepsilon_G$ is the error in terms of duality gap.
Optimization and Control
no code implementations • 31 Dec 2020 • Artem Agafonov, Dmitry Kamzolov, Pavel Dvurechensky, Alexander Gasnikov
We propose general non-accelerated and accelerated tensor methods under inexact information on the derivatives of the objective, analyze their convergence rate.
Optimization and Control
no code implementations • 11 Dec 2020 • Marina Danilova, Pavel Dvurechensky, Alexander Gasnikov, Eduard Gorbunov, Sergey Guminov, Dmitry Kamzolov, Innokentiy Shibaev
For this setting, we first present known results for the convergence rates of deterministic first-order methods, which are then followed by a general theoretical analysis of optimal stochastic and randomized gradient schemes, and an overview of the stochastic first-order methods.
no code implementations • 21 Sep 2020 • Abdurakhmon Sadiev, Aleksandr Beznosikov, Pavel Dvurechensky, Alexander Gasnikov
In particular, our analysis shows that in the case when the feasible set is a direct product of two simplices, our convergence rate for the stochastic term is only by a $\log n$ factor worse than for the first-order methods.
no code implementations • 11 Jun 2020 • Daniil Tiapkin, Alexander Gasnikov, Pavel Dvurechensky
This leads to a complicated stochastic optimization problem where the objective is given as an expectation of a function given as a solution to a random optimization problem.
2 code implementations • 18 Apr 2020 • Darina Dvinskikh, Dmitry Kamzolov, Alexander Gasnikov, Pavel Dvurechensky, Dmitry Pasechnyk, Vladislav Matykhin, Alexei Chernov
We propose an accelerated meta-algorithm, which allows to obtain accelerated methods for convex unconstrained minimization in different settings.
Optimization and Control
1 code implementation • 11 Feb 2020 • Pavel Dvurechensky, Petr Ostroukhov, Kamil Safin, Shimrit Shtern, Mathias Staudigl
Projection-free optimization via different variants of the Frank-Wolfe (FW), a. k. a.
1 code implementation • 19 Nov 2019 • Aleksandr Ogaltsov, Darina Dvinskikh, Pavel Dvurechensky, Alexander Gasnikov, Vladimir Spokoiny
In this paper we propose several adaptive gradient methods for stochastic optimization.
Optimization and Control
no code implementations • 4 Nov 2019 • Pavel Dvurechensky, Mathias Staudigl, César A. Uribe
Many problems in statistical learning, imaging, and computer vision involve the optimization of a non-convex objective function with singularities at the boundary of the feasible set.
no code implementations • 9 Jun 2019 • Sergey Guminov, Pavel Dvurechensky, Nazarii Tupitsa, Alexander Gasnikov
In this paper we combine AM and Nesterov's acceleration to propose an accelerated alternating minimization algorithm.
no code implementations • 8 Mar 2018 • César A. Uribe, Darina Dvinskikh, Pavel Dvurechensky, Alexander Gasnikov, Angelia Nedić
We propose a new \cu{class-optimal} algorithm for the distributed computation of Wasserstein Barycenters over networks.
1 code implementation • 25 Feb 2018 • Eduard Gorbunov, Pavel Dvurechensky, Alexander Gasnikov
In the two-point feedback setting, i. e. when pairs of function values are available, we propose an accelerated derivative-free algorithm together with its complexity analysis.
Optimization and Control Computational Complexity
1 code implementation • ICML 2018 • Pavel Dvurechensky, Alexander Gasnikov, Alexey Kroshnin
We analyze two algorithms for approximating the general optimal transport (OT) distance between two discrete distributions of size $n$, up to accuracy $\varepsilon$.
Data Structures and Algorithms Optimization and Control