Search Results for author: Grigory Malinovsky

Found 18 papers, 4 papers with code

From Local SGD to Local Fixed Point Methods for Federated Learning

no code implementations ICML 2020 Grigory Malinovsky, Dmitry Kovalev, Elnur Gasanov, Laurent Condat, Peter Richtarik

Most algorithms for solving optimization problems or finding saddle points of convex-concave functions are fixed point algorithms.

Federated Learning

Streamlining in the Riemannian Realm: Efficient Riemannian Optimization with Loopless Variance Reduction

no code implementations11 Mar 2024 Yury Demidovich, Grigory Malinovsky, Peter Richtárik

These methods replace the outer loop with probabilistic gradient computation triggered by a coin flip in each iteration, ensuring simpler proofs, efficient hyperparameter selection, and sharp convergence guarantees.

Distributed Optimization Riemannian optimization

MAST: Model-Agnostic Sparsified Training

1 code implementation27 Nov 2023 Yury Demidovich, Grigory Malinovsky, Egor Shulgin, Peter Richtárik

We introduce a novel optimization problem formulation that departs from the conventional way of minimizing machine learning model loss as a black-box function.

TAMUNA: Doubly Accelerated Distributed Optimization with Local Training, Compression, and Partial Participation

1 code implementation20 Feb 2023 Laurent Condat, Ivan Agarský, Grigory Malinovsky, Peter Richtárik

We propose TAMUNA, the first algorithm for distributed optimization that leveraged the two strategies of local training and compression jointly and allows for partial participation.

Distributed Optimization Federated Learning

Federated Learning with Regularized Client Participation

no code implementations7 Feb 2023 Grigory Malinovsky, Samuel Horváth, Konstantin Burlachenko, Peter Richtárik

Under this scheme, each client joins the learning process every $R$ communication rounds, which we refer to as a meta epoch.

Federated Learning

Can 5th Generation Local Training Methods Support Client Sampling? Yes!

no code implementations29 Dec 2022 Michał Grudzień, Grigory Malinovsky, Peter Richtárik

The celebrated FedAvg algorithm of McMahan et al. (2017) is based on three components: client sampling (CS), data sampling (DS) and local training (LT).

An Optimal Algorithm for Strongly Convex Min-min Optimization

no code implementations29 Dec 2022 Alexander Gasnikov, Dmitry Kovalev, Grigory Malinovsky

In this paper we study the smooth strongly convex minimization problem $\min_{x}\min_y f(x, y)$.

Minibatch Stochastic Three Points Method for Unconstrained Smooth Minimization

no code implementations16 Sep 2022 Soumia Boucherouite, Grigory Malinovsky, Peter Richtárik, El Houcine Bergou

In this paper, we propose a new zero order optimization method called minibatch stochastic three points (MiSTP) method to solve an unconstrained minimization problem in a setting where only an approximation of the objective function evaluation is possible.

Variance Reduced ProxSkip: Algorithm, Theory and Application to Federated Learning

1 code implementation9 Jul 2022 Grigory Malinovsky, Kai Yi, Peter Richtárik

We study distributed optimization methods based on the {\em local training (LT)} paradigm: achieving communication efficiency by performing richer local gradient-based training on the clients before parameter averaging.

Distributed Optimization Federated Learning

Federated Optimization Algorithms with Random Reshuffling and Gradient Compression

1 code implementation14 Jun 2022 Abdurakhmon Sadiev, Grigory Malinovsky, Eduard Gorbunov, Igor Sokolov, Ahmed Khaled, Konstantin Burlachenko, Peter Richtárik

To reveal the true advantages of RR in the distributed learning with compression, we propose a new method called DIANA-RR that reduces the compression variance and has provably better convergence rates than existing counterparts with with-replacement sampling of stochastic gradients.

Federated Learning Quantization

Federated Random Reshuffling with Compression and Variance Reduction

no code implementations8 May 2022 Grigory Malinovsky, Peter Richtárik

Random Reshuffling (RR), which is a variant of Stochastic Gradient Descent (SGD) employing sampling without replacement, is an immensely popular method for training supervised machine learning models via empirical risk minimization.

BIG-bench Machine Learning Federated Learning

ProxSkip: Yes! Local Gradient Steps Provably Lead to Communication Acceleration! Finally!

no code implementations18 Feb 2022 Konstantin Mishchenko, Grigory Malinovsky, Sebastian Stich, Peter Richtárik

The canonical approach to solving such problems is via the proximal gradient descent (ProxGD) algorithm, which is based on the evaluation of the gradient of $f$ and the prox operator of $\psi$ in each iteration.

Federated Learning

Server-Side Stepsizes and Sampling Without Replacement Provably Help in Federated Optimization

no code implementations26 Jan 2022 Grigory Malinovsky, Konstantin Mishchenko, Peter Richtárik

Together, our results on the advantage of large and small server-side stepsizes give a formal justification for the practice of adaptive server-side optimization in federated learning.

Federated Learning

Random Reshuffling with Variance Reduction: New Analysis and Better Rates

no code implementations19 Apr 2021 Grigory Malinovsky, Alibek Sailanbayev, Peter Richtárik

One of the tricks that works so well in practice that it is used as default in virtually all widely used machine learning software is {\em random reshuffling (RR)}.

BIG-bench Machine Learning

Distributed Proximal Splitting Algorithms with Rates and Acceleration

no code implementations2 Oct 2020 Laurent Condat, Grigory Malinovsky, Peter Richtárik

We analyze several generic proximal splitting algorithms well suited for large-scale convex nonsmooth optimization.

From Local SGD to Local Fixed-Point Methods for Federated Learning

no code implementations3 Apr 2020 Grigory Malinovsky, Dmitry Kovalev, Elnur Gasanov, Laurent Condat, Peter Richtárik

Most algorithms for solving optimization problems or finding saddle points of convex-concave functions are fixed-point algorithms.

Federated Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.