Search Results for author: Chulu Xiang

Found 3 papers, 2 papers with code

AdaBatchGrad: Combining Adaptive Batch Size and Adaptive Step Size

no code implementations7 Feb 2024 Petr Ostroukhov, Aigerim Zhumabayeva, Chulu Xiang, Alexander Gasnikov, Martin Takáč, Dmitry Kamzolov

To substantiate the efficacy of our method, we experimentally show, how the introduction of adaptive step size and adaptive batch size gradually improves the performance of regular SGD.

SANIA: Polyak-type Optimization Framework Leads to Scale Invariant Stochastic Algorithms

1 code implementation28 Dec 2023 Farshed Abdukhakimov, Chulu Xiang, Dmitry Kamzolov, Robert Gower, Martin Takáč

Adaptive optimization methods are widely recognized as among the most popular approaches for training Deep Neural Networks (DNNs).

Stochastic Gradient Descent with Preconditioned Polyak Step-size

1 code implementation3 Oct 2023 Farshed Abdukhakimov, Chulu Xiang, Dmitry Kamzolov, Martin Takáč

Stochastic Gradient Descent (SGD) is one of the many iterative optimization methods that are widely used in solving machine learning problems.

Cannot find the paper you are looking for? You can Submit a new open access paper.