Search Results for author: Dmitry Pasechnyuk

Found 4 papers, 2 papers with code

Judging Adam: Studying the Performance of Optimization Methods on ML4SE Tasks

no code implementations6 Mar 2023 Dmitry Pasechnyuk, Anton Prazdnichnykh, Mikhail Evtikhiev, Timofey Bryksin

In this work, we test the performance of various optimizers on deep learning models for source code and find that the choice of an optimizer can have a significant impact on the model quality, with up to two-fold score differences between some of the relatively well-performing optimizers.

On the Computational Efficiency of Catalyst Accelerated Coordinate Descent

no code implementations11 Mar 2021 Dmitry Pasechnyuk, Vladislav Matyukhin

This article is devoted to one particular case of using universal accelerated proximal envelopes to obtain computationally efficient accelerated versions of methods used to solve various optimization problem setups.

Optimization and Control 90C25 (Primary), 65K05 (Secondary) G.1.6

Adaptive Catalyst for Smooth Convex Optimization

1 code implementation25 Nov 2019 Anastasiya Ivanova, Dmitry Pasechnyuk, Dmitry Grishchenko, Egor Shulgin, Alexander Gasnikov, Vladislav Matyukhin

In this paper, we present a generic framework that allows accelerating almost arbitrary non-accelerated deterministic and randomized algorithms for smooth convex optimization problems.

Optimization and Control

Cannot find the paper you are looking for? You can Submit a new open access paper.