no code implementations • 8 Nov 2022 • Parvin Nazari, Ahmad Mousavi, Davoud Ataee Tarzanagh, George Michailidis
A key feature of the proposed algorithm is to estimate the hyper-gradient of the penalty function via decentralized computation of matrix-vector products and few vector communications, which is then integrated within an alternating algorithm to obtain finite-time convergence analysis under different convexity assumptions.
no code implementations • 4 Sep 2022 • Parvin Nazari, Esmaile Khorram
Adaptive gradient algorithms such as ADAGRAD and its variants have gained popularity in the training of deep neural networks.
1 code implementation • 6 Jul 2022 • Davoud Ataee Tarzanagh, Parvin Nazari, BoJian Hou, Li Shen, Laura Balzano
This paper introduces \textit{online bilevel optimization} in which a sequence of time-varying bilevel problems is revealed one after the other.
no code implementations • 29 Sep 2021 • Parvin Nazari, Esmaile Khorram
The online meta-learning framework has arisen as a powerful tool for the continual lifelong learning setting.
no code implementations • 19 May 2020 • Parvin Nazari, Davoud Ataee Tarzanagh, George Michailidis
In this paper, we design and analyze a new family of adaptive subgradient methods for solving an important class of weakly convex (possibly nonsmooth) stochastic optimization problems.
1 code implementation • ICLR 2019 • Parvin Nazari, Davoud Ataee Tarzanagh, George Michailidis
Adaptive gradient-based optimization methods such as \textsc{Adagrad}, \textsc{Rmsprop}, and \textsc{Adam} are widely used in solving large-scale machine learning problems including deep learning.