Search Results for author: Zhixia Jiang

Found 2 papers, 2 papers with code

Scaling transition from momentum stochastic gradient descent to plain stochastic gradient descent

1 code implementation12 Jun 2021 Kun Zeng, Jinlan Liu, Zhixia Jiang, Dongpo Xu

The momentum stochastic gradient descent uses the accumulated gradient as the updated direction of the current parameters, which has a faster training speed.

A decreasing scaling transition scheme from Adam to SGD

2 code implementations12 Jun 2021 Kun Zeng, Jinlan Liu, Zhixia Jiang, Dongpo Xu

Adaptive gradient algorithm (AdaGrad) and its variants, such as RMSProp, Adam, AMSGrad, etc, have been widely used in deep learning.

Cannot find the paper you are looking for? You can Submit a new open access paper.