no code implementations • 11 Jun 2020 • Shuheng Shen, Yifei Cheng, Jingchang Liu, Linli Xu
Distributed parallel stochastic gradient descent algorithms are workhorses for large scale machine learning tasks.
1 code implementation • 30 Dec 2019 • Xianfeng Liang, Shuheng Shen, Jingchang Liu, Zhen Pan, Enhong Chen, Yifei Cheng
To accelerate the training of machine learning models, distributed stochastic gradient descent (SGD) and its variants have been widely adopted, which apply multiple workers in parallel to speed up training.
no code implementations • 28 Jun 2019 • Shuheng Shen, Linli Xu, Jingchang Liu, Xianfeng Liang, Yifei Cheng
Nevertheless, although distributed stochastic gradient descent (SGD) algorithms can achieve a linear iteration speedup, they are limited significantly in practice by the communication cost, making it difficult to achieve a linear time speedup.
no code implementations • 15 Nov 2018 • Shuheng Shen, Linli Xu, Jingchang Liu, Junliang Guo, Qing Ling
Composition optimization has drawn a lot of attention in a wide variety of machine learning domains from risk management to reinforcement learning.
no code implementations • 7 Oct 2018 • Jingchang Liu, Linli Xu
(Mini-batch) Stochastic Gradient Descent is a popular optimization method which has been applied to many machine learning applications.