Search Results for author: Jingchang Liu

Found 5 papers, 1 papers with code

STL-SGD: Speeding Up Local SGD with Stagewise Communication Period

no code implementations11 Jun 2020 Shuheng Shen, Yifei Cheng, Jingchang Liu, Linli Xu

Distributed parallel stochastic gradient descent algorithms are workhorses for large scale machine learning tasks.

Variance Reduced Local SGD with Lower Communication Complexity

1 code implementation30 Dec 2019 Xianfeng Liang, Shuheng Shen, Jingchang Liu, Zhen Pan, Enhong Chen, Yifei Cheng

To accelerate the training of machine learning models, distributed stochastic gradient descent (SGD) and its variants have been widely adopted, which apply multiple workers in parallel to speed up training.

BIG-bench Machine Learning

Faster Distributed Deep Net Training: Computation and Communication Decoupled Stochastic Gradient Descent

no code implementations28 Jun 2019 Shuheng Shen, Linli Xu, Jingchang Liu, Xianfeng Liang, Yifei Cheng

Nevertheless, although distributed stochastic gradient descent (SGD) algorithms can achieve a linear iteration speedup, they are limited significantly in practice by the communication cost, making it difficult to achieve a linear time speedup.

Asynchronous Stochastic Composition Optimization with Variance Reduction

no code implementations15 Nov 2018 Shuheng Shen, Linli Xu, Jingchang Liu, Junliang Guo, Qing Ling

Composition optimization has drawn a lot of attention in a wide variety of machine learning domains from risk management to reinforcement learning.

Management Reinforcement Learning

Accelerating Stochastic Gradient Descent Using Antithetic Sampling

no code implementations7 Oct 2018 Jingchang Liu, Linli Xu

(Mini-batch) Stochastic Gradient Descent is a popular optimization method which has been applied to many machine learning applications.

Binary Classification General Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.