International Conference on Machine Learning 2019

Trading Redundancy for Communication: Speeding up Distributed SGD for Non-convex Optimization

International Conference on Machine Learning 2019 mmkamani7/RI-SGD

Communication overhead is one of the key challenges that hinder the scalability of distributed optimization algorithms to train large neural networks.

DISTRIBUTED OPTIMIZATION