signSGD: Compressed Optimisation for Non-Convex Problems

Training large neural networks requires distributing learning across multiple workers, where the cost of communicating gradients can be a significant bottleneck. signSGD alleviates this problem by transmitting just the sign of each minibatch stochastic gradient... (read more)

PDF Abstract ICML 2018 PDF ICML 2018 Abstract

Tasks


Datasets


Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper


METHOD TYPE
Adam
Stochastic Optimization
SGD
Stochastic Optimization