Accelerating Deep Neural Network Training with Inconsistent Stochastic Gradient Descent

17 Mar 2016Linnan WangYi YangMartin Renqiang MinSrimat Chakradhar

SGD is the widely adopted method to train CNN. Conceptually it approximates the population with a randomly sampled batch; then it evenly trains batches by conducting a gradient update on every batch in an epoch... (read more)

PDF Abstract


No code implementations yet. Submit your code now


Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods used in the Paper