Progressive Skeletonization: Trimming more fat from a network at initialization

16 Jun 2020Pau de JorgeAmartya SanyalHarkirat S. BehlPhilip H. S. TorrGregory RogezPuneet K. Dokania

Recent studies have shown that skeletonization (pruning parameters) of networks at initialization provides all the practical benefits of sparsity both at inference and training time, while only marginally degrading their performance. However, we observe that beyond a certain level of sparsity (approx 95%), these approaches fail to preserve the network performance, and to our surprise, in many cases perform even worse than trivial random pruning... (read more)

PDF Abstract


No code implementations yet. Submit your code now


Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.