1 code implementation • 1 Aug 2018 • Ini Oguntola, Subby Olubeko, Christopher Sweeney
We show that by combining pruning and knowledge distillation methods we can create a compressed network 85 times smaller than the original, all while retaining 96% of the original model's accuracy.