no code implementations • 21 Nov 2019 • Yunteng Luan, Hanyu Zhao, Zhi Yang, Yafei Dai
In this paper, we propose a general training framework named multi-self-distillation learning (MSD), which mining knowledge of different classifiers within the same network and increase every classifier accuracy.