no code implementations • 21 Nov 2019 • Yunteng Luan, Hanyu Zhao, Zhi Yang, Yafei Dai
In this paper, we propose a general training framework named multi-self-distillation learning (MSD), which mining knowledge of different classifiers within the same network and increase every classifier accuracy.
no code implementations • 19 Oct 2018 • Lingxiao Ma, Zhi Yang, Youshan Miao, Jilong Xue, Ming Wu, Lidong Zhou, Yafei Dai
This evolution has led to large graph-based irregular and sparse models that go beyond what existing deep learning frameworks are designed for.