no code implementations • 25 Oct 2022 • Jiongyu Guo, Defang Chen, Can Wang
Alignahead++ transfers structure and feature information in a student layer to the previous layer of another simultaneously trained student model in an alternating training procedure.
1 code implementation • 5 May 2022 • Jiongyu Guo, Defang Chen, Can Wang
Existing knowledge distillation methods on graph neural networks (GNNs) are almost offline, where the student model extracts knowledge from a powerful teacher model to improve its performance.