no code implementations • 3 Dec 2018 • Minghan Li, Tanli Zuo, Ruicheng Li, Martha White, Wei-Shi Zheng
Knowledge distillation is an effective technique that transfers knowledge from a large teacher model to a shallow student.
Knowledge Distillation Machine Translation +2