no code implementations • 6 Feb 2025 • Zony Yu, Yuqiao Wen, Lili Mou
Knowledge distillation (KD) is a popular method of transferring knowledge from a large "teacher" model to a small "student" model.
Knowledge Distillation