1 code implementation • 28 Mar 2024 • Mohsen Gholami, Mohammad Akbari, Cindy Hu, Vaden Masrani, Z. Jane Wang, Yong Zhang
Knowledge distillation from LLMs is essential for the efficient deployment of language models.
Knowledge Distillation