no code implementations • 20 Feb 2024 • Gyeongman Kim, Doohyuk Jang, Eunho Yang
Recent advancements in large language models (LLMs) have raised concerns about inference costs, increasing the need for research into model compression.
Instruction Following Knowledge Distillation +1