no code implementations • 28 Jun 2024 • Wonbeom Lee, Jungi Lee, Junghwan Seo, Jaewoong Sim
Transformer-based large language models (LLMs) demonstrate impressive performance across various natural language processing tasks.
no code implementations • 16 Jun 2024 • Jungi Lee, Wonbeom Lee, Jaewoong Sim
Large language models (LLMs) demonstrate outstanding performance in various tasks in machine learning and have thus become one of the most important workloads in today's computing landscape.
no code implementations • 29 May 2024 • Taehyun Kim, Kwanseok Choi, Youngmock Cho, Jaehoon Cho, Hyuk-Jae Lee, Jaewoong Sim
Mixture-of-Experts (MoE) large language models (LLM) have memory requirements that often exceed the GPU memory capacity, requiring costly parameter movement from secondary memories to the GPU for expert computation.