1 code implementation • 21 Jun 2024 • Sungbin Shin, Wonpyo Park, Jaeho Lee, Namhoon Lee
This work suggests fundamentally rethinking the current practice of pruning large language models (LLMs).
1 code implementation • 29 Nov 2023 • Sungbin Shin, Dongyeop Lee, Maksym Andriushchenko, Namhoon Lee
Training overparameterized neural networks often yields solutions with varying generalization capabilities, even when achieving similar training losses.
1 code implementation • 28 Feb 2023 • Sungbin Shin, Yohan Jo, Sungsoo Ahn, Namhoon Lee
Concept bottleneck models (CBMs) are a class of interpretable neural network models that predict the target response of a given input based on its high-level concepts.