no code implementations • 1 Jan 2021 • Se Jung Kwon, Dongsoo Lee, Yongkweon Jeon, Byeongwook Kim, Bae Seong Park, Yeonju Ro
As a practical model compression technique, parameter quantization is effective especially for language models associated with a large memory footprint.