Search Results for author: Kaokao Lv

Found 3 papers, 3 papers with code

TEQ: Trainable Equivalent Transformation for Quantization of LLMs

1 code implementation17 Oct 2023 Wenhua Cheng, Yiyang Cai, Kaokao Lv, Haihao Shen

As large language models (LLMs) become more prevalent, there is a growing need for new and improved quantization methods that can meet the computationalast layer demands of these modern architectures while maintaining the accuracy.

Quantization

Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs

1 code implementation11 Sep 2023 Wenhua Cheng, Weiwei Zhang, Haihao Shen, Yiyang Cai, Xin He, Kaokao Lv

As the number of bits decreases, the quantization grid broadens, thus emphasizing the importance of up and down rounding.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.