Search Results for author: Fangmin Liu

Found 4 papers, 0 papers with code

A General Error-Theoretical Analysis Framework for Constructing Compression Strategies

no code implementations19 Feb 2025 Boyang Zhang, Daning Cheng, Yunquan Zhang, Meiqi Tu, Fangmin Liu, Jiake Tian

The exponential growth in parameter size and computational complexity of deep models poses significant challenges for efficient deployment.

Quantization

Compression for Better: A General and Stable Lossless Compression Framework

no code implementations9 Dec 2024 Boyang Zhang, Daning Cheng, Yunquan Zhang, Fangmin Liu, WenGuang Chen

A key challenge is effectively leveraging compression errors and defining the boundaries for lossless compression to minimize model loss.

Computational Efficiency Model Compression +1

Lossless Model Compression via Joint Low-Rank Factorization Optimization

no code implementations9 Dec 2024 Boyang Zhang, Daning Cheng, Yunquan Zhang, Fangmin Liu, Jiake Tian

Low-rank factorization is a popular model compression technique that minimizes the error $\delta$ between approximated and original weight matrices.

Model Compression Model Optimization

FP=xINT:A Low-Bit Series Expansion Algorithm for Post-Training Quantization

no code implementations9 Dec 2024 Boyang Zhang, Daning Cheng, Yunquan Zhang, Fangmin Liu

We introduce a deep model series expansion framework to address this issue, enabling rapid and accurate approximation of unquantized models without calibration sets or fine-tuning.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.