Search Results for author: Daning Cheng

Found 2 papers, 0 papers with code

Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss

no code implementations20 Jul 2022 Daning Cheng, WenGuang Chen

Based on the model's resilience to computational noise, model quantization is important for compressing models and improving computing speed.

Quantization

Quantization in Layer's Input is Matter

no code implementations10 Feb 2022 Daning Cheng, WenGuang Chen

In this paper, we will show that the quantization in layer's input is more important than parameters' quantization for loss function.

Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.