Improving Neural Network Quantization without Retraining using Outlier Channel Splitting

28 Jan 2019Ritchie ZhaoYuwei HuJordan DotzelChristopher De SaZhiru Zhang

Quantization can improve the execution latency and energy efficiency of neural networks on both commodity GPUs and specialized accelerators. The majority of existing literature focuses on training quantized DNNs, while this work examines the less-studied topic of quantizing a floating-point model without (re)training... (read more)

PDF Abstract

Evaluation Results from the Paper

  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.