no code implementations • 4 Mar 2023 • Kaixin Xu, Alina Hui Xiu Lee, Ziyuan Zhao, Zhe Wang, Min Wu, Weisi Lin
A popular track of network compression approach is Quantization aware Training (QAT), which accelerates the forward pass during the neural network training and inference.