no code implementations • 18 Mar 2021 • DongHyun Lee, Minkyoung Cho, Seungwon Lee, Joonho Song, Changkyu Choi
Post-training quantization is a representative technique for compressing neural networks, making them smaller and more efficient for deployment on edge devices.