no code implementations • 15 Jun 2022 • Clemens JS Schaefer, Siddharth Joshi, Shan Li, Raul Blazquez
Quantizing the parameters and operations to lower bit-precision offers substantial memory and energy savings for neural network inference, facilitating the use of DNNs on edge computing platforms.