no code implementations • 10 Mar 2021 • Sedigh Ghamari, Koray Ozcan, Thu Dinh, Andrey Melnikov, Juan Carvajal, Jan Ernst, Sek Chai
We propose a Quantization Guided Training (QGT) method to guide DNN training towards optimized low-bit-precision targets and reach extreme compression levels below 8-bit precision.
no code implementations • 4 Nov 2020 • Thu Dinh, Andrey Melnikov, Vasilios Daskalopoulos, Sek Chai
Quantization for deep neural networks (DNN) have enabled developers to deploy models with less memory and more efficient low-power inference.