no code implementations • 4 Sep 2023 • Nilesh Prasad Pandey, Marios Fournarakis, Chirag Patel, Markus Nagel
Post-training quantization (PTQ) is the go-to compression technique for large generative models, such as stable diffusion or large language models.
no code implementations • 10 Jul 2023 • Jorn Peters, Marios Fournarakis, Markus Nagel, Mart van Baalen, Tijmen Blankevoort
By combining fast-to-compute sensitivities with efficient solvers during QAT, QBitOpt can produce mixed-precision networks with high task performance guaranteed to satisfy strict resource constraints.
no code implementations • 22 Jun 2022 • Kartik Gupta, Marios Fournarakis, Matthias Reisser, Christos Louizos, Markus Nagel
We perform extensive experiments on standard FL benchmarks to evaluate our proposed FedAvg variants for quantization robustness and provide a convergence analysis for our Quantization-Aware variants in FL.
1 code implementation • 21 Mar 2022 • Markus Nagel, Marios Fournarakis, Yelysei Bondarenko, Tijmen Blankevoort
These effects are particularly pronounced in low-bit ($\leq$ 4-bits) quantization of efficient networks with depth-wise separable layers, such as MobileNets and EfficientNets.
no code implementations • 20 Jan 2022 • Sangeetha Siddegowda, Marios Fournarakis, Markus Nagel, Tijmen Blankevoort, Chirag Patel, Abhijit Khobare
chapter 4) and quantization-aware training (QAT, cf.
no code implementations • 15 Jun 2021 • Markus Nagel, Marios Fournarakis, Rana Ali Amjad, Yelysei Bondarenko, Mart van Baalen, Tijmen Blankevoort
Neural network quantization is one of the most effective ways of achieving these savings but the additional noise it induces can lead to accuracy degradation.
no code implementations • 10 May 2021 • Marios Fournarakis, Markus Nagel
Quantization techniques applied to the inference of deep neural networks have enabled fast and efficient execution on resource-constraint devices.