no code implementations • 24 Jun 2023 • Tianhong Huang, Victor Agostinelli, Lizhong Chen
Compactness in deep learning can be critical to a model's viability in low-resource applications, and a common approach to extreme model compression is quantization.