Search Results for author: Jeffrey L. McKinstry

Found 5 papers, 1 papers with code

Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference

no code implementations30 Jan 2023 Deepika Bablani, Jeffrey L. McKinstry, Steven K. Esser, Rathinakumar Appuswamy, Dharmendra S. Modha

Using EAGL and ALPS for layer precision selection, full-precision accuracy is recovered with a mix of 4-bit and 2-bit layers for ResNet-50, ResNet-101 and BERT-base transformer networks, demonstrating enhanced performance across the entire accuracy-throughput frontier.

Efficient Neural Network Quantization

Learned Step Size Quantization

8 code implementations ICLR 2020 Steven K. Esser, Jeffrey L. McKinstry, Deepika Bablani, Rathinakumar Appuswamy, Dharmendra S. Modha

Deep networks run with low precision operations at inference time offer power and space advantages over high precision alternatives, but need to overcome the challenge of maintaining high accuracy as precision decreases.

Model Compression Quantization

Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference

no code implementations ICLR 2019 Jeffrey L. McKinstry, Steven K. Esser, Rathinakumar Appuswamy, Deepika Bablani, John V. Arthur, Izzet B. Yildiz, Dharmendra S. Modha

Therefore, we (a) reduce solution distance by starting with pretrained fp32 precision baseline networks and fine-tuning, and (b) combat gradient noise introduced by quantization by training longer and reducing learning rates.

General Classification Quantization

Cannot find the paper you are looking for? You can Submit a new open access paper.