Search Results for author: Rathinakumar Appuswamy

Found 7 papers, 1 papers with code

Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference

no code implementations30 Jan 2023 Deepika Bablani, Jeffrey L. McKinstry, Steven K. Esser, Rathinakumar Appuswamy, Dharmendra S. Modha

Using EAGL and ALPS for layer precision selection, full-precision accuracy is recovered with a mix of 4-bit and 2-bit layers for ResNet-50, ResNet-101 and BERT-base transformer networks, demonstrating enhanced performance across the entire accuracy-throughput frontier.

Efficient Neural Network Quantization

Learned Step Size Quantization

8 code implementations ICLR 2020 Steven K. Esser, Jeffrey L. McKinstry, Deepika Bablani, Rathinakumar Appuswamy, Dharmendra S. Modha

Deep networks run with low precision operations at inference time offer power and space advantages over high precision alternatives, but need to overcome the challenge of maintaining high accuracy as precision decreases.

Model Compression Quantization

Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference

no code implementations ICLR 2019 Jeffrey L. McKinstry, Steven K. Esser, Rathinakumar Appuswamy, Deepika Bablani, John V. Arthur, Izzet B. Yildiz, Dharmendra S. Modha

Therefore, we (a) reduce solution distance by starting with pretrained fp32 precision baseline networks and fine-tuning, and (b) combat gradient noise introduced by quantization by training longer and reducing learning rates.

General Classification Quantization

Structured Convolution Matrices for Energy-efficient Deep learning

no code implementations8 Jun 2016 Rathinakumar Appuswamy, Tapan Nayak, John Arthur, Steven Esser, Paul Merolla, Jeffrey Mckinstry, Timothy Melano, Myron Flickner, Dharmendra Modha

We derive a relationship between network representation in energy-efficient neuromorphic architectures and block Toplitz convolutional matrices.

General Classification

Deep neural networks are robust to weight binarization and other non-linear distortions

no code implementations7 Jun 2016 Paul Merolla, Rathinakumar Appuswamy, John Arthur, Steve K. Esser, Dharmendra Modha

Recent results show that deep neural networks achieve excellent performance even when, during training, weights are quantized and projected to a binary representation.

Binarization Data Augmentation +1

Backpropagation for Energy-Efficient Neuromorphic Computing

no code implementations NeurIPS 2015 Steve K. Esser, Rathinakumar Appuswamy, Paul Merolla, John V. Arthur, Dharmendra S. Modha

Solving real world problems with embedded neural networks requires both training algorithms that achieve high performance and compatible hardware that runs in real time while remaining energy efficient.

Cannot find the paper you are looking for? You can Submit a new open access paper.