Search Results for author: Ian Colbert

Found 11 papers, 2 papers with code

A2Q+: Improving Accumulator-Aware Weight Quantization

no code implementations19 Jan 2024 Ian Colbert, Alessandro Pappalardo, Jakoba Petri-Koenig, Yaman Umuroglu

Recent studies show that also reducing the precision of the accumulator can further improve hardware efficiency at the risk of numerical overflow, which introduces arithmetic errors that can degrade model accuracy.

Quantization

A2Q: Accumulator-Aware Quantization with Guaranteed Overflow Avoidance

no code implementations ICCV 2023 Ian Colbert, Alessandro Pappalardo, Jakoba Petri-Koenig

We apply our method to deep learning-based computer vision tasks to show that A2Q can train QNNs for low-precision accumulators while maintaining model accuracy competitive with a floating-point baseline.

Quantization

Quantized Neural Networks for Low-Precision Accumulation with Guaranteed Overflow Avoidance

no code implementations31 Jan 2023 Ian Colbert, Alessandro Pappalardo, Jakoba Petri-Koenig

Across all of our benchmark models trained with 8-bit weights and activations, we observe that constraining the hidden layers of quantized neural networks to fit into 16-bit accumulators yields an average 98. 2% sparsity with an estimated compression rate of 46. 5x all while maintaining 99. 2% of the floating-point performance.

Quantization

Robust Transferable Feature Extractors: Learning to Defend Pre-Trained Networks Against White Box Adversaries

no code implementations14 Sep 2022 Alexander Cann, Ian Colbert, Ihab Amer

The widespread adoption of deep neural networks in computer vision applications has brought forth a significant interest in adversarial robustness.

Adversarial Robustness

Human-Like Navigation Behavior: A Statistical Evaluation Framework

no code implementations10 Mar 2022 Ian Colbert, Mehdi Saeedi

Recent advancements in deep reinforcement learning have brought forth an impressive display of highly skilled artificial agents capable of complex intelligent behavior.

Training Deep Neural Networks with Joint Quantization and Pruning of Weights and Activations

2 code implementations15 Oct 2021 Xinyu Zhang, Ian Colbert, Ken Kreutz-Delgado, Srinjoy Das

State-of-the-art quantization techniques are currently applied to both the weights and activations; however, pruning is most often applied to only the weights of the network.

Network Pruning Quantization

An Energy-Efficient Edge Computing Paradigm for Convolution-based Image Upsampling

1 code implementation15 Jul 2021 Ian Colbert, Ken Kreutz-Delgado, Srinjoy Das

We analyze and compare the inference properties of convolution-based upsampling algorithms using a quantitative model of incurred time and energy costs and show that using deconvolution for inference at the edge improves both system latency and energy efficiency when compared to their sub-pixel or resize convolution counterparts.

Edge-computing

Generative and Discriminative Deep Belief Network Classifiers: Comparisons Under an Approximate Computing Framework

no code implementations31 Jan 2021 Siqiao Ruan, Ian Colbert, Ken Kreutz-Delgado, Srinjoy Das

The use of Deep Learning hardware algorithms for embedded applications is characterized by challenges such as constraints on device power consumption, availability of labeled data, and limited internet bandwidth for frequent training on cloud servers.

PT-MMD: A Novel Statistical Framework for the Evaluation of Generative Systems

no code implementations28 Oct 2019 Alexander Potapov, Ian Colbert, Ken Kreutz-Delgado, Alexander Cloninger, Srinjoy Das

Stochastic-sampling-based Generative Neural Networks, such as Restricted Boltzmann Machines and Generative Adversarial Networks, are now used for applications such as denoising, image occlusion removal, pattern completion, and motion synthesis.

Denoising Model Selection +1

AX-DBN: An Approximate Computing Framework for the Design of Low-Power Discriminative Deep Belief Networks

no code implementations11 Mar 2019 Ian Colbert, Ken Kreutz-Delgado, Srinjoy Das

The power budget for embedded hardware implementations of Deep Learning algorithms can be extremely tight.

Cannot find the paper you are looking for? You can Submit a new open access paper.