Search Results for author: Alberto Delmas Lascorz

Found 4 papers, 0 papers with code

APack: Off-Chip, Lossless Data Compression for Efficient Deep Learning Inference

no code implementations21 Jan 2022 Alberto Delmas Lascorz, Mostafa Mahmoud, Andreas Moshovos

When integrated with a Tensorcore-based accelerator, APack boosts the speedup and energy efficiency to 1. 44X and 1. 37X respectively.

Data Compression Quantization

BitPruning: Learning Bitlengths for Aggressive and Accurate Quantization

no code implementations8 Feb 2020 Miloš Nikolić, Ghouthi Boukli Hacene, Ciaran Bannon, Alberto Delmas Lascorz, Matthieu Courbariaux, Yoshua Bengio, Vincent Gripon, Andreas Moshovos

Neural networks have demonstrably achieved state-of-the art accuracy using low-bitlength integer quantization, yielding both execution time and energy benefits on existing hardware designs that support short bitlengths.

Quantization

Laconic Deep Learning Computing

no code implementations10 May 2018 Sayeh Sharify, Mostafa Mahmoud, Alberto Delmas Lascorz, Milos Nikolic, Andreas Moshovos

A Laconic configuration that uses a 1K-wire weight memory interface, outperforms the 2K-wire conventional accelerator by 15. 4x and is 1. 95x more energy efficient.

2k Image Classification

Loom: Exploiting Weight and Activation Precisions to Accelerate Convolutional Neural Networks

no code implementations23 Jun 2017 Sayeh Sharify, Alberto Delmas Lascorz, Kevin Siu, Patrick Judd, Andreas Moshovos

LM can trade-off accuracy for additional improvements in execution performance and energy efficiency and compares favorably to an accelerator that targeted only activation precisions.

Image Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.