Search Results for author: Ivan Lazarevich

Found 7 papers, 4 papers with code

QGen: On the Ability to Generalize in Quantization Aware Training

no code implementations17 Apr 2024 MohammadHossein AskariHemmat, Ahmadreza Jeddi, Reyhane Askari Hemmat, Ivan Lazarevich, Alexander Hoffman, Sudhakar Sah, Ehsan Saboori, Yvon Savaria, Jean-Pierre David

In this work, we investigate the generalization properties of quantized neural networks, a characteristic that has received little attention despite its implications on model performance.

Quantization

Accelerating Deep Neural Networks via Semi-Structured Activation Sparsity

1 code implementation12 Sep 2023 Matteo Grimaldi, Darshan C. Ganji, Ivan Lazarevich, Sudhakar Sah

The demand for efficient processing of deep neural networks (DNNs) on embedded devices is a significant challenge limiting their deployment.

Image Classification object-detection +1

YOLOBench: Benchmarking Efficient Object Detectors on Embedded Systems

1 code implementation26 Jul 2023 Ivan Lazarevich, Matteo Grimaldi, Ravish Kumar, Saptarshi Mitra, Shahrukh Khan, Sudhakar Sah

We present YOLOBench, a benchmark comprised of 550+ YOLO-based object detection models on 4 different datasets and 4 different embedded hardware platforms (x86 CPU, ARM CPU, Nvidia GPU, NPU).

Benchmarking Neural Architecture Search +3

QReg: On Regularization Effects of Quantization

no code implementations24 Jun 2022 MohammadHossein AskariHemmat, Reyhane Askari Hemmat, Alex Hoffman, Ivan Lazarevich, Ehsan Saboori, Olivier Mastropietro, Yvon Savaria, Jean-Pierre David

To confirm our analytical study, we performed an extensive list of experiments summarized in this paper in which we show that the regularization effects of quantization can be seen in various vision tasks and models, over various datasets.

Quantization

Post-training deep neural network pruning via layer-wise calibration

no code implementations30 Apr 2021 Ivan Lazarevich, Alexander Kozlov, Nikita Malinin

We present a post-training weight pruning method for deep neural networks that achieves accuracy levels tolerable for the production setting and that is sufficiently fast to be run on commodity hardware such as desktop CPUs or edge devices.

Network Pruning

Neural Network Compression Framework for fast model inference

2 code implementations20 Feb 2020 Alexander Kozlov, Ivan Lazarevich, Vasily Shamporov, Nikolay Lyalyushkin, Yury Gorbachev

In this work we present a new framework for neural networks compression with fine-tuning, which we called Neural Network Compression Framework (NNCF).

Binarization Neural Network Compression +1

Spikebench: An open benchmark for spike train time-series classification

3 code implementations9 Oct 2018 Ivan Lazarevich, Ilya Prokin, Boris Gutkin, Victor Kazantsev

Modern well-performing approaches to neural decoding are based on machine learning models such as decision tree ensembles and deep neural networks.

Feature Engineering Time Series +2

Cannot find the paper you are looking for? You can Submit a new open access paper.