Neural Network Compression

74 papers with code • 1 benchmarks • 1 datasets

This task has no description! Would you like to contribute one?

Libraries

Use these libraries to find Neural Network Compression models and implementations

Datasets


Towards Meta-Pruning via Optimal Transport

alexandertheus/intra-fusion 12 Feb 2024

Structural pruning of neural networks conventionally relies on identifying and discarding less important neurons, a practice often resulting in significant accuracy loss that necessitates subsequent fine-tuning efforts.

7
12 Feb 2024

Causal-DFQ: Causality Guided Data-free Network Quantization

42shawn/causal-dfq ICCV 2023

Inspired by the causal understanding, we propose the Causality-guided Data-free Network Quantization method, Causal-DFQ, to eliminate the reliance on data via approaching an equilibrium of causality-driven intervened distributions.

7
24 Sep 2023

A Survey on Deep Neural Network Pruning-Taxonomy, Comparison, Analysis, and Recommendations

hrcheng1066/awesome-pruning 13 Aug 2023

Modern deep neural networks, particularly recent large language models, come with massive model sizes that require significant computational and storage resources.

92
13 Aug 2023

Lightweight Attribute Localizing Models for Pedestrian Attribute Recognition

ashishjv1/LW-ALM 16 Jun 2023

Pedestrian Attribute Recognition (PAR) deals with the problem of identifying features in a pedestrian image.

2
16 Jun 2023

Implicit Compressibility of Overparametrized Neural Networks Trained with Heavy-Tailed SGD

mbarsbey/implicit-compressibility 13 Jun 2023

Neural network compression has been an increasingly important subject, not only due to its practical relevance, but also due to its theoretical implications, as there is an explicit connection between compressibility and generalization error.

0
13 Jun 2023

Variation Spaces for Multi-Output Neural Networks: Insights on Multi-Task Learning and Network Compression

joeshenouda/vv-spaces-nn-width 25 May 2023

This representer theorem establishes that shallow vector-valued neural networks are the solutions to data-fitting problems over these infinite-dimensional spaces, where the network widths are bounded by the square of the number of training data.

3
25 May 2023

SwiftTron: An Efficient Hardware Accelerator for Quantized Transformers

albertomarchisio/swifttron 8 Apr 2023

In particular, fixed-point quantization is desirable to ease the computations using lightweight blocks, like adders and multipliers, of the underlying hardware.

22
08 Apr 2023

WHC: Weighted Hybrid Criterion for Filter Pruning on Convolutional Neural Networks

ShaowuChen/WHC 16 Feb 2023

Filter pruning has attracted increasing attention in recent years for its capacity in compressing and accelerating convolutional neural networks.

6
16 Feb 2023

DepGraph: Towards Any Structural Pruning

VainF/Torch-Pruning CVPR 2023

Structural pruning enables model acceleration by removing structurally-grouped parameters from neural networks.

2,296
30 Jan 2023

Magnitude and Similarity based Variable Rate Filter Pruning for Efficient Convolution Neural Networks

ghimiredhikura/msvsp-filter-pruning Applied Sciences 2022

We studied several filter selection criteria based on filter magnitude and similarity among filters within a convolution layer, and based on the assumption that the sensitivity of each layer throughout the network is different, unlike conventional fixed rate pruning methods, our algorithm using loss-aware filter selection criteria automatically finds the suitable pruning rate for each layer throughout the network.

5
27 Dec 2022