Network Pruning

213 papers with code • 5 benchmarks • 5 datasets

Network Pruning is a popular approach to reduce a heavy network to obtain a light-weight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on come criterions, and finally fine-tuned to achieve comparable performance with reduced parameters.

Source: Ensemble Knowledge Distillation for Learning Improved and Efficient Networks

Libraries

Use these libraries to find Network Pruning models and implementations

Do Localization Methods Actually Localize Memorized Data in LLMs? A Tale of Two Benchmarks

terarachang/mempi 15 Nov 2023

On the other hand, even successful methods identify neurons that are not specific to a single memorized sequence.

1
15 Nov 2023

Beyond Size: How Gradients Shape Pruning Decisions in Large Language Models

rocktimjyotidas/gblm-pruner 8 Nov 2023

GBLM-Pruner leverages the first-order term of the Taylor expansion, operating in a training-free manner by harnessing properly normalized gradients from a few calibration samples to determine the pruning metric, and substantially outperforms competitive counterparts like SparseGPT and Wanda in multiple benchmarks.

26
08 Nov 2023

Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLMs

zyxxmu/dsnot 13 Oct 2023

Inspired by the Dynamic Sparse Training, DSnoT minimizes the reconstruction error between the dense and sparse LLMs, in the fashion of performing iterative weight pruning-and-growing on top of sparse LLMs.

26
13 Oct 2023

Filter Pruning For CNN With Enhanced Linear Representation Redundancy

bojue-wang/ccm-lrr 10 Oct 2023

In this paper, we propose a new structured pruning method.

1
10 Oct 2023

Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity

luuyin/owl 8 Oct 2023

Large Language Models (LLMs), renowned for their remarkable performance across diverse domains, present a challenge when it comes to practical deployment due to their colossal model size.

39
08 Oct 2023

SWAP: Sparse Entropic Wasserstein Regression for Robust Network Pruning

youlei202/entropic-wasserstein-pruning 7 Oct 2023

This study addresses the challenge of inaccurate gradients in computing the empirical Fisher Information Matrix during neural network pruning.

0
07 Oct 2023

Feather: An Elegant Solution to Effective DNN Sparsification

athglentis/feather 3 Oct 2023

Neural Network pruning is an increasingly popular way for producing compact and efficient models, suitable for resource-limited environments, while preserving high performance.

7
03 Oct 2023

EDAC: Efficient Deployment of Audio Classification Models For COVID-19 Detection

edac-ml4h/edac-ml4h 11 Sep 2023

Various researchers made use of machine learning methods in an attempt to detect COVID-19.

0
11 Sep 2023

A Survey on Deep Neural Network Pruning-Taxonomy, Comparison, Analysis, and Recommendations

hrcheng1066/awesome-pruning 13 Aug 2023

Modern deep neural networks, particularly recent large language models, come with massive model sizes that require significant computational and storage resources.

95
13 Aug 2023

Distilled Pruning: Using Synthetic Data to Win the Lottery

luke-mcdermott-mi/distilled-pruning 7 Jul 2023

This work introduces a novel approach to pruning deep learning models by using distilled data.

0
07 Jul 2023