Browse > Methodology > Model Compression

Model Compression

21 papers with code · Methodology

State-of-the-art leaderboards

You can find evaluation results in the subtasks. You can also submitting evaluation metrics for this task.

Greatest papers with code

SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size

24 Feb 2016pytorch/vision

(2) Smaller DNNs require less bandwidth to export a new model from the cloud to an autonomous car.

MODEL COMPRESSION

The State of Sparsity in Deep Neural Networks

25 Feb 2019google-research/google-research

We rigorously evaluate three state-of-the-art techniques for inducing sparsity in deep neural networks on two large-scale learning tasks: Transformer trained on WMT 2014 English-to-German, and ResNet-50 trained on ImageNet.

MODEL COMPRESSION

Model compression via distillation and quantization

ICLR 2018 NervanaSystems/distiller

Deep neural networks (DNNs) continue to make significant advances, solving tasks from image classification to translation or reinforcement learning.

MODEL COMPRESSION QUANTIZATION

AMC: AutoML for Model Compression and Acceleration on Mobile Devices

ECCV 2018 NervanaSystems/distiller

Model compression is a critical technique to efficiently deploy neural network models on mobile devices which have limited computation resources and tight power budgets.

AUTOML MODEL COMPRESSION

Discrimination-aware Channel Pruning for Deep Neural Networks

NeurIPS 2018 SCUT-AILab/DCP

Channel pruning is one of the predominant approaches for deep model compression.

MODEL COMPRESSION

MicroExpNet: An Extremely Small and Fast Model For Expression Recognition From Frontal Face Images

19 Nov 2017cuguilke/microexpnet

This paper is aimed at creating extremely small and fast convolutional neural networks (CNN) for the problem of facial expression recognition (FER) from frontal face images.

FACIAL EXPRESSION RECOGNITION MODEL COMPRESSION

Learning Intrinsic Sparse Structures within Long Short-Term Memory

ICLR 2018 wenwei202/iss-rnns

This work aims to learn structurally-sparse Long Short-Term Memory (LSTM) by reducing the sizes of basic structures within LSTM units, including input updates, gates, hidden states, cell states and outputs.

LANGUAGE MODELLING MODEL COMPRESSION QUESTION ANSWERING

Ternary Weight Networks

16 May 2016fengfu-chris/caffe-twns

We introduce ternary weight networks (TWNs) - neural networks with weights constrained to +1, 0 and -1.

MODEL COMPRESSION

Dynamic Channel Pruning: Feature Boosting and Suppression

ICLR 2019 deep-fry/mayo

Making deep convolutional neural networks more accurate typically comes at the cost of increased computational and memory resources.

MODEL COMPRESSION

On-Device Neural Language Model Based Word Prediction

COLING 2018 meinwerk/WordPrediction

Recent developments in deep learning with application to language modeling have led to success in tasks of text processing, summarizing and machine translation.

LANGUAGE MODELLING MACHINE TRANSLATION MODEL COMPRESSION NETWORK PRUNING SPEECH RECOGNITION