Home

Awesome

Awesome Pruning Awesome

A curated list of neural network pruning and related resources. Inspired by awesome-deep-vision, awesome-adversarial-machine-learning, awesome-deep-learning-papers and Awesome-NAS.

Please feel free to pull requests or open an issue to add papers.

Table of Contents

Type of Pruning

TypeFWSOther
ExplanationFilter pruningWeight pruningSpecial Networksother types

A Survey of Structured Pruning (arXiv version and IEEE T-PAMI version)

Please cite our paper if it's helpful:

@article{he2024structured,
  author={He, Yang and Xiao, Lingao},
  journal={IEEE Transactions on Pattern Analysis and Machine Intelligence}, 
  title={Structured Pruning for Deep Convolutional Neural Networks: A Survey}, 
  year={2024},
  volume={46},
  number={5},
  pages={2900-2919},
  doi={10.1109/TPAMI.2023.3334614}}

The related papers are categorized as below: Structured Pruning Taxonomy

2023

TitleVenueTypeCode
Revisiting Pruning at Initialization Through the Lens of Ramanujan GraphICLRWPyTorch(Author)(Releasing)
Unmasking the Lottery Ticket Hypothesis: What's Encoded in a Winning Ticket's Mask?ICLRW-
Bit-Pruning: A Sparse Multiplication-Less Dot-ProductICLRWCode Deleted
NTK-SAP: Improving neural network pruning by aligning training dynamicsICLRW-
A Unified Framework for Soft Threshold PruningICLRWPyTorch(Author)
CrAM: A Compression-Aware MinimizerICLRW-
Trainability Preserving Neural PruningICLRF-
DFPC: Data flow driven pruning of coupled channels without dataICLRFPyTorch(Author)
TVSPrune - Pruning Non-discriminative filters via Total Variation separability of intermediate representations without fine tuningICLRFPyTorch(Author)
HomoDistil: Homotopic Task-Agnostic Distillation of Pre-trained TransformersICLRF-
MECTA: Memory-Economic Continual Test-Time Model AdaptationICLRF-
DepthFL : Depthwise Federated Learning for Heterogeneous ClientsICLRF-
OTOv2: Automatic, Generic, User-FriendlyICLRFPyTorch(Author)
Over-parameterized Model Optimization with Polyak-Lojasiewicz ConditionICLRF-
Pruning Deep Neural Networks from a Sparsity PerspectiveICLRWFPyTorch(Author)
Holistic Adversarially Robust PruningICLRWF-
How I Learned to Stop Worrying and Love RetrainingICLRWFPyTorch(Author)
Symmetric Pruning in Quantum Neural NetworksICLRS-
Rethinking Graph Lottery Tickets: Graph Sparsity MattersICLRS-
Joint Edge-Model Sparse Learning is Provably Efficient for Graph Neural NetworksICLRS-
Searching Lottery Tickets in Graph Neural Networks: A Dual PerspectiveICLRS-
Diffusion Models for Causal Discovery via Topological OrderingICLRS-
A General Framework For Proving The Equivariant Strong Lottery Ticket HypothesisICLROther-
Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!ICLROther-
Minimum Variance Unbiased N:M Sparsity for the Neural GradientsICLROther-

2022

TitleVenueTypeCode
Parameter-Efficient Masking NetworksNeurIPSWPyTorch(Author)
"Lossless" Compression of Deep Neural Networks: A High-dimensional Neural Tangent Kernel ApproachNeurIPSWPyTorch(Author)
Losses Can Be Blessings: Routing Self-Supervised Speech Representations Towards Efficient Multilingual and Multitask Speech ProcessingNeurIPSWPyTorch(Author)
Models Out of Line: A Fourier Lens on Distribution Shift RobustnessNeurIPSWPyTorch(Author)
Robust Binary Models by Pruning Randomly-initialized NetworksNeurIPSWPyTorch(Author)
Rare Gems: Finding Lottery Tickets at InitializationNeurIPSWPyTorch(Author)
Optimal Brain Compression: A Framework for Accurate Post-Training Quantization and PruningNeurIPSWPyTorch(Author)
Pruning’s Effect on Generalization Through the Lens of Training and RegularizationNeurIPSW-
Back Razor: Memory-Efficient Transfer Learning by Self-Sparsified BackpropagationNeurIPSWPyTorch(Author)
Analyzing Lottery Ticket Hypothesis from PAC-Bayesian Theory PerspectiveNeurIPSW-
Sparse Winning Tickets are Data-Efficient Image RecognizersNeurIPSWPyTorch(Author)
Lottery Tickets on a Data Diet: Finding Initializations with Sparse Trainable NetworksNeurIPSW-
Weighted Mutual Learning with Diversity-Driven Model CompressionNeurIPSF-
SInGE: Sparsity via Integrated Gradients Estimation of Neuron RelevanceNeurIPSF-
Data-Efficient Structured Pruning via Submodular OptimizationNeurIPSFPyTorch(Author)
Structural Pruning via Latency-Saliency KnapsackNeurIPSFPyTorch(Author)
Recall Distortion in Neural Network Pruning and the Undecayed Pruning AlgorithmNeurIPSWF-
Pruning Neural Networks via Coresets and Convex Geometry: Towards No AssumptionsNeurIPSWF-
Controlled Sparsity via Constrained Optimization or: How I Learned to Stop Tuning Penalties and Love ConstraintsNeurIPSWFPyTorch(Author)
Advancing Model Pruning via Bi-level OptimizationNeurIPSWFPyTorch(Author)
Emergence of Hierarchical Layers in a Single Sheet of Self-Organizing Spiking NeuronsNeurIPSS-
CryptoGCN: Fast and Scalable Homomorphically Encrypted Graph Convolutional Network InferenceNeurIPSSPyTorch(Author)(Releasing)
Transform Once: Efficient Operator Learning in Frequency DomainNeurIPSOtherPyTorch(Author)(Releasing)
Most Activation Functions Can Win the Lottery Without Excessive DepthNeurIPSOtherPyTorch(Author)
Pruning has a disparate impact on model accuracyNeurIPSOther-
Model Preserving Compression for Neural NetworksNeurIPSOtherPyTorch(Author)
Prune Your Model Before Distill ItECCVWPyTorch(Author)
FedLTN: Federated Learning for Sparse and Personalized Lottery Ticket NetworksECCVW-
FairGRAPE: Fairness-Aware GRAdient Pruning mEthod for Face Attribute ClassificationECCVFPyTorch(Author)
SuperTickets: Drawing Task-Agnostic Lottery Tickets from Supernets via Jointly Architecture Searching and Parameter PruningECCVFPyTorch(Author)
Ensemble Knowledge Guided Sub-network Search and Fine-Tuning for Filter PruningECCVFPyTorch(Author)
CPrune: Compiler-Informed Model Pruning for Efficient Target-Aware DNN ExecutionECCVFPyTorch(Author)
Soft Masking for Cost-Constrained Channel PruningECCVFPyTorch(Author)
Filter Pruning via Feature Discrimination in Deep Neural NetworksECCVF-
Disentangled Differentiable Network PruningECCVF-
Interpretations Steered Network Pruning via Amortized Inferred Saliency MapsECCVFPyTorch(Author)
Bayesian Optimization with Clustering and Rollback for CNN Auto PruningECCVFPyTorch(Author)
Multi-granularity Pruning for Model Acceleration on Mobile DevicesECCVWF-
Exploring Lottery Ticket Hypothesis in Spiking Neural NetworksECCVSPyTorch(Author)
Towards Ultra Low Latency Spiking Neural Networks for Vision and Sequential Tasks Using Temporal PruningECCVS-
Recent Advances on Neural Network Pruning at InitializationIJCAIWPyTorch(Author)
FedDUAP: Federated Learning with Dynamic Update and Adaptive Pruning Using Shared Data on the ServerIJCAIF-
On the Channel Pruning using Graph Convolution Network for Convolutional Neural Network AccelerationIJCAIF-
Pruning-as-Search: Efficient Neural Architecture Search via Channel Pruning and Structural ReparameterizationIJCAIF-
Neural Network Pruning by Cooperative CoevolutionIJCAIF-
SPDY: Accurate Pruning with Speedup GuaranteesICMLWPyTorch(Author)
Sparse Double Descent: Where Network Pruning Aggravates OverfittingICMLWPyTorch(Author)
The Combinatorial Brain Surgeon: Pruning Weights That Cancel One Another in Neural NetworksICMLWPyTorch(Author)
Linearity Grafting: Relaxed Neuron Pruning Helps Certifiable RobustnessICMLFPyTorch(Author)
Winning the Lottery Ahead of Time: Efficient Early Network PruningICMLFPyTorch(Author)
Topology-Aware Network Pruning using Multi-stage Graph Embedding and Reinforcement LearningICMLFPyTorch(Author)
Fast Lossless Neural Compression with Integer-Only Discrete FlowsICMLFPyTorch(Author)
DepthShrinker: A New Compression Paradigm Towards Boosting Real-Hardware Efficiency of Compact Neural NetworksICMLOtherPyTorch(Author)
PAC-Net: A Model Pruning Approach to Inductive Transfer LearningICMLOther-
Neural Network Pruning Denoises the Features and Makes Local Connectivity Emerge in Visual TasksICMLOtherPyTorch(Author)
Interspace Pruning: Using Adaptive Filter Representations To Improve Training of Sparse CNNsCVPRW-
Masking Adversarial Damage: Finding Adversarial Saliency for Robust and Sparse NetworkCVPRW-
When To Prune? A Policy Towards Early Structural PruningCVPRF-
Fire Together Wire Together: A Dynamic Pruning Approach With Self-Supervised Mask PredictionFire Together Wire Together: A Dynamic Pruning Approach With Self-Supervised Mask PredictionCVPRF-
Revisiting Random Channel Pruning for Neural Network CompressionCVPRFPyTorch(Author)(Releasing)
Learning Bayesian Sparse Networks With Full Experience Replay for Continual LearningCVPRF-
DECORE: Deep Compression With Reinforcement LearningCVPRF-
CHEX: CHannel EXploration for CNN Model CompressionCVPRF-
Compressing Models With Few Samples: Mimicking Then ReplacingCVPRFPyTorch(Author)(Releasing)
Contrastive Dual Gating: Learning Sparse Features With Contrastive LearningCVPRWF-
DiSparse: Disentangled Sparsification for Multitask Model CompressionCVPROtherPyTorch(Author)
Learning Pruning-Friendly Networks via Frank-Wolfe: One-Shot, Any-Sparsity, And No RetrainingICLR (Spotlight)WPyTorch(Author)
On Lottery Tickets and Minimal Task Representations in Deep Reinforcement LearningICLR (Spotlight)W-
An Operator Theoretic View On Pruning Deep Neural NetworksICLRWPyTorch(Author)
Effective Model Sparsification by Scheduled Grow-and-Prune MethodsICLRWPyTorch(Author)
Signing the Supermask: Keep, Hide, InvertICLRW-
How many degrees of freedom do we need to train deep networks: a loss landscape perspectiveICLRWPyTorch(Author)
Dual Lottery Ticket HypothesisICLRWPyTorch(Author)
Peek-a-Boo: What (More) is Disguised in a Randomly Weighted Neural Network, and How to Find It EfficientlyICLRWPyTorch(Author)
Sparsity Winning Twice: Better Robust Generalization from More Efficient TrainingICLRWPyTorch(Author)
SOSP: Efficiently Capturing Global Correlations by Second-Order Structured PruningICLR (Spotlight)FPyTorch(Author)(Releasing)
Pixelated Butterfly: Simple and Efficient Sparse training for Neural Network ModelsICLR (Spotlight)FPyTorch(Author)
Revisit Kernel Pruning with Lottery Regulated Grouped ConvolutionsICLRFPyTorch(Author)
Plant 'n' Seek: Can You Find the Winning Ticket?ICLRFPyTorch(Author)
Proving the Lottery Ticket Hypothesis for Convolutional Neural NetworksICLRFPyTorch(Author)
On the Existence of Universal Lottery TicketsICLRFPyTorch(Author)
Training Structured Neural Networks Through Manifold Identification and Variance ReductionICLRFPyTorch(Author)
Learning Efficient Image Super-Resolution Networks via Structure-Regularized PruningICLRFPyTorch(Author)
Prospect Pruning: Finding Trainable Weights at Initialization using Meta-GradientsICLRWFPyTorch(Author)
The Unreasonable Effectiveness of Random Pruning: Return of the Most Naive Baseline for Sparse TrainingICLROtherPyTorch(Author)
Prune and Tune Ensembles: Low-Cost Ensemble Learning with Sparse Independent SubnetworksAAAIW-
Prior Gradient Mask Guided Pruning-Aware Fine-TuningAAAIF-
Convolutional Neural Network Compression through Generalized Kronecker Product DecompositionAAAIOther-

2021

TitleVenueTypeCode
Validating the Lottery Ticket Hypothesis with Inertial Manifold TheoryNeurIPSW-
The Elastic Lottery Ticket HypothesisNeurIPSWPyTorch(Author)
Sanity Checks for Lottery Tickets: Does Your Winning Ticket Really Win the Jackpot?NeurIPSWPyTorch(Author)
Why Lottery Ticket Wins? A Theoretical Perspective of Sample Complexity on Sparse Neural NetworksNeurIPSW-
You are caught stealing my winning lottery ticket! Making a lottery ticket claim its ownershipNeurIPSWPyTorch(Author)
Pruning Randomly Initialized Neural Networks with Iterative RandomizationNeurIPSWPyTorch(Author)
Sparse Training via Boosting Pruning Plasticity with NeuroregenerationNeurIPSWPyTorch(Author)
AC/DC: Alternating Compressed/DeCompressed Training of Deep Neural NetworksNeurIPSWPyTorch(Author)
A Winning Hand: Compressing Deep Networks Can Improve Out-of-Distribution RobustnessNeurIPSWPyTorch(Author)
Rethinking the Pruning Criteria for Convolutional Neural NetworkNeurIPSF-
Only Train Once: A One-Shot Neural Network Training And Pruning FrameworkNeurIPSFPyTorch(Author)
CHIP: CHannel Independence-based Pruning for Compact Neural NetworksNeurIPSFPyTorch(Author)
RED : Looking for Redundancies for Data-FreeStructured Compression of Deep Neural NetworksNeurIPSF-
Compressing Neural Networks: Towards Determining the Optimal Layer-wise DecompositionNeurIPSFPyTorch(Author)
Sparse Flows: Pruning Continuous-depth ModelsNeurIPSWFPyTorch(Author)
Scaling Up Exact Neural Network Compression by ReLU StabilityNeurIPSSPyTorch(Author)
Discriminator in GAN Compression: A Generator-discriminator Cooperative Compression SchemeNeurIPSSPyTorch(Author)
Heavy Tails in SGD and Compressibility of Overparametrized Neural NetworksNeurIPSOtherPyTorch(Author)
ResRep: Lossless CNN Pruning via Decoupling Remembering and ForgettingICCVFPyTorch(Author)
Achieving on-Mobile Real-Time Super-Resolution with Neural Architecture and Pruning SearchICCVF-
GDP: Stabilized Neural Network Pruning via Gates with Differentiable PolarizationICCVF-
Auto Graph Encoder-Decoder for Neural Network PruningICCVF-
Exploration and Estimation for Model CompressionICCVF-
Sub-Bit Neural Networks: Learning To Compress and Accelerate Binary Neural NetworksICCVOtherPyTorch(Author)
On the Predictability of Pruning Across ScalesICMLW-
A Probabilistic Approach to Neural Network PruningICMLF-
Accelerate CNNs from Three Dimensions: A Comprehensive Pruning FrameworkICMLF-
Group Fisher Pruning for Practical Network CompressionICMLFPyTorch(Author)
Towards Compact CNNs via Collaborative CompressionCVPRFPyTorch(Author)
Permute, Quantize, and Fine-tune: Efficient Compression of Neural NetworksCVPRFPyTorch(Author)
NPAS: A Compiler-aware Framework of Unified Network Pruning andArchitecture Search for Beyond Real-Time Mobile AccelerationCVPRF-
Network Pruning via Performance MaximizationCVPRF-
Convolutional Neural Network Pruning with Structural Redundancy ReductionCVPRF-
Manifold Regularized Dynamic Network PruningCVPRF-
Joint-DetNAS: Upgrade Your Detector with NAS, Pruning and Dynamic DistillationCVPRFO-
Content-Aware GAN CompressionCVPRSPyTorch(Author)
Multi-Prize Lottery Ticket Hypothesis: Finding Accurate Binary Neural Networks by Pruning A Randomly Weighted NetworkICLRWPyTorch(Author)
Layer-adaptive Sparsity for the Magnitude-based PruningICLRWPyTorch(Author)
Pruning Neural Networks at Initialization: Why Are We Missing the Mark?ICLRW-
Robust Pruning at InitializationICLRW-
A Gradient Flow Framework For Analyzing Network PruningICLRFPyTorch(Author)
Neural Pruning via Growing RegularizationICLRFPyTorch(Author)
ChipNet: Budget-Aware Pruning with Heaviside Continuous ApproximationsICLRFPyTorch(Author)
Network Pruning That Matters: A Case Study on Retraining VariantsICLRFPyTorch(Author)

2020

TitleVenueTypeCode
Optimal Lottery Tickets via Subset Sum: Logarithmic Over-Parameterization is SufficientNeurIPSW-
Winning the Lottery with Continuous SparsificationNeurIPSWPyTorch(Author)
HYDRA: Pruning Adversarially Robust Neural NetworksNeurIPSWPyTorch(Author)
Logarithmic Pruning is All You NeedNeurIPSW-
Directional Pruning of Deep Neural NetworksNeurIPSW-
Movement Pruning: Adaptive Sparsity by Fine-TuningNeurIPSWPyTorch(Author)
Sanity-Checking Pruning Methods: Random Tickets can Win the JackpotNeurIPSWPyTorch(Author)
Neuron Merging: Compensating for Pruned NeuronsNeurIPSFPyTorch(Author)
Neuron-level Structured Pruning using Polarization RegularizerNeurIPSFPyTorch(Author)
SCOP: Scientific Control for Reliable Neural Network PruningNeurIPSFPyTorch(Author)
Storage Efficient and Dynamic Flexible Runtime Channel Pruning via Deep Reinforcement LearningNeurIPSF-
The Generalization-Stability Tradeoff In Neural Network PruningNeurIPSFPyTorch(Author)
Greedy Optimization Provably Wins the Lottery: Logarithmic Number of Winning Tickets is EnoughNeurIPSWF-
Pruning Filter in FilterNeurIPSOtherPyTorch(Author)
Position-based Scaled Gradient for Model Quantization and PruningNeurIPSOtherPyTorch(Author)
Bayesian Bits: Unifying Quantization and PruningNeurIPSOther-
Pruning neural networks without any data by iteratively conserving synaptic flowNeurIPSOtherPyTorch(Author)
Meta-Learning with Network PruningECCVW-
Accelerating CNN Training by Pruning Activation GradientsECCVW-
EagleEye: Fast Sub-net Evaluation for Efficient Neural Network PruningECCV (Oral)FPyTorch(Author)
DSA: More Efficient Budgeted Pruning via Differentiable Sparsity AllocationECCVF-
DHP: Differentiable Meta Pruning via HyperNetworksECCVFPyTorch(Author)
DA-NAS: Data Adapted Pruning for Efficient Neural Architecture Search SECCVOther-
Differentiable Joint Pruning and Quantization for Hardware EfficiencyECCVOther-
Channel Pruning via Automatic Structure SearchIJCAIFPyTorch(Author)
Adversarial Neural Pruning with Latent Vulnerability SuppressionICMLW-
Proving the Lottery Ticket Hypothesis: Pruning is All You NeedICMLW-
Network Pruning by Greedy Subnetwork SelectionICMLF-
Operation-Aware Soft Channel Pruning using Differentiable MasksICMLF-
DropNet: Reducing Neural Network Complexity via Iterative PruningICMLF-
Soft Threshold Weight Reparameterization for Learnable SparsityICMLWFPytorch(Author)
Structured Compression by Weight Encryption for Unstructured Pruning and QuantizationCVPRW-
Automatic Neural Network Compression by Sparsity-Quantization Joint Learning: A Constrained Optimization-Based ApproachCVPRW-
Towards Efficient Model Compression via Learned Global RankingCVPR (Oral)FPytorch(Author)
HRank: Filter Pruning using High-Rank Feature MapCVPR (Oral)FPytorch(Author)
Neural Network Pruning with Residual-Connections and Limited-DataCVPR (Oral)F-
DMCP: Differentiable Markov Channel Pruning for Neural NetworksCVPR (Oral)FTensorFlow(Author)
Group Sparsity: The Hinge Between Filter Pruning and Decomposition for Network CompressionCVPRFPyTorch(Author)
Few Sample Knowledge Distillation for Efficient Network CompressionCVPRF-
Discrete Model Compression With Resource Constraint for Deep Neural NetworksCVPRF-
Learning Filter Pruning Criteria for Deep Convolutional Neural Networks AccelerationCVPRF-
APQ: Joint Search for Network Architecture, Pruning and Quantization PolicyCVPRF-
Multi-Dimensional Pruning: A Unified Framework for Model CompressionCVPR (Oral)WF-
A Signal Propagation Perspective for Pruning Neural Networks at InitializationICLR (Spotlight)W-
ProxSGD: Training Structured Neural Networks under Regularization and ConstraintsICLRWTF+PT(Author)
One-Shot Pruning of Recurrent Neural Networks by Jacobian Spectrum EvaluationICLRW-
Lookahead: A Far-sighted Alternative of Magnitude-based PruningICLRWPyTorch(Author)
Data-Independent Neural Pruning via CoresetsICLRW-
Provable Filter Pruning for Efficient Neural NetworksICLRF-
Dynamic Model Pruning with FeedbackICLRWF-
Comparing Rewinding and Fine-tuning in Neural Network PruningICLR (Oral)WFTensorFlow(Author)
AutoCompress: An Automatic DNN Structured Pruning Framework for Ultra-High Compression RatesAAAIF-
Reborn filters: Pruning convolutional neural networks with limited dataAAAIF-
DARB: A Density-Aware Regular-Block Pruning for Deep Neural NetworksAAAIOther-
Pruning from ScratchAAAIOther-

2019

TitleVenueTypeCode
Deconstructing Lottery Tickets: Zeros, Signs, and the SupermaskNeurIPSWTensorFlow(Author)
One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizersNeurIPSW-
Global Sparse Momentum SGD for Pruning Very Deep Neural NetworksNeurIPSWPyTorch(Author)
AutoPrune: Automatic Network Pruning by Regularizing Auxiliary ParametersNeurIPSW-
Network Pruning via Transformable Architecture SearchNeurIPSFPyTorch(Author)
Gate Decorator: Global Filter Pruning Method for Accelerating Deep Convolutional Neural NetworksNeurIPSFPyTorch(Author)
Model Compression with Adversarial Robustness: A Unified Optimization FrameworkNeurIPSOtherPyTorch(Author)
Adversarial Robustness vs Model Compression, or Both?ICCVWPyTorch(Author)
MetaPruning: Meta Learning for Automatic Neural Network Channel PruningICCVFPyTorch(Author)
Accelerate CNN via Recursive Bayesian PruningICCVF-
Learning Filter Basis for Convolutional Neural Network CompressionICCVOther-
Co-Evolutionary Compression for Unpaired Image TranslationICCVS-
COP: Customized Deep Model Compression via Regularized Correlation-Based Filter-Level PruningIJCAIFTensorflow(Author)
Filter Pruning via Geometric Median for Deep Convolutional Neural Networks AccelerationCVPR (Oral)FPyTorch(Author)
Towards Optimal Structured CNN Pruning via Generative Adversarial LearningCVPRFPyTorch(Author)
Centripetal SGD for Pruning Very Deep Convolutional Networks with Complicated StructureCVPRFPyTorch(Author)
On Implicit Filter Level Sparsity in Convolutional Neural Networks, Extension1, Extension2CVPRFPyTorch(Author)
Structured Pruning of Neural Networks with Budget-Aware RegularizationCVPRF-
Importance Estimation for Neural Network PruningCVPRFPyTorch(Author)
OICSR: Out-In-Channel Sparsity Regularization for Compact Deep Neural NetworksCVPRF-
Variational Convolutional Neural Network PruningCVPRF-
Partial Order Pruning: for Best Speed/Accuracy Trade-off in Neural Architecture SearchCVPROtherTensorFlow(Author)
Collaborative Channel Pruning for Deep NetworksICMLF-
Approximated Oracle Filter Pruning for Destructive CNN Width Optimization githubICMLF-
EigenDamage: Structured Pruning in the Kronecker-Factored EigenbasisICMLFPyTorch(Author)
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural NetworksICLR (Best)WTensorFlow(Author)
SNIP: Single-shot Network Pruning based on Connection SensitivityICLRWTensorFLow(Author)
Dynamic Channel Pruning: Feature Boosting and SuppressionICLRFTensorFlow(Author)
Rethinking the Value of Network PruningICLRFPyTorch(Author)
Dynamic Sparse Graph for Efficient Deep LearningICLRFCUDA(3rd)

2018

TitleVenueTypeCode
Frequency-Domain Dynamic Pruning for Convolutional Neural NetworksNeurIPSW-
Discrimination-aware Channel Pruning for Deep Neural NetworksNeurIPSFTensorFlow(Author)
Learning Sparse Neural Networks via Sensitivity-Driven RegularizationNeurIPSWF-
Constraint-Aware Deep Neural Network CompressionECCVWSkimCaffe(Author)
A Systematic DNN Weight Pruning Framework using Alternating Direction Method of MultipliersECCVWCaffe(Author)
Amc: Automl for model compression and acceleration on mobile devicesECCVFTensorFlow(3rd)
Data-Driven Sparse Structure Selection for Deep Neural NetworksECCVFMXNet(Author)
Coreset-Based Neural Network CompressionECCVFPyTorch(Author)
Soft Filter Pruning for Accelerating Deep Convolutional Neural NetworksIJCAIFPyTorch(Author)
Accelerating Convolutional Networks via Global & Dynamic Filter PruningIJCAIF-
Weightless: Lossy weight encoding for deep neural network compressionICMLW-
Compressing Neural Networks using the Variational Information BottleneckICMLFPyTorch(Author)
Deep k-Means: Re-Training and Parameter Sharing with Harder Cluster Assignments for Compressing Deep ConvolutionsICMLOtherPyTorch(Author)
CLIP-Q: Deep Network Compression Learning by In-Parallel Pruning-QuantizationCVPRW-
“Learning-Compression” Algorithms for Neural Net PruningCVPRW-
PackNet: Adding Multiple Tasks to a Single Network by Iterative PruningCVPRFPyTorch(Author)
NISP: Pruning Networks using Neuron Importance Score PropagationCVPRF-
To prune, or not to prune: exploring the efficacy of pruning for model compressionICLRW-
Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution LayersICLRFTensorFlow(Author), PyTorch(3rd)

2017

TitleVenueTypeCode
Net-Trim: Convex Pruning of Deep Neural Networks with Performance GuaranteeNeurIPSWTensorFlow(Author)
Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain SurgeonNeurIPSWPyTorch(Author)
Runtime Neural PruningNeurIPSF-
Structured Bayesian Pruning via Log-Normal Multiplicative NoiseNeurIPSF-
Bayesian Compression for Deep LearningNeurIPSF-
ThiNet: A Filter Level Pruning Method for Deep Neural Network CompressionICCVFCaffe(Author), PyTorch(3rd)
Channel pruning for accelerating very deep neural networksICCVFCaffe(Author)
Learning Efficient Convolutional Networks Through Network SlimmingICCVFPyTorch(Author)
Variational Dropout Sparsifies Deep Neural NetworksICMLW-
Combined Group and Exclusive Sparsity for Deep Neural NetworksICMLWF-
Designing Energy-Efficient Convolutional Neural Networks using Energy-Aware PruningCVPRW-
Pruning Filters for Efficient ConvNetsICLRFPyTorch(3rd)
Pruning Convolutional Neural Networks for Resource Efficient InferenceICLRFTensorFlow(3rd)

2016

TitleVenueTypeCode
Dynamic Network Surgery for Efficient DNNsNeurIPSWCaffe(Author)
Learning the Number of Neurons in Deep NetworksNeurIPSF-
Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman CodingICLR (Best)WCaffe(Author)

2015

TitleVenueTypeCode
Learning both Weights and Connections for Efficient Neural NetworksNeurIPSWPyTorch(3rd)

Related Repo

Awesome-model-compression-and-acceleration

EfficientDNNs

Embedded-Neural-Network

awesome-AutoML-and-Lightweight-Models

Model-Compression-Papers

knowledge-distillation-papers

Network-Speed-and-Compression