Awesome
Magnitude and Similarity based Variable Rate Filter Pruning for Efficient Convolution Neural Networks
This repo is the official implementation of the following paper.
Pruning the selected set of filters & restructuring the network is based on VainF/Torch-Pruning.
Table of Contents
Requirements
- Python 3.9.7
- PyTorch 1.10.2
- TorchVision 0.11.2
- matplotlib 3.5.1
- scipy 1.8.0
Note: These are the verified version of the tools used in the experiment. You can test with other versions as well.
Models
The trained baseline models are uploaded in baseline_models and corresponding pruned models are uploaded in pruned_models. The code and models for ResNet in CIFAR10 needs some cleanup and will be released later.
VGGNet on CIFAR10
The CIFAR-10 dataset will be downloaded automatically when we run the script for the first time. No extra step is required to manage this dataset.
The file main_vgg_cifar10.py is used for training, pruning, and evaluating VGGNet in the CIFAR-10 dataset.
Training
python main_vgg_cifar10.py --dataset cifar10 --depth 16 \
--mode train \
--save_path "./save/path/" \
--total_epoches 160 \
--schedule 40 80 120 \
--gammas 0.2 0.2 0.2 \
--lr 0.1 --decay 0.0005 --batch_size 256
Also, script scripts/run_vgg16_cifar10_baseline_train.sh can be used for training. In the case of the CIFAR-10 dataset, an average of three experiments are used for evaluation.
Pruning
python main_vgg_cifar10.py --dataset cifar10 --depth 16 \
--mode prune \
--pretrain_path "./pretrain/path/" \
--save_path "./save/path/" \
--method mix \
--rate_flop 0.342 \
--total_epoches 160 \
--schedule 40 80 120 \
--gammas 0.2 0.2 0.2 \
--lr 0.01 --decay 0.0005 --batch_size 256
Also, refer to the script scripts/run_vgg16_cifar10_pruning.sh for more detail instructions.
Evaluation
# you can directly run this script for evaluation of one vgg16 model.
# pruned model is present in the given path.
python main_vgg_cifar10.py --dataset cifar10 --depth 16 \
--mode eval \
--depth 16 \
--pretrain_path "./baseline_models/CIFAR10-VGG16_BASELINE/vgg16_round_1/model_best.pth.tar" \
--pruned_path "./pruned_models/CIFAR10-VGG16-PRUNED/vgg16_round_1_flop_0.342/model_best.pth.tar" \
--save_path "./eval_pruning_cifar10_vgg16"
Also, refer to the script scripts/run_vgg16_cifar10_pruning_eval.sh for more detail instructions.
ResNet on ImageNet
The file main_resnet_imagenet.py is used for training, pruning, and evaluating ResNets in the ImageNet dataset.
Download and prepare ImageNet dataset
-
Download the images from http://image-net.org/download-images
-
Extract the training data:
mkdir train && mv ILSVRC2012_img_train.tar train/ && cd train
tar -xvf ILSVRC2012_img_train.tar && rm -f ILSVRC2012_img_train.tar
find . -name "*.tar" | while read NAME ; do mkdir -p "${NAME%.tar}"; tar -xvf "${NAME}" -C "${NAME%.tar}"; rm -f "${NAME}"; done
cd ..
- Extract the validation data and move images to subfolders:
mkdir val && mv ILSVRC2012_img_val.tar val/ && cd val && tar -xvf ILSVRC2012_img_val.tar
wget -qO- https://raw.githubusercontent.com/soumith/imagenetloader.torch/master/valprep.sh | bash
Note: Here we suppose the ImageNet dataset is stored in "C:/ImageNet"
directory.
Training
python main_resnet_imagenet.py --data_path "C:/ImageNet" \
--mode train \
--arch resnet18 \
--save_path "./save/path/" \
--total_epoches 100 \
--decay_epoch_step 30 \
--workers 8 \
--lr 0.1 --decay 0.0001 --batch_size 256
Also, refer to the script scripts/run_resnet_imagenet_baseline_train.sh for more detail instructions.
Pruning
python main_resnet_imagenet.py --data_path "C:/ImageNet" \
--mode prune \
--arch resnet18 \
--pretrain_path "pretrain/path" \
--save_path "save/path" \
--method mix \
--rate_flop 0.41 \
--total_epoches 100 \
--start_epoch 20 \
--recover_epoch 2 \
--retrain_step_flop 0.03 \
--decay_epoch_step 30 \
--workers 8 \
--lr 0.01 --decay 0.0001 --batch_size 256
Also, refer to the script scripts/run_resnet_imagenet_pruning.sh for more detail instructions.
Evaluation
# you can directly run this script for evaluation of one resnet50 model.
# pruned model is present in the given path.
# even if you do not have the ImageNet dataset, you can still calculate FLOPs reduction rate
python main_resnet_imagenet.py --data_path "C:/ImageNet" \
--mode eval \
--arch resnet50 \
--pretrain_path "./baseline_models/ImageNet-RESNET-BASELINE/resnet50/resnet50.model_best.pth.tar" \
--pruned_path "./pruned_models/ImageNet-RESNET-PRUNED/resnet50_flop[0.42]/resnet50.model_best.pth.tar" \
--save_path "save/path" \
--workers 8 \
--batch_size 256
Also, refer to the script scripts/run_resnet_imagenet_pruning_eval.sh for more detail instructions.
ResNet on CIFAR-10
-- Will be released soon ...
How to Cite this Work?
@article{ghimire2023magnitude,
title={Magnitude and Similarity Based Variable Rate Filter Pruning for Efficient Convolution Neural Networks},
author={Ghimire, Deepak and Kim, Seong-Heum},
journal={Applied Sciences},
volume={13},
number={1},
pages={316},
year={2022},
publisher={Multidisciplinary Digital Publishing Institute}
}