Home

Awesome

ResNet-Matconvnet

:x: I have stopped maintaining this repo. For fine-tuning ResNet, I would suggest using Torch version from Facebook repo.

This repository is a Matconvnet re-implementation of "Deep Residual Learning for Image Recognition",Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun. You can train Deep Residual Network on ImageNet from Scratch or fine-tune pre-trained model on your own dataset. This repo is created by Hang Zhang.

Table of Contents

  1. Get Started
  2. Train from Scratch
  3. Fine-tune Your Own
  4. Changes

Get Started

The code relies on vlfeat, and matconvnet, which should be downloaded and built before running the experiments. You can use the following commend to download them.

git clone -b v1.0 --recurse-submodules https://github.com/zhanghang1989/ResNet-Matconvnet.git

If you have problem with compiling, please refer to the link.

Train from Scratch

  1. Cifar. Reproducing Figure 6 from the original paper.

    run_cifar_experiments([20 32 44 56 110], 'plain', 'gpus', [1]);
    run_cifar_experiments([20 32 44 56 110], 'resnet', 'gpus', [1]);
    
    <p style="text-align:center; font-size:75%; font-style: italic;">Cifar Experiments</p> <div style="text-align:center"><img src ="https://raw.githubusercontent.com/zhanghang1989/ResNet-Matconvnet/master/figure/plain_cifar.png" width="420" /><img src ="https://raw.githubusercontent.com/zhanghang1989/ResNet-Matconvnet/master/figure/resnet_cifar.png" width="420" /></div>

    Reproducing the experiments in Facebook blog. Removing ReLU layer at the end of each residual unit, we observe a small but significant improvement in test performance and the converging progress becomes smoother.

    res_cifar(20, 'modelType', 'resnet', 'reLUafterSum', false,...
    	'expDir', 'data/exp/cifar-resNOrelu-20', 'gpus', [2])
    plot_results_mix('data/exp','cifar',[],[],'plots',{'resnet','resNOrelu'})
    
    <div style="text-align:center"><img src ="https://raw.githubusercontent.com/zhanghang1989/ResNet-Matconvnet/master/figure/resnet_relu.png" width="420" /></div>
  2. Imagenet2012. download the dataset to data/ILSVRC2012 and follow the instructions in setup_imdb_imagenet.m.

    run_experiments([50 101 152], 'gpus', [1 2 3 4 5 6 7 8]);
    
  3. Your own dataset.

    run_experiments([18 34],'datasetName', 'minc',...
    'datafn', @setup_imdb_minc, 'nClasses', 23, 'gpus', [1 2]);
    

Fine-tune Your Own

  1. Download

  2. Fine-tuning

    res_finetune('datasetName', 'minc', 'datafn',...
    @setup_imdb_minc, 'gpus',[1 2]);
    

Changes

  1. 06/21/2016:

  2. 05/17/2016:

    • Reproducing the experiments in Facebook blog, removing ReLU layer at the end of each residual unit.
  3. 05/02/2016:

    • Supported official Matconvnet version.
    • Added Cifar experiments and plots.
  4. 04/27/2016: Re-implementation of Residual Network: