资源算法 ResNet-Matconvnet

ResNet-Matconvnet

2020-02-21 | |  34 |   0 |   0

ResNet-Matconvnet

❌ I have stopped maintaining this repo. For fine-tuning ResNet, I would suggest using Torch version from Facebook repo.

This repository is a Matconvnet re-implementation of "Deep Residual Learning for Image Recognition",Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun. You can train Deep Residual Network on ImageNet from Scratch or fine-tune pre-trained model on your own dataset. This repo is created by Hang Zhang.

Table of Contents

  1. Get Started

  2. Train from Scratch

  3. Fine-tune Your Own

  4. Changes

Get Started

The code relies on vlfeat, and matconvnet, which should be downloaded and built before running the experiments. You can use the following commend to download them.

git clone -b v1.0 --recurse-submodules https://github.com/zhanghang1989/ResNet-Matconvnet.git

If you have problem with compiling, please refer to the link.

Train from Scratch

  1. Cifar. Reproducing Figure 6 from the original paper.

    run_cifar_experiments([20 32 44 56 110], 'plain', 'gpus', [1]);
    run_cifar_experiments([20 32 44 56 110], 'resnet', 'gpus', [1]);

    Cifar Experiments

    Reproducing the experiments in Facebook blog. Removing ReLU layer at the end of each residual unit, we observe a small but significant improvement in test performance and the converging progress becomes smoother.

    res_cifar(20, 'modelType', 'resnet', 'reLUafterSum', false,...
    	'expDir', 'data/exp/cifar-resNOrelu-20', 'gpus', [2])
    plot_results_mix('data/exp','cifar',[],[],'plots',{'resnet','resNOrelu'})

  2. Imagenet2012. download the dataset to data/ILSVRC2012 and follow the instructions in setup_imdb_imagenet.m.

    run_experiments([50 101 152], 'gpus', [1 2 3 4 5 6 7 8]);
  3. Your own dataset.

    run_experiments([18 34],'datasetName', 'minc',...'datafn', @setup_imdb_minc, 'nClasses', 23, 'gpus', [1 2]);

Fine-tune Your Own

  1. Download

  2. Fine-tuning

    res_finetune('datasetName', 'minc', 'datafn',...@setup_imdb_minc, 'gpus',[1 2]);

Changes

  1. 06/21/2016:

  2. 05/17/2016:

    • Reproducing the experiments in Facebook blog, removing ReLU layer at the end of each residual unit.

  3. 05/02/2016:

    • Supported official Matconvnet version.

    • Added Cifar experiments and plots.

  4. 04/27/2016: Re-implementation of Residual Network:


上一篇:resnet-tf

下一篇: resnet-cppn-gan-tensorflow

用户评价
全部评价

热门资源

  • seetafaceJNI

    项目介绍 基于中科院seetaface2进行封装的JAVA...

  • spark-corenlp

    This package wraps Stanford CoreNLP annotators ...

  • Keras-ResNeXt

    Keras ResNeXt Implementation of ResNeXt models...

  • capsnet-with-caps...

    CapsNet with capsule-wise convolution Project ...

  • shih-styletransfer

    shih-styletransfer Code from Style Transfer ...