Xception
This is a MXNet implementation of Xception architecture as described in the paper Xception: Deep Learning with Depthwise Separable Convolutions proposed by Franois Chollet at CVPR 2017 from google.
The author's keras implementation can be found in his repo on GitHub.
This is an illustration of Depthwise Separable Convolution with Depthwise convolution and Pointwise convolution.
M indicates the number of input feature map, N indicates the number of output channel.
This MXNet implementation is based on Franois Chollet's keras version. Furthermore, I also refered one implementation from u1234x1234, but my version is more compact. I also attach the training code if you want to train your own data with Xception architecture by yourself. Of course, if you just want to use the pretrained model by keras with MXNet, you can refer the python code from u1234x1234's repo.
In original paper, author deploys Xception on 60 NVIDIA K80 GPUs for training ImageNet dataset. The ImageNet experiments took approximately 3 days. However, I only have 4 GPUs. I'm afraid the computational effiency on MXNet. I'm not sure whether MXNet has any optimization for depthwise separable convolution or not. If there is some more efficient implementation or api from MXNet, please let me know. I'm appreciate for your kindness.
It is worth noting that I decrease the filter_list from [64, 128, 256, 728, 1024, 1536, 2048] (original version) to [64, 64, 128, 364, 512, 768, 1024] (smaller one). It is just because of computational efficency. Moreover, different from original version, I add dropout layer followed the last global average pooling layer.
Install MXNet(0.8.0 or later version) on GPUs mechine with NVIDIA CUDA 8.0, and it's better also installed with cuDNN v5 or later version (I'm not testing cuDNN v7).
ImageNet'12 dataset
Imagenet 1000 class dataset with 1.2 million images. Because this dataset is about 120GB, so you have to download by yourself. Sorry for this inconvenience.
For data preparation, you can refer my pervious part of densenet or you can also visit the repo of Wei Wu. In his page, there is a very detailed information about how to prepare your data.
When you finised data preparation, please make sure the data locates the same folder of source codes. You also need to change path of path_imgrec in line 53 and line 77 of train_xception.py. Then you can run the training cmd just like this (here, I use 4 gpus for training):
python -u train_xception.py --data-dir data/imagenet --data-type imagenet --batch-size 256 --gpus=0,1,2,3
Maybe you should change batch-size from 256 to 128 due to the memory size of GPU.
When we want to train the large dataset and hope to change learning rate manually, or the machine is suddenly shutdown due to some reason, of course, we definitely hope we can continue to train model with previous trained weights. Then, your can use this cmd:
python -u train_xception.py --data-dir data/imagenet --data-type imagenet --batch-size 256 --gpus=0,1,2,3 --model-load-epoch=50 --lr 0.001 --retrain
This means you can retrain your xception model from epoch 50 and change lr=0.001 using 4 GPU.
The training procedure is ongoing. The computational efficency is not optimistic. So, I hope anyone who are mxnet fun can test this code with me. When I finish, I will update more information about training and validation.
[1] Franois Chollet. "Xception: Deep Learning with Depthwise Separable Convolutions" http://openaccess.thecvf.com/content_cvpr_2017/papers/Chollet_Xception_Deep_Learning_CVPR_2017_paper.pdf
[2] Keras implementation of Xception from Franois Chollet
[3] MXNet implementation of Xception from u1234x1234
上一篇:wavenet
下一篇:SSD+Focal Loss
还没有评论,说两句吧!
热门资源
seetafaceJNI
项目介绍 基于中科院seetaface2进行封装的JAVA...
Keras-ResNeXt
Keras ResNeXt Implementation of ResNeXt models...
spark-corenlp
This package wraps Stanford CoreNLP annotators ...
shih-styletransfer
shih-styletransfer Code from Style Transfer ...
inferno-boilerplate
This is a very basic boilerplate example for pe...
智能在线
400-630-6780
聆听.建议反馈
E-mail: support@tusaishared.com