资源算法Face-alignment-mobilenet-v2

Face-alignment-mobilenet-v2

2020-02-27 | |  40 |   0 |   0

Face-alignment-mobilenet-v2

Face Alignment by MobileNetv2. Note that MTCNN is used to provided the input boundingbox. You need to modify the path of images in order to run the demo.

Network Structure

The most important part of the mobilenet-v2 network is the design of bottleneck. In our experiments, we crop the face image by the boundingbox and resize it to , which is the input size of the network. Based on this, we can design the structure of our customized mobilenet-v2 for facial landmark lacalization. Note that the receptive field is a key factor to the design of the network.

InputOperatortchannelsnstride
conv2d-1612
bottleneck62412
conv2d62411
conv2d63212
conv2d63211
conv2d66412
conv2d66411
inner product-2001-
200inner product-2001-
200inner product-501-
50inner product-1361-

Note that this structure mainly has two features:

  • Use LeakyReLU rather than ReLU.

  • Use bottleneck embedding, which is 50 in our experiments.

Training

The training data including:

  • Training data of 300W dataset

  • Training data of Menpo dataset

Data Augmentation

Data augmentation is important to the performance of face alignment. I have tried several kinds of data augmentation method, including:

  • Random Flip.

  • Random Shift.

  • Random Scale.

  • Random Rotation. The image is rotated by the degree sampled from -30 to 30.

  • Random Noise. Gaussian noise is added to the input images.

Performance

The performance on 300W is not good enough. May be I need to try more times. If you have any ideas, please contact me or open an issue.

MethodInput SizeCommonChallengeFull setTraining Data
VGG-Shadow(With Dropout)70 * 605.6610.826.67300W
Mobilenet-v2-stage164 * 646.0710.606.96300W and Menpo
Mobilenet-v2-stage264 * 645.768.936.39300W and Menpo

Dataset

DatasetNumber of images for training
300-W3148
Menpo12006

Result on 300W

图片.png

The ground truth landmarks is donated by white color while the predicted ones blue.

Pre-train Models

The pre-train models can be downloaded from baiduyun or GoogleDisk.

Demo

I write a demo to view the alignment results. Besides, the yaw, row and pitch parameters are estimated by the predicted landmarks. To run the domo, please do:

  1. Download and compile caffe. Compile pycaffe.

  2. Use MTCNN to detect face of the images and save the boundingbox of faces.

  3. Modified the path in demo.py.

  4. Run.

图片.png

Dependence

To use my code to reproduce the results, you need to use my caffe. I have added some useful layers.

Reference:


上一篇:MobileNet-V2

下一篇:Mobilenet-_v1-Mask-RCNN-for-detection

用户评价
全部评价

热门资源

  • Keras-ResNeXt

    Keras ResNeXt Implementation of ResNeXt models...

  • seetafaceJNI

    项目介绍 基于中科院seetaface2进行封装的JAVA...

  • spark-corenlp

    This package wraps Stanford CoreNLP annotators ...

  • capsnet-with-caps...

    CapsNet with capsule-wise convolution Project ...

  • inferno-boilerplate

    This is a very basic boilerplate example for pe...