variational-autoencoder
Chainer implementation of Variational AutoEncoder(VAE) model M1, M2, M1+M2
Chainer 1.8+
sklearn
To visualize results, you need
matplotlib.patches
PIL
pandas
run mnist-tools.py
to download and extract MNIST.
You can provide label information by filename.
format:
{label_id}_{unique_filename}.{extension}
regex:
([0-9]+)_.+.(bmp|png|jpg)
e.g. MNIST
| params | value | |:-----------|------------:| | OS | Windows 7 | | GPU | GeForce GTX 970M | | ndim_z | 2 | | encoder_apply_dropout | False | | decoder_apply_dropout | False | | encoder_apply_batchnorm | True | | decoder_apply_batchnorm | True | | encoder_apply_batchnorm_to_input | True | | decoder_apply_batchnorm_to_input | True | | encoder_units | [600, 600] | | decoder_units | [600, 600] | | gradient_clipping | 1.0 | | learning_rate | 0.0003 | | gradient_momentum | 0.9 | | gradient_clipping | 1.0 | | nonlinear | softplus|
| params | value | |:-----------|------------:| | OS | Windows 7 | | GPU | GeForce GTX 970M | | ndim_z | 50 | | encoder_xy_z_apply_dropout | False | | encoder_x_y_apply_dropout | False | | decoder_apply_dropout | False | | encoder_xy_z_apply_batchnorm_to_input | True | | encoder_x_y_apply_batchnorm_to_input | True | | decoder_apply_batchnorm_to_input | True | | encoder_xy_z_apply_batchnorm | True | | encoder_x_y_apply_batchnorm | True | | decoder_apply_batchnorm | True | | encoder_xy_z_hidden_units | [500] | | encoder_x_y_hidden_units | [500] | | decoder_hidden_units | [500] | | batchnorm_before_activation | True | | gradient_clipping | 5.0 | | learning_rate | 0.0003 | | gradient_momentum | 0.9 | | gradient_clipping | 1.0 | | nonlinear | softplus|
| data | # | |:-----------|------------:| | labeled | 100 | | unlabeled | 49900 | | validation | 10000 | | test | 10000 |
| * | # | |:-----------|------------:| | epochs | 490 | | minutes | 1412 | | weight updates per epoch | 2000 |
run analogy.py
after training
Model was trained with...
| data | # | |:-----------|------------:| | labeled | 100 | | unlabeled | 49900 |
| data | # | |:-----------|------------:| | labeled | 10000 | | unlabeled | 40000 |
| data | # | |:-----------|------------:| | labeled | 50000 | | unlabeled | 0 |
| params | value | |:-----------|------------:| | OS | Windows 7 | | GPU | GeForce GTX 970M | | ndim_z | 2 | | encoder_apply_dropout | False | | decoder_apply_dropout | False | | encoder_apply_batchnorm | True | | decoder_apply_batchnorm | True | | encoder_apply_batchnorm_to_input | True | | decoder_apply_batchnorm_to_input | True | | encoder_units | [600, 600] | | decoder_units | [600, 600] | | gradient_clipping | 1.0 | | learning_rate | 0.0003 | | gradient_momentum | 0.9 | | gradient_clipping | 1.0 | | nonlinear | softplus|
We trained M1 for 500 epochs before starting training of M2.
| * | # | |:-----------|------------:| | epochs | 500 | | minutes | 860 | | weight updates per epoch | 2000 |
| params | value | |:-----------|------------:| | OS | Windows 7 | | GPU | GeForce GTX 970M | | ndim_z | 50 | | encoder_xy_z_apply_dropout | False | | encoder_x_y_apply_dropout | False | | decoder_apply_dropout | False | | encoder_xy_z_apply_batchnorm_to_input | True | | encoder_x_y_apply_batchnorm_to_input | True | | decoder_apply_batchnorm_to_input | True | | encoder_xy_z_apply_batchnorm | True | | encoder_x_y_apply_batchnorm | True | | decoder_apply_batchnorm | True | | encoder_xy_z_hidden_units | [500] | | encoder_x_y_hidden_units | [500] | | decoder_hidden_units | [500] | | batchnorm_before_activation | True | | gradient_clipping | 5.0 | | learning_rate | 0.0003 | | gradient_momentum | 0.9 | | gradient_clipping | 1.0 | | nonlinear | softplus|
| data | # | |:-----------|------------:| | labeled | 100 | | unlabeled | 49900 | | validation | 10000 | | test | 10000 |
| * | # | |:-----------|------------:| | epochs | 600 | | minutes | 4920 | | weight updates per epoch | 5000 |
seed1: 0.954
seed2: 0.951
上一篇:mia
还没有评论,说两句吧!
热门资源
Keras-ResNeXt
Keras ResNeXt Implementation of ResNeXt models...
seetafaceJNI
项目介绍 基于中科院seetaface2进行封装的JAVA...
spark-corenlp
This package wraps Stanford CoreNLP annotators ...
capsnet-with-caps...
CapsNet with capsule-wise convolution Project ...
inferno-boilerplate
This is a very basic boilerplate example for pe...
智能在线
400-630-6780
聆听.建议反馈
E-mail: support@tusaishared.com