资源算法Attentive Recurrent Comparators

Attentive Recurrent Comparators

2019-09-17 | |  73 |   0 |   0

arc-pytorch

PyTorch implementation of Attentive Recurrent Comparators by Shyam et al.

blog explaining Attentive Recurrent Comparators

Visualizing Attention

On Same characters

 

On Different Characters

 

How to run?

Download data

python download_data.py

A one-time 52MB download. Shouldn't take more than a few minutes.

Train

python train.py --cuda

Let it train until the accuracy rises to at least 80%. Early stopping is not implemented yet. You will have to manually kill the process.

Visualize

python viz.py --cuda --load 0.13591022789478302 --same

Run with exactly the same parameters as train.py and specify the model to load. Specify "--same" if you want to generate a sample with same characters in both images. The script dumps images to a directory in visualization. The name of directory is taken from --name parameter if specified, else name is a function of the parameters of network.


上一篇:convnet-aig

下一篇:FractalNet

用户评价
全部评价

热门资源

  • seetafaceJNI

    项目介绍 基于中科院seetaface2进行封装的JAVA...

  • Keras-ResNeXt

    Keras ResNeXt Implementation of ResNeXt models...

  • spark-corenlp

    This package wraps Stanford CoreNLP annotators ...

  • shih-styletransfer

    shih-styletransfer Code from Style Transfer ...

  • inferno-boilerplate

    This is a very basic boilerplate example for pe...