资源算法DQN

DQN

2019-09-17 | |  66 |   0 |   0

dqn

This is a very basic DQN (with experience replay) implementation, which uses OpenAI's gym environment and Keras/Theano neural networks.

Requirements

  • gym

  • keras

  • theano

  • numpy

and all their dependencies.

Usage

To run, python example.py <env_name>. It runs MsPacman-v0 if no env is specified. Uncomment the env.render() line to see the game while training, however, this is likely to make training slow.

Currently, it assumes that the observation is an image, i.e. a 3d array, which is the case for all Atari games, and other Atari-like environments.

Purpose

This is meant to be a very simple implementation, to be used as a starter code. I aimed it to be easy-to-comprehend rather than feature-complete.

Pull requests welcome!

References

  • https://www.cs.toronto.edu/~vmnih/docs/dqn.pdf

TODO

  • Extend to other environemnts. Currently only works for Atari and Atari-like environments where the observation space is a 3D Box.

上一篇:GradientEpisodicMemory

下一篇:fcn

用户评价
全部评价

热门资源

  • Keras-ResNeXt

    Keras ResNeXt Implementation of ResNeXt models...

  • seetafaceJNI

    项目介绍 基于中科院seetaface2进行封装的JAVA...

  • spark-corenlp

    This package wraps Stanford CoreNLP annotators ...

  • capsnet-with-caps...

    CapsNet with capsule-wise convolution Project ...

  • inferno-boilerplate

    This is a very basic boilerplate example for pe...