Basic DQN implementation
This is a very basic DQN (with experience replay) implementation, which uses OpenAI's gym environment and Keras/Theano neural networks.
and all their dependencies.
python example.py. It runs
MsPacman-v0if no env is specified. Uncomment the
env.render()line to see the game while training, however, this is likely to make training slow.
Currently, it assumes that the observation is an image, i.e. a 3d array, which is the case for all Atari games, and other Atari-like environments.
This is meant to be a very simple implementation, to be used as a starter code. I aimed it to be easy-to-comprehend rather than feature-complete.
Pull requests welcome!