Skip to content

Commit 451a822

Browse files
committed
repository info
1 parent 1b61483 commit 451a822

File tree

2 files changed

+8
-43
lines changed

2 files changed

+8
-43
lines changed

README.md

Lines changed: 7 additions & 42 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
# Introduction
22

3-
This package provides a Lasagne/Theano-based implementation of the deep
3+
This repository is a fork of the Nathan Sprague implementation of the deep
44
Q-learning algorithm described in:
55

66
[Playing Atari with Deep Reinforcement Learning](http://arxiv.org/abs/1312.5602)
@@ -11,9 +11,7 @@ and
1111

1212
Mnih, Volodymyr, et al. "Human-level control through deep reinforcement learning." Nature 518.7540 (2015): 529-533.
1313

14-
Here is a video showing a trained network playing breakout (using an earlier version of the code):
15-
16-
http://youtu.be/SZ88F82KLX4
14+
We use the DQN algorithm to learn the strategies for Atari games from the RAM state of the machine.
1715

1816
# Dependencies
1917

@@ -22,53 +20,20 @@ Here is a video showing a trained network playing breakout (using an earlier ver
2220
* [Theano](http://deeplearning.net/software/theano/) ([https://github.com/Theano/Theano](https://github.com/Theano/Theano))
2321
* [Lasagne](http://lasagne.readthedocs.org/en/latest/) ([https://github.com/Lasagne/Lasagne](https://github.com/Lasagne/Lasagne)
2422
* [Pylearn2](http://deeplearning.net/software/pylearn2/) ([https://github.com/lisa-lab/pylearn2](https://github.com/lisa-lab/pylearn2))
25-
* [Arcade Learning Environment](http://www.arcadelearningenvironment.org/) ([https://github.com/mgbellemare/Arcade-Learning-Environment](https://github.com/mgbellemare/Arcade-Learning-Environment))
23+
* [Arcade Learning Environment](http://www.arcadelearningenvironment.org/) ([https://github.com/sygi/Arcade-Learning-Environment](https://github.com/sygi/Arcade-Learning-Environment) (we use our fork of ALE which draws state of the RAM))
2624

2725
The script `dep_script.sh` can be used to install all dependencies under Ubuntu.
2826

2927

3028
# Running
3129

32-
Use the scripts `run_nips.py` or `run_nature.py` to start all the necessary processes:
33-
34-
`$ ./run_nips.py --rom breakout`
35-
36-
`$ ./run_nature.py --rom breakout`
37-
38-
The `run_nips.py` script uses parameters consistent with the original
39-
NIPS workshop paper. This code should take 2-4 days to complete. The
40-
`run_nature.py` script uses parameters consistent with the Nature
41-
paper. The final policies should be better, but it will take 6-10
42-
days to finish training.
43-
44-
Either script will store output files in a folder prefixed with the
45-
name of the ROM. Pickled version of the network objects are stored
46-
after every epoch. The file `results.csv` will contain the testing
47-
output. You can plot the progress by executing `plot_results.py`:
48-
49-
`$ python plot_results.py breakout_05-28-17-09_0p00025_0p99/results.csv`
50-
51-
After training completes, you can watch the network play using the
52-
`ale_run_watch.py` script:
30+
TODO: describe all the experiments and how to run the code.
5331

54-
`$ python ale_run_watch.py breakout_05-28-17-09_0p00025_0p99/network_file_99.pkl`
55-
56-
# Performance Tuning
57-
58-
## Theano Configuration
59-
60-
Setting `allow_gc=False` in `THEANO_FLAGS` or in the `.theanorc` file
61-
significantly improves performance at the expense of a slight increase
62-
in memory usage on the GPU.
63-
64-
65-
# Getting Help
32+
# See Also
6633

67-
The [deep Q-learning web-forum](https://groups.google.com/forum/#!forum/deep-q-learning)
68-
can be used for discussion and advice related to deep Q-learning in
69-
general and this package in particular.
34+
* https://github.com/spragunr/deep_q_rl
7035

71-
# See Also
36+
Original Nathan Sprague implementation of DQN.
7237

7338
* https://sites.google.com/a/deepmind.com/dqn
7439

dep_script.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ echo "==>installing ALE ..."
4141
# dependencies ...
4242
sudo apt-get install libsdl1.2-dev libsdl-gfx1.2-dev libsdl-image1.2-dev cmake
4343

44-
git clone https://github.com/mgbellemare/Arcade-Learning-Environment ALE
44+
git clone https://github.com/sygi/Arcade-Learning-Environment ALE
4545
cd ./ALE
4646
cmake -DUSE_SDL=ON -DUSE_RLGLUE=OFF .
4747
make -j2

0 commit comments

Comments
 (0)