Atari, Dockerfile, PPO
New features and improvements
- some code cleanup to prepare for the next version
- DQN Atari working, not optimized yet
- Dockerfile finished, ready to run lab at scale on server
- implemented PPO in tensorflow from OpenAI, along with the utils