You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Dec 11, 2022. It is now read-only.
@@ -52,13 +55,6 @@ Contacting the Coach development team is also possible through the email [coach@
52
55
53
56
One of the main challenges when building a research project, or a solution based on a published algorithm, is getting a concrete and reliable baseline that reproduces the algorithm's results, as reported by its authors. To address this problem, we are releasing a set of [benchmarks](benchmarks) that shows Coach reliably reproduces many state of the art algorithm results.
54
57
55
-
## Documentation
56
-
57
-
Framework documentation, algorithm description and instructions on how to contribute a new agent/environment can be found [here](https://nervanasystems.github.io/coach/).
58
-
59
-
Jupyter notebooks demonstrating how to run Coach from command line or as a library, implement an algorithm, or integrate an environment can be found [here](https://github.com/NervanaSystems/coach/tree/master/tutorials).
60
-
61
-
62
58
## Installation
63
59
64
60
Note: Coach has only been tested on Ubuntu 16.04 LTS, and with Python 3.5.
@@ -113,9 +109,16 @@ If a GPU is present, Coach's pip package will install tensorflow-gpu, by default
113
109
114
110
In addition to OpenAI Gym, several other environments were tested and are supported. Please follow the instructions in the Supported Environments section below in order to install more environments.
115
111
116
-
## Usage
112
+
## Getting Started
113
+
114
+
### Tutorials and Documentation
115
+
[Jupyter notebooks demonstrating how to run Coach from command line or as a library, implement an algorithm, or integrate an environment](https://github.com/NervanaSystems/coach/tree/master/tutorials).
116
+
117
+
[Framework documentation, algorithm description and instructions on how to contribute a new agent/environment](https://nervanasystems.github.io/coach/).
118
+
119
+
### Basic Usage
117
120
118
-
### Running Coach
121
+
####Running Coach
119
122
120
123
To allow reproducing results in Coach, we defined a mechanism called _preset_.
121
124
There are several available presets under the `presets` directory.
@@ -167,17 +170,7 @@ It is easy to create new presets for different levels or environments by followi
167
170
168
171
More usage examples can be found [here](https://github.com/NervanaSystems/coach/blob/master/tutorials/0.%20Quick%20Start%20Guide.ipynb).
169
172
170
-
### Distributed Multi-Node Coach
171
-
172
-
As of release 0.11.0, Coach supports horizontal scaling for training RL agents on multiple nodes. In release 0.11.0 this was tested on the ClippedPPO and DQN agents.
173
-
For usage instructions please refer to the documentation [here](https://nervanasystems.github.io/coach/dist_usage.html).
174
-
175
-
### Batch Reinforcement Learning
176
-
177
-
Training and evaluating an agent from a dataset of experience, where no simulator is available, is supported in Coach.
178
-
There are [example](https://github.com/NervanaSystems/coach/blob/master/rl_coach/presets/CartPole_DDQN_BatchRL.py)[presets](https://github.com/NervanaSystems/coach/blob/master/rl_coach/presets/Acrobot_DDQN_BCQ_BatchRL.py) and a [tutorial](https://github.com/NervanaSystems/coach/blob/master/tutorials/4.%20Batch%20Reinforcement%20Learning.ipynb).
179
-
180
-
### Running Coach Dashboard (Visualization)
173
+
#### Running Coach Dashboard (Visualization)
181
174
Training an agent to solve an environment can be tricky, at times.
182
175
183
176
In order to debug the training process, Coach outputs several signals, per trained algorithm, in order to track algorithmic performance.
As of release 0.11.0, Coach supports horizontal scaling for training RL agents on multiple nodes. In release 0.11.0 this was tested on the ClippedPPO and DQN agents.
194
+
For usage instructions please refer to the documentation [here](https://nervanasystems.github.io/coach/dist_usage.html).
195
+
196
+
### Batch Reinforcement Learning
197
+
198
+
Training and evaluating an agent from a dataset of experience, where no simulator is available, is supported in Coach.
199
+
There are [example](https://github.com/NervanaSystems/coach/blob/master/rl_coach/presets/CartPole_DDQN_BatchRL.py)[presets](https://github.com/NervanaSystems/coach/blob/master/rl_coach/presets/Acrobot_DDQN_BCQ_BatchRL.py) and a [tutorial](https://github.com/NervanaSystems/coach/blob/master/tutorials/4.%20Batch%20Reinforcement%20Learning.ipynb).
0 commit comments