@@ -126,7 +126,9 @@ If you would like to contribute environments, please see our
126
126
* Vector Action space: (Discrete) Size of 6, corresponding to turn clockwise
127
127
and counterclockwise and move along four different face directions.
128
128
* Visual Observations (Optional): One first-person camera. Use
129
- ` VisualPushBlock ` scene.
129
+ ` VisualPushBlock ` scene. __ The visual observation version of
130
+ this environment does not train with the provided default
131
+ training parameters.__
130
132
* Reset Parameters: None.
131
133
* Benchmark Mean Reward: 4.5
132
134
* Optional Imitation Learning scene: ` PushBlockIL ` .
@@ -219,7 +221,9 @@ If you would like to contribute environments, please see our
219
221
* Rotation (3 possible actions: Rotate Left, Rotate Right, No Action)
220
222
* Laser (2 possible actions: Laser, No Action)
221
223
* Visual Observations (Optional): First-person camera per-agent. Use
222
- ` VisualBanana ` scene.
224
+ ` VisualBanana ` scene. __ The visual observation version of
225
+ this environment does not train with the provided default
226
+ training parameters.__
223
227
* Reset Parameters: None.
224
228
* Benchmark Mean Reward: 10
225
229
* Optional Imitation Learning scene: ` BananaIL ` .
@@ -243,7 +247,9 @@ If you would like to contribute environments, please see our
243
247
* Vector Action space: (Discrete) 1 Branch, 4 actions corresponding to agent
244
248
rotation and forward/backward movement.
245
249
* Visual Observations (Optional): First-person view for the agent. Use
246
- ` VisualHallway ` scene.
250
+ ` VisualHallway ` scene. __ The visual observation version of
251
+ this environment does not train with the provided default
252
+ training parameters.__
247
253
* Reset Parameters: None.
248
254
* Benchmark Mean Reward: 0.7
249
255
* To speed up training, you can enable curiosity by adding ` use_curiosity: true ` in ` config/trainer_config.yaml `
@@ -300,7 +306,7 @@ If you would like to contribute environments, please see our
300
306
* Visual Observations: None.
301
307
* Reset Parameters: None
302
308
* Benchmark Mean Reward (Striker & Goalie Brain): 0 (the means will be inverse
303
- of each other and criss crosses during training)
309
+ of each other and criss crosses during training) __ Note that our trainer is currently unable to consistently train this environment __
304
310
305
311
## Walker
306
312
@@ -345,7 +351,9 @@ If you would like to contribute environments, please see our
345
351
* Vector Action space: (Discrete) 4 corresponding to agent rotation and
346
352
forward/backward movement.
347
353
* Visual Observations (Optional): First-person camera per-agent. Us
348
- ` VisualPyramids ` scene.
354
+ ` VisualPyramids ` scene. __ The visual observation version of
355
+ this environment does not train with the provided default
356
+ training parameters.__
349
357
* Reset Parameters: None.
350
358
* Optional Imitation Learning scene: ` PyramidsIL ` .
351
359
* Benchmark Mean Reward: 1.75
0 commit comments