Commit ·
fef09a8
1
Parent(s): 8f5b016
Environment Solved
Browse files
README.md
CHANGED
|
@@ -53,4 +53,11 @@ python -m sf_examples.atari.train_atari --algo=APPO --env=atari_atlantis --train
|
|
| 53 |
```
|
| 54 |
|
| 55 |
Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
|
| 56 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
```
|
| 54 |
|
| 55 |
Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
|
| 56 |
+
|
| 57 |
+
## SOTA Performance
|
| 58 |
+
|
| 59 |
+
This model as with all the others was trained at 10 million steps to create a baseline. Interestingly, in this environment, it reaches SOTA performance at even this level suggesting that the Atlantis game is pretty easy to beat.
|
| 60 |
+
|
| 61 |
+
For more information on this environment see: https://www.endtoend.ai/envs/gym/atari/atlantis/. Because rewards are plentiful and the Gorgons have to pass 4 times to reach attack range the environment is relatively easy to reach SOTA on.
|
| 62 |
+
|
| 63 |
+
I have now compared this with the performance of the TQC, SAC and the DQN models which all underperformed PPO. I now consider this atari environment solved.
|