First Push
Browse files- README.md +14 -8
- SnowballTarget.onnx +1 -1
- SnowballTarget/SnowballTarget-159936.onnx +3 -0
- SnowballTarget/SnowballTarget-159936.pt +3 -0
- SnowballTarget/SnowballTarget-164992.onnx +3 -0
- SnowballTarget/SnowballTarget-164992.pt +3 -0
- SnowballTarget/SnowballTarget-169976.onnx +3 -0
- SnowballTarget/SnowballTarget-169976.pt +3 -0
- SnowballTarget/SnowballTarget-174952.onnx +3 -0
- SnowballTarget/SnowballTarget-174952.pt +3 -0
- SnowballTarget/SnowballTarget-179992.onnx +3 -0
- SnowballTarget/SnowballTarget-179992.pt +3 -0
- SnowballTarget/SnowballTarget-184992.onnx +3 -0
- SnowballTarget/SnowballTarget-184992.pt +3 -0
- SnowballTarget/SnowballTarget-189968.onnx +3 -0
- SnowballTarget/SnowballTarget-189968.pt +3 -0
- SnowballTarget/SnowballTarget-194944.onnx +3 -0
- SnowballTarget/SnowballTarget-194944.pt +3 -0
- SnowballTarget/SnowballTarget-199984.onnx +3 -0
- SnowballTarget/SnowballTarget-199984.pt +3 -0
- SnowballTarget/SnowballTarget-200112.onnx +3 -0
- SnowballTarget/SnowballTarget-200112.pt +3 -0
- SnowballTarget/checkpoint.pt +1 -1
- SnowballTarget/events.out.tfevents.1686583121.7c4b9d5ff59a.1945.0 +3 -0
- config.json +1 -1
- configuration.yaml +5 -5
- run_logs/Player-0.log +13 -13
- run_logs/timers.json +165 -165
- run_logs/training_status.json +86 -14
README.md
CHANGED
|
@@ -8,21 +8,27 @@ tags:
|
|
| 8 |
---
|
| 9 |
|
| 10 |
# **ppo** Agent playing **SnowballTarget**
|
| 11 |
-
This is a trained model of a **ppo** agent playing **SnowballTarget**
|
| 12 |
-
|
|
|
|
| 13 |
## Usage (with ML-Agents)
|
| 14 |
-
The Documentation: https://github.
|
| 15 |
-
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
|
| 16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
### Resume the training
|
| 19 |
-
```
|
| 20 |
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
|
| 21 |
```
|
|
|
|
| 22 |
### Watch your Agent play
|
| 23 |
-
You can watch your agent **playing directly in your browser
|
| 24 |
-
|
| 25 |
-
1.
|
| 26 |
2. Step 1: Find your model_id: Jeibros/ppo-SnowballTarget
|
| 27 |
3. Step 2: Select your *.nn /*.onnx file
|
| 28 |
4. Click on Watch the agent play 👀
|
|
|
|
| 8 |
---
|
| 9 |
|
| 10 |
# **ppo** Agent playing **SnowballTarget**
|
| 11 |
+
This is a trained model of a **ppo** agent playing **SnowballTarget**
|
| 12 |
+
using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
|
| 13 |
+
|
| 14 |
## Usage (with ML-Agents)
|
| 15 |
+
The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
|
|
|
|
| 16 |
|
| 17 |
+
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
|
| 18 |
+
- A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
|
| 19 |
+
browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
|
| 20 |
+
- A *longer tutorial* to understand how works ML-Agents:
|
| 21 |
+
https://huggingface.co/learn/deep-rl-course/unit5/introduction
|
| 22 |
|
| 23 |
### Resume the training
|
| 24 |
+
```bash
|
| 25 |
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
|
| 26 |
```
|
| 27 |
+
|
| 28 |
### Watch your Agent play
|
| 29 |
+
You can watch your agent **playing directly in your browser**
|
| 30 |
+
|
| 31 |
+
1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
|
| 32 |
2. Step 1: Find your model_id: Jeibros/ppo-SnowballTarget
|
| 33 |
3. Step 2: Select your *.nn /*.onnx file
|
| 34 |
4. Click on Watch the agent play 👀
|
SnowballTarget.onnx
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 646532
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:633f985e4f98e1d52c9489d09daf84bd770912de0aee7aba76dbde510f7d6e23
|
| 3 |
size 646532
|
SnowballTarget/SnowballTarget-159936.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ed85638f7708034e423b0abf1f3b81a2db321668093f34c8f1574b543c2ff00
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-159936.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:18a929e10e2fad1d862de661b55855642fb220c2918b3f05449ac7356c85831e
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-164992.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6304e9a854f8528c3596962a322ff78eeb09f4a3cb14e9ad534e1898356b62d0
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-164992.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:acb69b88dc7bd28d8ecb635ec403c245e468e9842be3eb49b884e1f7e7cca6fe
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-169976.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a7f64d7ec112340eeee5f350345f6b30853db2d7bb5117d05b72e1d309fe19d0
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-169976.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:34e31bb00b75d55216ec78d23b8ecb1b6c070fdb0e83c079bba62ded9597d778
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-174952.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:fc21dde0d963d2a98395b4b027b29c7b28c0eb4fdebc5a911f753d4ebe0f50f4
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-174952.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:604c9a0ac1b0e79d131ee2d19b8b75d2005322fe2c2d477e16292e3f317a5b6f
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-179992.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a43ad048744271848239b5b3ee11f71addeb2f62d3dab777a0f1e07abbb8164d
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-179992.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:bb6838e47bb06785e762675e91a9c6f66e6b9768a6bc4e83ffc18d5333c0c899
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-184992.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b4ca4be6c6304b6609f1387a40bf61f16370af2fdb247e06a4cb6a7a6c4e94f7
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-184992.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:66dbb30638e0a55769e5311da7b61f2335c3ef03243499b677770b029ea105d6
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-189968.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:03717e967de98a32a8db0ff5498a663e071b45cc0f15a24ff52512f7f168d97b
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-189968.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:74081c21203e9934e5184e97e11282506f57d25310cda2376095f3d18b4286d5
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-194944.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eb69e46003375a8862cbca5f4470984263b20046720aebcbf4931d357aa91288
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-194944.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e9527991f5601037aa584dfab762eb226309821e75773c16b4e7b6ac048927e9
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-199984.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:633f985e4f98e1d52c9489d09daf84bd770912de0aee7aba76dbde510f7d6e23
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-199984.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6b7fe56cc55a1a03b3a4f7ca9babc8da823432ea667706df762c5b660fbeb2d9
|
| 3 |
+
size 3843968
|
SnowballTarget/SnowballTarget-200112.onnx
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:633f985e4f98e1d52c9489d09daf84bd770912de0aee7aba76dbde510f7d6e23
|
| 3 |
+
size 646532
|
SnowballTarget/SnowballTarget-200112.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8d9db1d5668e126c4785ec5d985db9a85319e4dac70b925a3dbd39d9cf1ba83f
|
| 3 |
+
size 3843968
|
SnowballTarget/checkpoint.pt
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3843968
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8d9db1d5668e126c4785ec5d985db9a85319e4dac70b925a3dbd39d9cf1ba83f
|
| 3 |
size 3843968
|
SnowballTarget/events.out.tfevents.1686583121.7c4b9d5ff59a.1945.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e2356caafceac69c4e24fe65704e740bc09376ea83da92aee1f638f2f72985e7
|
| 3 |
+
size 28090
|
config.json
CHANGED
|
@@ -1 +1 @@
|
|
| 1 |
-
{"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval":
|
|
|
|
| 1 |
+
{"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 5000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20.0, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
|
configuration.yaml
CHANGED
|
@@ -14,7 +14,7 @@ behaviors:
|
|
| 14 |
learning_rate_schedule: linear
|
| 15 |
beta_schedule: linear
|
| 16 |
epsilon_schedule: linear
|
| 17 |
-
checkpoint_interval:
|
| 18 |
network_settings:
|
| 19 |
normalize: false
|
| 20 |
hidden_units: 256
|
|
@@ -36,11 +36,11 @@ behaviors:
|
|
| 36 |
goal_conditioning_type: hyper
|
| 37 |
deterministic: false
|
| 38 |
init_path: null
|
| 39 |
-
keep_checkpoints:
|
| 40 |
even_checkpoints: false
|
| 41 |
-
max_steps:
|
| 42 |
time_horizon: 64
|
| 43 |
-
summary_freq:
|
| 44 |
threaded: true
|
| 45 |
self_play: null
|
| 46 |
behavioral_cloning: null
|
|
@@ -58,7 +58,7 @@ engine_settings:
|
|
| 58 |
width: 84
|
| 59 |
height: 84
|
| 60 |
quality_level: 5
|
| 61 |
-
time_scale: 20
|
| 62 |
target_frame_rate: -1
|
| 63 |
capture_frame_rate: 60
|
| 64 |
no_graphics: true
|
|
|
|
| 14 |
learning_rate_schedule: linear
|
| 15 |
beta_schedule: linear
|
| 16 |
epsilon_schedule: linear
|
| 17 |
+
checkpoint_interval: 5000
|
| 18 |
network_settings:
|
| 19 |
normalize: false
|
| 20 |
hidden_units: 256
|
|
|
|
| 36 |
goal_conditioning_type: hyper
|
| 37 |
deterministic: false
|
| 38 |
init_path: null
|
| 39 |
+
keep_checkpoints: 10
|
| 40 |
even_checkpoints: false
|
| 41 |
+
max_steps: 200000
|
| 42 |
time_horizon: 64
|
| 43 |
+
summary_freq: 10000
|
| 44 |
threaded: true
|
| 45 |
self_play: null
|
| 46 |
behavioral_cloning: null
|
|
|
|
| 58 |
width: 84
|
| 59 |
height: 84
|
| 60 |
quality_level: 5
|
| 61 |
+
time_scale: 20.0
|
| 62 |
target_frame_rate: -1
|
| 63 |
capture_frame_rate: 60
|
| 64 |
no_graphics: true
|
run_logs/Player-0.log
CHANGED
|
@@ -34,7 +34,7 @@ ALSA lib pcm.c:2642:(snd_pcm_open_noupdate) Unknown PCM default
|
|
| 34 |
FMOD failed to initialize the output device.: "Error initializing output device. " (60)
|
| 35 |
FMOD initialized on nosound output
|
| 36 |
Begin MonoManager ReloadAssembly
|
| 37 |
-
- Completed reload, in 0.
|
| 38 |
ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 39 |
ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 40 |
ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
|
@@ -45,7 +45,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
|
|
| 45 |
WARNING: Shader Unsupported: 'Standard' - All subshaders removed
|
| 46 |
WARNING: Shader Did you use #pragma only_renderers and omit this platform?
|
| 47 |
WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
|
| 48 |
-
UnloadTime:
|
| 49 |
ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 50 |
requesting resize 84 x 84
|
| 51 |
Setting up 1 worker threads for Enlighten.
|
|
@@ -53,7 +53,7 @@ Memory Statistics:
|
|
| 53 |
[ALLOC_TEMP_TLS] TLS Allocator
|
| 54 |
StackAllocators :
|
| 55 |
[ALLOC_TEMP_MAIN]
|
| 56 |
-
Peak usage frame count: [8.0 KB-16.0 KB]:
|
| 57 |
Initial Block Size 4.0 MB
|
| 58 |
Current Block Size 4.0 MB
|
| 59 |
Peak Allocated Bytes 2.0 MB
|
|
@@ -169,19 +169,19 @@ Memory Statistics:
|
|
| 169 |
Peak Allocated Bytes 0 B
|
| 170 |
Overflow Count 0
|
| 171 |
[ALLOC_DEFAULT] Dual Thread Allocator
|
| 172 |
-
Peak main deferred allocation count
|
| 173 |
[ALLOC_BUCKET]
|
| 174 |
Large Block size 4.0 MB
|
| 175 |
Used Block count 1
|
| 176 |
Peak Allocated bytes 1.0 MB
|
| 177 |
[ALLOC_DEFAULT_MAIN]
|
| 178 |
-
Peak usage frame count: [4.0 MB-8.0 MB]:
|
| 179 |
Requested Block Size 16.0 MB
|
| 180 |
Peak Block count 1
|
| 181 |
-
Peak Allocated memory
|
| 182 |
Peak Large allocation bytes 0 B
|
| 183 |
[ALLOC_DEFAULT_THREAD]
|
| 184 |
-
Peak usage frame count: [16.0 MB-32.0 MB]:
|
| 185 |
Requested Block Size 16.0 MB
|
| 186 |
Peak Block count 1
|
| 187 |
Peak Allocated memory 17.9 MB
|
|
@@ -213,13 +213,13 @@ Memory Statistics:
|
|
| 213 |
Used Block count 1
|
| 214 |
Peak Allocated bytes 1.0 MB
|
| 215 |
[ALLOC_GFX_MAIN]
|
| 216 |
-
Peak usage frame count: [32.0 KB-64.0 KB]:
|
| 217 |
Requested Block Size 16.0 MB
|
| 218 |
Peak Block count 1
|
| 219 |
Peak Allocated memory 66.8 KB
|
| 220 |
Peak Large allocation bytes 0 B
|
| 221 |
[ALLOC_GFX_THREAD]
|
| 222 |
-
Peak usage frame count: [32.0 KB-64.0 KB]:
|
| 223 |
Requested Block Size 16.0 MB
|
| 224 |
Peak Block count 1
|
| 225 |
Peak Allocated memory 39.6 KB
|
|
@@ -231,13 +231,13 @@ Memory Statistics:
|
|
| 231 |
Used Block count 1
|
| 232 |
Peak Allocated bytes 1.0 MB
|
| 233 |
[ALLOC_CACHEOBJECTS_MAIN]
|
| 234 |
-
Peak usage frame count: [0.5 MB-1.0 MB]:
|
| 235 |
Requested Block Size 4.0 MB
|
| 236 |
Peak Block count 1
|
| 237 |
Peak Allocated memory 0.6 MB
|
| 238 |
Peak Large allocation bytes 0 B
|
| 239 |
[ALLOC_CACHEOBJECTS_THREAD]
|
| 240 |
-
Peak usage frame count: [0.5 MB-1.0 MB]:
|
| 241 |
Requested Block Size 4.0 MB
|
| 242 |
Peak Block count 1
|
| 243 |
Peak Allocated memory 2.2 MB
|
|
@@ -249,13 +249,13 @@ Memory Statistics:
|
|
| 249 |
Used Block count 1
|
| 250 |
Peak Allocated bytes 1.0 MB
|
| 251 |
[ALLOC_TYPETREE_MAIN]
|
| 252 |
-
Peak usage frame count: [0-1.0 KB]:
|
| 253 |
Requested Block Size 2.0 MB
|
| 254 |
Peak Block count 1
|
| 255 |
Peak Allocated memory 1.0 KB
|
| 256 |
Peak Large allocation bytes 0 B
|
| 257 |
[ALLOC_TYPETREE_THREAD]
|
| 258 |
-
Peak usage frame count: [1.0 KB-2.0 KB]:
|
| 259 |
Requested Block Size 2.0 MB
|
| 260 |
Peak Block count 1
|
| 261 |
Peak Allocated memory 1.7 KB
|
|
|
|
| 34 |
FMOD failed to initialize the output device.: "Error initializing output device. " (60)
|
| 35 |
FMOD initialized on nosound output
|
| 36 |
Begin MonoManager ReloadAssembly
|
| 37 |
+
- Completed reload, in 0.136 seconds
|
| 38 |
ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 39 |
ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 40 |
ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
|
|
|
| 45 |
WARNING: Shader Unsupported: 'Standard' - All subshaders removed
|
| 46 |
WARNING: Shader Did you use #pragma only_renderers and omit this platform?
|
| 47 |
WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
|
| 48 |
+
UnloadTime: 1.011150 ms
|
| 49 |
ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
|
| 50 |
requesting resize 84 x 84
|
| 51 |
Setting up 1 worker threads for Enlighten.
|
|
|
|
| 53 |
[ALLOC_TEMP_TLS] TLS Allocator
|
| 54 |
StackAllocators :
|
| 55 |
[ALLOC_TEMP_MAIN]
|
| 56 |
+
Peak usage frame count: [8.0 KB-16.0 KB]: 5377 frames, [16.0 KB-32.0 KB]: 91 frames, [2.0 MB-4.0 MB]: 1 frames
|
| 57 |
Initial Block Size 4.0 MB
|
| 58 |
Current Block Size 4.0 MB
|
| 59 |
Peak Allocated Bytes 2.0 MB
|
|
|
|
| 169 |
Peak Allocated Bytes 0 B
|
| 170 |
Overflow Count 0
|
| 171 |
[ALLOC_DEFAULT] Dual Thread Allocator
|
| 172 |
+
Peak main deferred allocation count 44
|
| 173 |
[ALLOC_BUCKET]
|
| 174 |
Large Block size 4.0 MB
|
| 175 |
Used Block count 1
|
| 176 |
Peak Allocated bytes 1.0 MB
|
| 177 |
[ALLOC_DEFAULT_MAIN]
|
| 178 |
+
Peak usage frame count: [4.0 MB-8.0 MB]: 5469 frames
|
| 179 |
Requested Block Size 16.0 MB
|
| 180 |
Peak Block count 1
|
| 181 |
+
Peak Allocated memory 6.6 MB
|
| 182 |
Peak Large allocation bytes 0 B
|
| 183 |
[ALLOC_DEFAULT_THREAD]
|
| 184 |
+
Peak usage frame count: [16.0 MB-32.0 MB]: 5469 frames
|
| 185 |
Requested Block Size 16.0 MB
|
| 186 |
Peak Block count 1
|
| 187 |
Peak Allocated memory 17.9 MB
|
|
|
|
| 213 |
Used Block count 1
|
| 214 |
Peak Allocated bytes 1.0 MB
|
| 215 |
[ALLOC_GFX_MAIN]
|
| 216 |
+
Peak usage frame count: [32.0 KB-64.0 KB]: 5056 frames, [64.0 KB-128.0 KB]: 413 frames
|
| 217 |
Requested Block Size 16.0 MB
|
| 218 |
Peak Block count 1
|
| 219 |
Peak Allocated memory 66.8 KB
|
| 220 |
Peak Large allocation bytes 0 B
|
| 221 |
[ALLOC_GFX_THREAD]
|
| 222 |
+
Peak usage frame count: [32.0 KB-64.0 KB]: 5469 frames
|
| 223 |
Requested Block Size 16.0 MB
|
| 224 |
Peak Block count 1
|
| 225 |
Peak Allocated memory 39.6 KB
|
|
|
|
| 231 |
Used Block count 1
|
| 232 |
Peak Allocated bytes 1.0 MB
|
| 233 |
[ALLOC_CACHEOBJECTS_MAIN]
|
| 234 |
+
Peak usage frame count: [0.5 MB-1.0 MB]: 5469 frames
|
| 235 |
Requested Block Size 4.0 MB
|
| 236 |
Peak Block count 1
|
| 237 |
Peak Allocated memory 0.6 MB
|
| 238 |
Peak Large allocation bytes 0 B
|
| 239 |
[ALLOC_CACHEOBJECTS_THREAD]
|
| 240 |
+
Peak usage frame count: [0.5 MB-1.0 MB]: 5468 frames, [2.0 MB-4.0 MB]: 1 frames
|
| 241 |
Requested Block Size 4.0 MB
|
| 242 |
Peak Block count 1
|
| 243 |
Peak Allocated memory 2.2 MB
|
|
|
|
| 249 |
Used Block count 1
|
| 250 |
Peak Allocated bytes 1.0 MB
|
| 251 |
[ALLOC_TYPETREE_MAIN]
|
| 252 |
+
Peak usage frame count: [0-1.0 KB]: 5469 frames
|
| 253 |
Requested Block Size 2.0 MB
|
| 254 |
Peak Block count 1
|
| 255 |
Peak Allocated memory 1.0 KB
|
| 256 |
Peak Large allocation bytes 0 B
|
| 257 |
[ALLOC_TYPETREE_THREAD]
|
| 258 |
+
Peak usage frame count: [1.0 KB-2.0 KB]: 5469 frames
|
| 259 |
Requested Block Size 2.0 MB
|
| 260 |
Peak Block count 1
|
| 261 |
Peak Allocated memory 1.7 KB
|
run_logs/timers.json
CHANGED
|
@@ -2,213 +2,213 @@
|
|
| 2 |
"name": "root",
|
| 3 |
"gauges": {
|
| 4 |
"SnowballTarget.Policy.Entropy.mean": {
|
| 5 |
-
"value": 0.
|
| 6 |
-
"min": 0.
|
| 7 |
-
"max": 2.
|
| 8 |
-
"count":
|
| 9 |
},
|
| 10 |
"SnowballTarget.Policy.Entropy.sum": {
|
| 11 |
-
"value":
|
| 12 |
-
"min":
|
| 13 |
-
"max":
|
| 14 |
-
"count":
|
| 15 |
},
|
| 16 |
"SnowballTarget.Step.mean": {
|
| 17 |
-
"value":
|
| 18 |
-
"min":
|
| 19 |
-
"max":
|
| 20 |
-
"count":
|
| 21 |
},
|
| 22 |
"SnowballTarget.Step.sum": {
|
| 23 |
-
"value":
|
| 24 |
-
"min":
|
| 25 |
-
"max":
|
| 26 |
-
"count":
|
| 27 |
},
|
| 28 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
|
| 29 |
-
"value": 13.
|
| 30 |
-
"min":
|
| 31 |
-
"max": 13.
|
| 32 |
-
"count":
|
| 33 |
},
|
| 34 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
|
| 35 |
-
"value":
|
| 36 |
-
"min":
|
| 37 |
-
"max":
|
| 38 |
-
"count":
|
| 39 |
},
|
| 40 |
"SnowballTarget.Environment.EpisodeLength.mean": {
|
| 41 |
"value": 199.0,
|
| 42 |
"min": 199.0,
|
| 43 |
"max": 199.0,
|
| 44 |
-
"count":
|
| 45 |
},
|
| 46 |
"SnowballTarget.Environment.EpisodeLength.sum": {
|
| 47 |
-
"value":
|
| 48 |
-
"min":
|
| 49 |
-
"max":
|
| 50 |
-
"count":
|
| 51 |
},
|
| 52 |
"SnowballTarget.Losses.PolicyLoss.mean": {
|
| 53 |
-
"value": 0.
|
| 54 |
-
"min": 0.
|
| 55 |
-
"max": 0.
|
| 56 |
-
"count":
|
| 57 |
},
|
| 58 |
"SnowballTarget.Losses.PolicyLoss.sum": {
|
| 59 |
-
"value":
|
| 60 |
-
"min":
|
| 61 |
-
"max":
|
| 62 |
-
"count":
|
| 63 |
},
|
| 64 |
"SnowballTarget.Losses.ValueLoss.mean": {
|
| 65 |
-
"value": 0.
|
| 66 |
-
"min": 0.
|
| 67 |
-
"max": 0.
|
| 68 |
-
"count":
|
| 69 |
},
|
| 70 |
"SnowballTarget.Losses.ValueLoss.sum": {
|
| 71 |
-
"value":
|
| 72 |
-
"min":
|
| 73 |
-
"max":
|
| 74 |
-
"count":
|
| 75 |
},
|
| 76 |
"SnowballTarget.Policy.LearningRate.mean": {
|
| 77 |
-
"value":
|
| 78 |
-
"min":
|
| 79 |
-
"max": 0.
|
| 80 |
-
"count":
|
| 81 |
},
|
| 82 |
"SnowballTarget.Policy.LearningRate.sum": {
|
| 83 |
-
"value":
|
| 84 |
-
"min":
|
| 85 |
-
"max": 0.
|
| 86 |
-
"count":
|
| 87 |
},
|
| 88 |
"SnowballTarget.Policy.Epsilon.mean": {
|
| 89 |
-
"value": 0.
|
| 90 |
-
"min": 0.
|
| 91 |
-
"max": 0.
|
| 92 |
-
"count":
|
| 93 |
},
|
| 94 |
"SnowballTarget.Policy.Epsilon.sum": {
|
| 95 |
-
"value":
|
| 96 |
-
"min":
|
| 97 |
-
"max":
|
| 98 |
-
"count":
|
| 99 |
},
|
| 100 |
"SnowballTarget.Policy.Beta.mean": {
|
| 101 |
-
"value": 0.
|
| 102 |
-
"min": 0.
|
| 103 |
-
"max": 0.
|
| 104 |
-
"count":
|
| 105 |
},
|
| 106 |
"SnowballTarget.Policy.Beta.sum": {
|
| 107 |
-
"value": 0.
|
| 108 |
-
"min": 0.
|
| 109 |
-
"max": 0.
|
| 110 |
-
"count":
|
| 111 |
},
|
| 112 |
"SnowballTarget.Environment.CumulativeReward.mean": {
|
| 113 |
-
"value":
|
| 114 |
-
"min":
|
| 115 |
-
"max":
|
| 116 |
-
"count":
|
| 117 |
},
|
| 118 |
"SnowballTarget.Environment.CumulativeReward.sum": {
|
| 119 |
-
"value":
|
| 120 |
-
"min":
|
| 121 |
-
"max":
|
| 122 |
-
"count":
|
| 123 |
},
|
| 124 |
"SnowballTarget.Policy.ExtrinsicReward.mean": {
|
| 125 |
-
"value":
|
| 126 |
-
"min":
|
| 127 |
-
"max":
|
| 128 |
-
"count":
|
| 129 |
},
|
| 130 |
"SnowballTarget.Policy.ExtrinsicReward.sum": {
|
| 131 |
-
"value":
|
| 132 |
-
"min":
|
| 133 |
-
"max":
|
| 134 |
-
"count":
|
| 135 |
},
|
| 136 |
"SnowballTarget.IsTraining.mean": {
|
| 137 |
"value": 1.0,
|
| 138 |
"min": 1.0,
|
| 139 |
"max": 1.0,
|
| 140 |
-
"count":
|
| 141 |
},
|
| 142 |
"SnowballTarget.IsTraining.sum": {
|
| 143 |
"value": 1.0,
|
| 144 |
"min": 1.0,
|
| 145 |
"max": 1.0,
|
| 146 |
-
"count":
|
| 147 |
}
|
| 148 |
},
|
| 149 |
"metadata": {
|
| 150 |
"timer_format_version": "0.1.0",
|
| 151 |
-
"start_time_seconds": "
|
| 152 |
-
"python_version": "3.10.
|
| 153 |
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
|
| 154 |
"mlagents_version": "0.31.0.dev0",
|
| 155 |
"mlagents_envs_version": "0.31.0.dev0",
|
| 156 |
"communication_protocol_version": "1.5.0",
|
| 157 |
"pytorch_version": "1.11.0+cu102",
|
| 158 |
"numpy_version": "1.21.2",
|
| 159 |
-
"end_time_seconds": "
|
| 160 |
},
|
| 161 |
-
"total":
|
| 162 |
"count": 1,
|
| 163 |
-
"self": 0.
|
| 164 |
"children": {
|
| 165 |
"run_training.setup": {
|
| 166 |
-
"total": 0.
|
| 167 |
"count": 1,
|
| 168 |
-
"self": 0.
|
| 169 |
},
|
| 170 |
"TrainerController.start_learning": {
|
| 171 |
-
"total":
|
| 172 |
"count": 1,
|
| 173 |
-
"self":
|
| 174 |
"children": {
|
| 175 |
"TrainerController._reset_env": {
|
| 176 |
-
"total":
|
| 177 |
"count": 1,
|
| 178 |
-
"self":
|
| 179 |
},
|
| 180 |
"TrainerController.advance": {
|
| 181 |
-
"total":
|
| 182 |
-
"count":
|
| 183 |
-
"self": 0.
|
| 184 |
"children": {
|
| 185 |
"env_step": {
|
| 186 |
-
"total":
|
| 187 |
-
"count":
|
| 188 |
-
"self":
|
| 189 |
"children": {
|
| 190 |
"SubprocessEnvManager._take_step": {
|
| 191 |
-
"total":
|
| 192 |
-
"count":
|
| 193 |
-
"self":
|
| 194 |
"children": {
|
| 195 |
"TorchPolicy.evaluate": {
|
| 196 |
-
"total":
|
| 197 |
-
"count":
|
| 198 |
-
"self":
|
| 199 |
}
|
| 200 |
}
|
| 201 |
},
|
| 202 |
"workers": {
|
| 203 |
-
"total": 0.
|
| 204 |
-
"count":
|
| 205 |
"self": 0.0,
|
| 206 |
"children": {
|
| 207 |
"worker_root": {
|
| 208 |
-
"total":
|
| 209 |
-
"count":
|
| 210 |
"is_parallel": true,
|
| 211 |
-
"self":
|
| 212 |
"children": {
|
| 213 |
"run_training.setup": {
|
| 214 |
"total": 0.0,
|
|
@@ -217,48 +217,48 @@
|
|
| 217 |
"self": 0.0,
|
| 218 |
"children": {
|
| 219 |
"steps_from_proto": {
|
| 220 |
-
"total": 0.
|
| 221 |
"count": 1,
|
| 222 |
"is_parallel": true,
|
| 223 |
-
"self": 0.
|
| 224 |
"children": {
|
| 225 |
"_process_rank_one_or_two_observation": {
|
| 226 |
-
"total": 0.
|
| 227 |
"count": 10,
|
| 228 |
"is_parallel": true,
|
| 229 |
-
"self": 0.
|
| 230 |
}
|
| 231 |
}
|
| 232 |
},
|
| 233 |
"UnityEnvironment.step": {
|
| 234 |
-
"total": 0.
|
| 235 |
"count": 1,
|
| 236 |
"is_parallel": true,
|
| 237 |
-
"self": 0.
|
| 238 |
"children": {
|
| 239 |
"UnityEnvironment._generate_step_input": {
|
| 240 |
-
"total": 0.
|
| 241 |
"count": 1,
|
| 242 |
"is_parallel": true,
|
| 243 |
-
"self": 0.
|
| 244 |
},
|
| 245 |
"communicator.exchange": {
|
| 246 |
-
"total": 0.
|
| 247 |
"count": 1,
|
| 248 |
"is_parallel": true,
|
| 249 |
-
"self": 0.
|
| 250 |
},
|
| 251 |
"steps_from_proto": {
|
| 252 |
-
"total": 0.
|
| 253 |
"count": 1,
|
| 254 |
"is_parallel": true,
|
| 255 |
-
"self": 0.
|
| 256 |
"children": {
|
| 257 |
"_process_rank_one_or_two_observation": {
|
| 258 |
-
"total": 0.
|
| 259 |
"count": 10,
|
| 260 |
"is_parallel": true,
|
| 261 |
-
"self": 0.
|
| 262 |
}
|
| 263 |
}
|
| 264 |
}
|
|
@@ -267,34 +267,34 @@
|
|
| 267 |
}
|
| 268 |
},
|
| 269 |
"UnityEnvironment.step": {
|
| 270 |
-
"total":
|
| 271 |
-
"count":
|
| 272 |
"is_parallel": true,
|
| 273 |
-
"self":
|
| 274 |
"children": {
|
| 275 |
"UnityEnvironment._generate_step_input": {
|
| 276 |
-
"total":
|
| 277 |
-
"count":
|
| 278 |
"is_parallel": true,
|
| 279 |
-
"self":
|
| 280 |
},
|
| 281 |
"communicator.exchange": {
|
| 282 |
-
"total":
|
| 283 |
-
"count":
|
| 284 |
"is_parallel": true,
|
| 285 |
-
"self":
|
| 286 |
},
|
| 287 |
"steps_from_proto": {
|
| 288 |
-
"total":
|
| 289 |
-
"count":
|
| 290 |
"is_parallel": true,
|
| 291 |
-
"self":
|
| 292 |
"children": {
|
| 293 |
"_process_rank_one_or_two_observation": {
|
| 294 |
-
"total":
|
| 295 |
-
"count":
|
| 296 |
"is_parallel": true,
|
| 297 |
-
"self":
|
| 298 |
}
|
| 299 |
}
|
| 300 |
}
|
|
@@ -309,9 +309,9 @@
|
|
| 309 |
}
|
| 310 |
},
|
| 311 |
"trainer_threads": {
|
| 312 |
-
"total":
|
| 313 |
"count": 1,
|
| 314 |
-
"self":
|
| 315 |
"children": {
|
| 316 |
"thread_root": {
|
| 317 |
"total": 0.0,
|
|
@@ -320,36 +320,36 @@
|
|
| 320 |
"self": 0.0,
|
| 321 |
"children": {
|
| 322 |
"trainer_advance": {
|
| 323 |
-
"total":
|
| 324 |
-
"count":
|
| 325 |
"is_parallel": true,
|
| 326 |
-
"self":
|
| 327 |
"children": {
|
| 328 |
"process_trajectory": {
|
| 329 |
-
"total":
|
| 330 |
-
"count":
|
| 331 |
"is_parallel": true,
|
| 332 |
-
"self":
|
| 333 |
"children": {
|
| 334 |
"RLTrainer._checkpoint": {
|
| 335 |
-
"total":
|
| 336 |
-
"count":
|
| 337 |
"is_parallel": true,
|
| 338 |
-
"self":
|
| 339 |
}
|
| 340 |
}
|
| 341 |
},
|
| 342 |
"_update_policy": {
|
| 343 |
-
"total":
|
| 344 |
-
"count":
|
| 345 |
"is_parallel": true,
|
| 346 |
-
"self":
|
| 347 |
"children": {
|
| 348 |
"TorchPPOOptimizer.update": {
|
| 349 |
-
"total":
|
| 350 |
-
"count":
|
| 351 |
"is_parallel": true,
|
| 352 |
-
"self":
|
| 353 |
}
|
| 354 |
}
|
| 355 |
}
|
|
@@ -360,14 +360,14 @@
|
|
| 360 |
}
|
| 361 |
},
|
| 362 |
"TrainerController._save_models": {
|
| 363 |
-
"total": 0.
|
| 364 |
"count": 1,
|
| 365 |
-
"self": 0.
|
| 366 |
"children": {
|
| 367 |
"RLTrainer._checkpoint": {
|
| 368 |
-
"total": 0.
|
| 369 |
"count": 1,
|
| 370 |
-
"self": 0.
|
| 371 |
}
|
| 372 |
}
|
| 373 |
}
|
|
|
|
| 2 |
"name": "root",
|
| 3 |
"gauges": {
|
| 4 |
"SnowballTarget.Policy.Entropy.mean": {
|
| 5 |
+
"value": 0.8867369890213013,
|
| 6 |
+
"min": 0.8867369890213013,
|
| 7 |
+
"max": 2.8655693531036377,
|
| 8 |
+
"count": 20
|
| 9 |
},
|
| 10 |
"SnowballTarget.Policy.Entropy.sum": {
|
| 11 |
+
"value": 8505.5810546875,
|
| 12 |
+
"min": 8505.5810546875,
|
| 13 |
+
"max": 29409.337890625,
|
| 14 |
+
"count": 20
|
| 15 |
},
|
| 16 |
"SnowballTarget.Step.mean": {
|
| 17 |
+
"value": 199984.0,
|
| 18 |
+
"min": 9952.0,
|
| 19 |
+
"max": 199984.0,
|
| 20 |
+
"count": 20
|
| 21 |
},
|
| 22 |
"SnowballTarget.Step.sum": {
|
| 23 |
+
"value": 199984.0,
|
| 24 |
+
"min": 9952.0,
|
| 25 |
+
"max": 199984.0,
|
| 26 |
+
"count": 20
|
| 27 |
},
|
| 28 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
|
| 29 |
+
"value": 13.046875,
|
| 30 |
+
"min": 0.492266982793808,
|
| 31 |
+
"max": 13.046875,
|
| 32 |
+
"count": 20
|
| 33 |
},
|
| 34 |
"SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
|
| 35 |
+
"value": 2544.140625,
|
| 36 |
+
"min": 95.49979400634766,
|
| 37 |
+
"max": 2648.087890625,
|
| 38 |
+
"count": 20
|
| 39 |
},
|
| 40 |
"SnowballTarget.Environment.EpisodeLength.mean": {
|
| 41 |
"value": 199.0,
|
| 42 |
"min": 199.0,
|
| 43 |
"max": 199.0,
|
| 44 |
+
"count": 20
|
| 45 |
},
|
| 46 |
"SnowballTarget.Environment.EpisodeLength.sum": {
|
| 47 |
+
"value": 10945.0,
|
| 48 |
+
"min": 8756.0,
|
| 49 |
+
"max": 10945.0,
|
| 50 |
+
"count": 20
|
| 51 |
},
|
| 52 |
"SnowballTarget.Losses.PolicyLoss.mean": {
|
| 53 |
+
"value": 0.05997834385397807,
|
| 54 |
+
"min": 0.05997834385397807,
|
| 55 |
+
"max": 0.07576822043851923,
|
| 56 |
+
"count": 20
|
| 57 |
},
|
| 58 |
"SnowballTarget.Losses.PolicyLoss.sum": {
|
| 59 |
+
"value": 0.2399133754159123,
|
| 60 |
+
"min": 0.2399133754159123,
|
| 61 |
+
"max": 0.37884110219259615,
|
| 62 |
+
"count": 20
|
| 63 |
},
|
| 64 |
"SnowballTarget.Losses.ValueLoss.mean": {
|
| 65 |
+
"value": 0.20376944804892821,
|
| 66 |
+
"min": 0.12399268490901472,
|
| 67 |
+
"max": 0.2859430268114688,
|
| 68 |
+
"count": 20
|
| 69 |
},
|
| 70 |
"SnowballTarget.Losses.ValueLoss.sum": {
|
| 71 |
+
"value": 0.8150777921957129,
|
| 72 |
+
"min": 0.49597073963605887,
|
| 73 |
+
"max": 1.4297151340573442,
|
| 74 |
+
"count": 20
|
| 75 |
},
|
| 76 |
"SnowballTarget.Policy.LearningRate.mean": {
|
| 77 |
+
"value": 8.082097306000005e-06,
|
| 78 |
+
"min": 8.082097306000005e-06,
|
| 79 |
+
"max": 0.000291882002706,
|
| 80 |
+
"count": 20
|
| 81 |
},
|
| 82 |
"SnowballTarget.Policy.LearningRate.sum": {
|
| 83 |
+
"value": 3.232838922400002e-05,
|
| 84 |
+
"min": 3.232838922400002e-05,
|
| 85 |
+
"max": 0.00138516003828,
|
| 86 |
+
"count": 20
|
| 87 |
},
|
| 88 |
"SnowballTarget.Policy.Epsilon.mean": {
|
| 89 |
+
"value": 0.10269400000000001,
|
| 90 |
+
"min": 0.10269400000000001,
|
| 91 |
+
"max": 0.19729400000000002,
|
| 92 |
+
"count": 20
|
| 93 |
},
|
| 94 |
"SnowballTarget.Policy.Epsilon.sum": {
|
| 95 |
+
"value": 0.41077600000000003,
|
| 96 |
+
"min": 0.41077600000000003,
|
| 97 |
+
"max": 0.96172,
|
| 98 |
+
"count": 20
|
| 99 |
},
|
| 100 |
"SnowballTarget.Policy.Beta.mean": {
|
| 101 |
+
"value": 0.0001444306000000001,
|
| 102 |
+
"min": 0.0001444306000000001,
|
| 103 |
+
"max": 0.0048649706,
|
| 104 |
+
"count": 20
|
| 105 |
},
|
| 106 |
"SnowballTarget.Policy.Beta.sum": {
|
| 107 |
+
"value": 0.0005777224000000004,
|
| 108 |
+
"min": 0.0005777224000000004,
|
| 109 |
+
"max": 0.023089828,
|
| 110 |
+
"count": 20
|
| 111 |
},
|
| 112 |
"SnowballTarget.Environment.CumulativeReward.mean": {
|
| 113 |
+
"value": 25.772727272727273,
|
| 114 |
+
"min": 4.204545454545454,
|
| 115 |
+
"max": 25.90909090909091,
|
| 116 |
+
"count": 20
|
| 117 |
},
|
| 118 |
"SnowballTarget.Environment.CumulativeReward.sum": {
|
| 119 |
+
"value": 1134.0,
|
| 120 |
+
"min": 185.0,
|
| 121 |
+
"max": 1411.0,
|
| 122 |
+
"count": 20
|
| 123 |
},
|
| 124 |
"SnowballTarget.Policy.ExtrinsicReward.mean": {
|
| 125 |
+
"value": 25.772727272727273,
|
| 126 |
+
"min": 4.204545454545454,
|
| 127 |
+
"max": 25.90909090909091,
|
| 128 |
+
"count": 20
|
| 129 |
},
|
| 130 |
"SnowballTarget.Policy.ExtrinsicReward.sum": {
|
| 131 |
+
"value": 1134.0,
|
| 132 |
+
"min": 185.0,
|
| 133 |
+
"max": 1411.0,
|
| 134 |
+
"count": 20
|
| 135 |
},
|
| 136 |
"SnowballTarget.IsTraining.mean": {
|
| 137 |
"value": 1.0,
|
| 138 |
"min": 1.0,
|
| 139 |
"max": 1.0,
|
| 140 |
+
"count": 20
|
| 141 |
},
|
| 142 |
"SnowballTarget.IsTraining.sum": {
|
| 143 |
"value": 1.0,
|
| 144 |
"min": 1.0,
|
| 145 |
"max": 1.0,
|
| 146 |
+
"count": 20
|
| 147 |
}
|
| 148 |
},
|
| 149 |
"metadata": {
|
| 150 |
"timer_format_version": "0.1.0",
|
| 151 |
+
"start_time_seconds": "1686583119",
|
| 152 |
+
"python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
|
| 153 |
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
|
| 154 |
"mlagents_version": "0.31.0.dev0",
|
| 155 |
"mlagents_envs_version": "0.31.0.dev0",
|
| 156 |
"communication_protocol_version": "1.5.0",
|
| 157 |
"pytorch_version": "1.11.0+cu102",
|
| 158 |
"numpy_version": "1.21.2",
|
| 159 |
+
"end_time_seconds": "1686583698"
|
| 160 |
},
|
| 161 |
+
"total": 578.381406132,
|
| 162 |
"count": 1,
|
| 163 |
+
"self": 0.9531930520000742,
|
| 164 |
"children": {
|
| 165 |
"run_training.setup": {
|
| 166 |
+
"total": 0.07172308900004509,
|
| 167 |
"count": 1,
|
| 168 |
+
"self": 0.07172308900004509
|
| 169 |
},
|
| 170 |
"TrainerController.start_learning": {
|
| 171 |
+
"total": 577.356489991,
|
| 172 |
"count": 1,
|
| 173 |
+
"self": 0.7381638260003456,
|
| 174 |
"children": {
|
| 175 |
"TrainerController._reset_env": {
|
| 176 |
+
"total": 5.815364906000013,
|
| 177 |
"count": 1,
|
| 178 |
+
"self": 5.815364906000013
|
| 179 |
},
|
| 180 |
"TrainerController.advance": {
|
| 181 |
+
"total": 570.5417217109995,
|
| 182 |
+
"count": 18230,
|
| 183 |
+
"self": 0.38011827400202947,
|
| 184 |
"children": {
|
| 185 |
"env_step": {
|
| 186 |
+
"total": 570.1616034369974,
|
| 187 |
+
"count": 18230,
|
| 188 |
+
"self": 415.2228667380073,
|
| 189 |
"children": {
|
| 190 |
"SubprocessEnvManager._take_step": {
|
| 191 |
+
"total": 154.57025875799974,
|
| 192 |
+
"count": 18230,
|
| 193 |
+
"self": 2.329320297002539,
|
| 194 |
"children": {
|
| 195 |
"TorchPolicy.evaluate": {
|
| 196 |
+
"total": 152.2409384609972,
|
| 197 |
+
"count": 18230,
|
| 198 |
+
"self": 152.2409384609972
|
| 199 |
}
|
| 200 |
}
|
| 201 |
},
|
| 202 |
"workers": {
|
| 203 |
+
"total": 0.368477940990374,
|
| 204 |
+
"count": 18230,
|
| 205 |
"self": 0.0,
|
| 206 |
"children": {
|
| 207 |
"worker_root": {
|
| 208 |
+
"total": 574.9013252950081,
|
| 209 |
+
"count": 18230,
|
| 210 |
"is_parallel": true,
|
| 211 |
+
"self": 264.8934394190214,
|
| 212 |
"children": {
|
| 213 |
"run_training.setup": {
|
| 214 |
"total": 0.0,
|
|
|
|
| 217 |
"self": 0.0,
|
| 218 |
"children": {
|
| 219 |
"steps_from_proto": {
|
| 220 |
+
"total": 0.006853279999972983,
|
| 221 |
"count": 1,
|
| 222 |
"is_parallel": true,
|
| 223 |
+
"self": 0.004323302999978296,
|
| 224 |
"children": {
|
| 225 |
"_process_rank_one_or_two_observation": {
|
| 226 |
+
"total": 0.0025299769999946875,
|
| 227 |
"count": 10,
|
| 228 |
"is_parallel": true,
|
| 229 |
+
"self": 0.0025299769999946875
|
| 230 |
}
|
| 231 |
}
|
| 232 |
},
|
| 233 |
"UnityEnvironment.step": {
|
| 234 |
+
"total": 0.03998278200003824,
|
| 235 |
"count": 1,
|
| 236 |
"is_parallel": true,
|
| 237 |
+
"self": 0.0006589900000335547,
|
| 238 |
"children": {
|
| 239 |
"UnityEnvironment._generate_step_input": {
|
| 240 |
+
"total": 0.00029402000001255146,
|
| 241 |
"count": 1,
|
| 242 |
"is_parallel": true,
|
| 243 |
+
"self": 0.00029402000001255146
|
| 244 |
},
|
| 245 |
"communicator.exchange": {
|
| 246 |
+
"total": 0.0366775109999935,
|
| 247 |
"count": 1,
|
| 248 |
"is_parallel": true,
|
| 249 |
+
"self": 0.0366775109999935
|
| 250 |
},
|
| 251 |
"steps_from_proto": {
|
| 252 |
+
"total": 0.0023522609999986344,
|
| 253 |
"count": 1,
|
| 254 |
"is_parallel": true,
|
| 255 |
+
"self": 0.0004488470000296729,
|
| 256 |
"children": {
|
| 257 |
"_process_rank_one_or_two_observation": {
|
| 258 |
+
"total": 0.0019034139999689614,
|
| 259 |
"count": 10,
|
| 260 |
"is_parallel": true,
|
| 261 |
+
"self": 0.0019034139999689614
|
| 262 |
}
|
| 263 |
}
|
| 264 |
}
|
|
|
|
| 267 |
}
|
| 268 |
},
|
| 269 |
"UnityEnvironment.step": {
|
| 270 |
+
"total": 310.0078858759867,
|
| 271 |
+
"count": 18229,
|
| 272 |
"is_parallel": true,
|
| 273 |
+
"self": 12.6898613589849,
|
| 274 |
"children": {
|
| 275 |
"UnityEnvironment._generate_step_input": {
|
| 276 |
+
"total": 6.402676535993692,
|
| 277 |
+
"count": 18229,
|
| 278 |
"is_parallel": true,
|
| 279 |
+
"self": 6.402676535993692
|
| 280 |
},
|
| 281 |
"communicator.exchange": {
|
| 282 |
+
"total": 248.24702683600935,
|
| 283 |
+
"count": 18229,
|
| 284 |
"is_parallel": true,
|
| 285 |
+
"self": 248.24702683600935
|
| 286 |
},
|
| 287 |
"steps_from_proto": {
|
| 288 |
+
"total": 42.66832114499874,
|
| 289 |
+
"count": 18229,
|
| 290 |
"is_parallel": true,
|
| 291 |
+
"self": 8.275188611003898,
|
| 292 |
"children": {
|
| 293 |
"_process_rank_one_or_two_observation": {
|
| 294 |
+
"total": 34.393132533994844,
|
| 295 |
+
"count": 182290,
|
| 296 |
"is_parallel": true,
|
| 297 |
+
"self": 34.393132533994844
|
| 298 |
}
|
| 299 |
}
|
| 300 |
}
|
|
|
|
| 309 |
}
|
| 310 |
},
|
| 311 |
"trainer_threads": {
|
| 312 |
+
"total": 0.0001654920000646598,
|
| 313 |
"count": 1,
|
| 314 |
+
"self": 0.0001654920000646598,
|
| 315 |
"children": {
|
| 316 |
"thread_root": {
|
| 317 |
"total": 0.0,
|
|
|
|
| 320 |
"self": 0.0,
|
| 321 |
"children": {
|
| 322 |
"trainer_advance": {
|
| 323 |
+
"total": 565.8226938610748,
|
| 324 |
+
"count": 546312,
|
| 325 |
"is_parallel": true,
|
| 326 |
+
"self": 12.58049569512616,
|
| 327 |
"children": {
|
| 328 |
"process_trajectory": {
|
| 329 |
+
"total": 320.60923400694793,
|
| 330 |
+
"count": 546312,
|
| 331 |
"is_parallel": true,
|
| 332 |
+
"self": 305.64768057594847,
|
| 333 |
"children": {
|
| 334 |
"RLTrainer._checkpoint": {
|
| 335 |
+
"total": 14.961553430999459,
|
| 336 |
+
"count": 40,
|
| 337 |
"is_parallel": true,
|
| 338 |
+
"self": 14.961553430999459
|
| 339 |
}
|
| 340 |
}
|
| 341 |
},
|
| 342 |
"_update_policy": {
|
| 343 |
+
"total": 232.63296415900066,
|
| 344 |
+
"count": 90,
|
| 345 |
"is_parallel": true,
|
| 346 |
+
"self": 86.53489770399852,
|
| 347 |
"children": {
|
| 348 |
"TorchPPOOptimizer.update": {
|
| 349 |
+
"total": 146.09806645500214,
|
| 350 |
+
"count": 4587,
|
| 351 |
"is_parallel": true,
|
| 352 |
+
"self": 146.09806645500214
|
| 353 |
}
|
| 354 |
}
|
| 355 |
}
|
|
|
|
| 360 |
}
|
| 361 |
},
|
| 362 |
"TrainerController._save_models": {
|
| 363 |
+
"total": 0.2610740560000977,
|
| 364 |
"count": 1,
|
| 365 |
+
"self": 0.003333339000050728,
|
| 366 |
"children": {
|
| 367 |
"RLTrainer._checkpoint": {
|
| 368 |
+
"total": 0.25774071700004697,
|
| 369 |
"count": 1,
|
| 370 |
+
"self": 0.25774071700004697
|
| 371 |
}
|
| 372 |
}
|
| 373 |
}
|
run_logs/training_status.json
CHANGED
|
@@ -2,31 +2,103 @@
|
|
| 2 |
"SnowballTarget": {
|
| 3 |
"checkpoints": [
|
| 4 |
{
|
| 5 |
-
"steps":
|
| 6 |
-
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-
|
| 7 |
-
"reward":
|
| 8 |
-
"creation_time":
|
| 9 |
"auxillary_file_paths": [
|
| 10 |
-
"results/SnowballTarget1/SnowballTarget/SnowballTarget-
|
| 11 |
]
|
| 12 |
},
|
| 13 |
{
|
| 14 |
-
"steps":
|
| 15 |
-
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-
|
| 16 |
-
"reward":
|
| 17 |
-
"creation_time":
|
| 18 |
"auxillary_file_paths": [
|
| 19 |
-
"results/SnowballTarget1/SnowballTarget/SnowballTarget-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
]
|
| 21 |
}
|
| 22 |
],
|
| 23 |
"final_checkpoint": {
|
| 24 |
-
"steps":
|
| 25 |
"file_path": "results/SnowballTarget1/SnowballTarget.onnx",
|
| 26 |
-
"reward":
|
| 27 |
-
"creation_time":
|
| 28 |
"auxillary_file_paths": [
|
| 29 |
-
"results/SnowballTarget1/SnowballTarget/SnowballTarget-
|
| 30 |
]
|
| 31 |
}
|
| 32 |
},
|
|
|
|
| 2 |
"SnowballTarget": {
|
| 3 |
"checkpoints": [
|
| 4 |
{
|
| 5 |
+
"steps": 159936,
|
| 6 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-159936.onnx",
|
| 7 |
+
"reward": 25.636363636363637,
|
| 8 |
+
"creation_time": 1686583580.4983654,
|
| 9 |
"auxillary_file_paths": [
|
| 10 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-159936.pt"
|
| 11 |
]
|
| 12 |
},
|
| 13 |
{
|
| 14 |
+
"steps": 164992,
|
| 15 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-164992.onnx",
|
| 16 |
+
"reward": 24.4,
|
| 17 |
+
"creation_time": 1686583596.0095642,
|
| 18 |
"auxillary_file_paths": [
|
| 19 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-164992.pt"
|
| 20 |
+
]
|
| 21 |
+
},
|
| 22 |
+
{
|
| 23 |
+
"steps": 169976,
|
| 24 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-169976.onnx",
|
| 25 |
+
"reward": 25.545454545454547,
|
| 26 |
+
"creation_time": 1686583609.0519998,
|
| 27 |
+
"auxillary_file_paths": [
|
| 28 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-169976.pt"
|
| 29 |
+
]
|
| 30 |
+
},
|
| 31 |
+
{
|
| 32 |
+
"steps": 174952,
|
| 33 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-174952.onnx",
|
| 34 |
+
"reward": 25.09090909090909,
|
| 35 |
+
"creation_time": 1686583623.4727666,
|
| 36 |
+
"auxillary_file_paths": [
|
| 37 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-174952.pt"
|
| 38 |
+
]
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"steps": 179992,
|
| 42 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-179992.onnx",
|
| 43 |
+
"reward": 26.272727272727273,
|
| 44 |
+
"creation_time": 1686583638.0229383,
|
| 45 |
+
"auxillary_file_paths": [
|
| 46 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-179992.pt"
|
| 47 |
+
]
|
| 48 |
+
},
|
| 49 |
+
{
|
| 50 |
+
"steps": 184992,
|
| 51 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-184992.onnx",
|
| 52 |
+
"reward": 25.454545454545453,
|
| 53 |
+
"creation_time": 1686583653.674834,
|
| 54 |
+
"auxillary_file_paths": [
|
| 55 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-184992.pt"
|
| 56 |
+
]
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"steps": 189968,
|
| 60 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-189968.onnx",
|
| 61 |
+
"reward": 25.636363636363637,
|
| 62 |
+
"creation_time": 1686583668.41145,
|
| 63 |
+
"auxillary_file_paths": [
|
| 64 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-189968.pt"
|
| 65 |
+
]
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"steps": 194944,
|
| 69 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-194944.onnx",
|
| 70 |
+
"reward": 25.272727272727273,
|
| 71 |
+
"creation_time": 1686583682.1123881,
|
| 72 |
+
"auxillary_file_paths": [
|
| 73 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-194944.pt"
|
| 74 |
+
]
|
| 75 |
+
},
|
| 76 |
+
{
|
| 77 |
+
"steps": 199984,
|
| 78 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
|
| 79 |
+
"reward": 26.363636363636363,
|
| 80 |
+
"creation_time": 1686583696.930381,
|
| 81 |
+
"auxillary_file_paths": [
|
| 82 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
|
| 83 |
+
]
|
| 84 |
+
},
|
| 85 |
+
{
|
| 86 |
+
"steps": 200112,
|
| 87 |
+
"file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
|
| 88 |
+
"reward": 26.363636363636363,
|
| 89 |
+
"creation_time": 1686583697.2291028,
|
| 90 |
+
"auxillary_file_paths": [
|
| 91 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
|
| 92 |
]
|
| 93 |
}
|
| 94 |
],
|
| 95 |
"final_checkpoint": {
|
| 96 |
+
"steps": 200112,
|
| 97 |
"file_path": "results/SnowballTarget1/SnowballTarget.onnx",
|
| 98 |
+
"reward": 26.363636363636363,
|
| 99 |
+
"creation_time": 1686583697.2291028,
|
| 100 |
"auxillary_file_paths": [
|
| 101 |
+
"results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
|
| 102 |
]
|
| 103 |
}
|
| 104 |
},
|