Roberto commited on
Commit
9ce5be9
·
1 Parent(s): b01b5b0
Files changed (31) hide show
  1. SnowballTarget.onnx +1 -1
  2. SnowballTarget/{SnowballTarget-149984.onnx → SnowballTarget-1599976.onnx} +1 -1
  3. SnowballTarget/{SnowballTarget-200112.pt → SnowballTarget-1599976.pt} +1 -1
  4. SnowballTarget/{SnowballTarget-200112.onnx → SnowballTarget-1649992.onnx} +1 -1
  5. SnowballTarget/{SnowballTarget-149984.pt → SnowballTarget-1649992.pt} +1 -1
  6. SnowballTarget/{SnowballTarget-199984.onnx → SnowballTarget-1699936.onnx} +1 -1
  7. SnowballTarget/{SnowballTarget-49936.pt → SnowballTarget-1699936.pt} +1 -1
  8. SnowballTarget/{SnowballTarget-49936.onnx → SnowballTarget-1749960.onnx} +1 -1
  9. SnowballTarget/{SnowballTarget-199984.pt → SnowballTarget-1749960.pt} +1 -1
  10. SnowballTarget/SnowballTarget-1799984.onnx +3 -0
  11. SnowballTarget/SnowballTarget-1799984.pt +3 -0
  12. SnowballTarget/SnowballTarget-1849984.onnx +3 -0
  13. SnowballTarget/SnowballTarget-1849984.pt +3 -0
  14. SnowballTarget/SnowballTarget-1899944.onnx +3 -0
  15. SnowballTarget/SnowballTarget-1899944.pt +3 -0
  16. SnowballTarget/SnowballTarget-1949968.onnx +3 -0
  17. SnowballTarget/SnowballTarget-1949968.pt +3 -0
  18. SnowballTarget/SnowballTarget-1999992.onnx +3 -0
  19. SnowballTarget/SnowballTarget-1999992.pt +3 -0
  20. SnowballTarget/SnowballTarget-2000376.onnx +3 -0
  21. SnowballTarget/SnowballTarget-2000376.pt +3 -0
  22. SnowballTarget/SnowballTarget-99960.onnx +0 -3
  23. SnowballTarget/SnowballTarget-99960.pt +0 -3
  24. SnowballTarget/checkpoint.pt +1 -1
  25. SnowballTarget/events.out.tfevents.1673800027.lamarmite.457457.0 +0 -3
  26. SnowballTarget/events.out.tfevents.1673823722.lamarmite.489332.0 +3 -0
  27. config.json +1 -1
  28. configuration.yaml +3 -3
  29. run_logs/Player-0.log +16 -19
  30. run_logs/timers.json +159 -159
  31. run_logs/training_status.json +75 -30
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e038a9e36e91b807a82bf3158a92856090d77a335b33b945ffec9cc8f64b9ef0
3
  size 645119
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acabf98140a6dc8ab34f9f2c68cdb740bedc5635763d13106971e2699b5c6f21
3
  size 645119
SnowballTarget/{SnowballTarget-149984.onnx → SnowballTarget-1599976.onnx} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02abbe8672d5f82480297b463fee125f687959d5843fe12520614b2a309f6d13
3
  size 645119
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7933853f98178301d95060023013db82b527b77fa06f17738fe2943855c490c1
3
  size 645119
SnowballTarget/{SnowballTarget-200112.pt → SnowballTarget-1599976.pt} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7fbea28e9a3c8b6adef2503f2f41fa7453d1f83c0a61c87f36dd907a7ab20e2d
3
  size 3844788
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10c86fe78bc1226ed4779d8d08689f768ab09d883c46085972e0619e1d0a243a
3
  size 3844788
SnowballTarget/{SnowballTarget-200112.onnx → SnowballTarget-1649992.onnx} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e038a9e36e91b807a82bf3158a92856090d77a335b33b945ffec9cc8f64b9ef0
3
  size 645119
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ab4c6b348a63561d55a1695bb8c1be8c990005436e6af9b125d8a8c74151d5f
3
  size 645119
SnowballTarget/{SnowballTarget-149984.pt → SnowballTarget-1649992.pt} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:daf4da98e80f7ad24eb82cdd048f4696dd42023105ec24d5db5ad52cd1baf244
3
  size 3844788
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08829efcb81b159036fd83ea2035db6738f9fef870f0dc7fd7c3259cab8e241e
3
  size 3844788
SnowballTarget/{SnowballTarget-199984.onnx → SnowballTarget-1699936.onnx} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e038a9e36e91b807a82bf3158a92856090d77a335b33b945ffec9cc8f64b9ef0
3
  size 645119
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3721232ff4cad21c66d35474dcac52002cef30b7a78bf7038b6459d14eedfca7
3
  size 645119
SnowballTarget/{SnowballTarget-49936.pt → SnowballTarget-1699936.pt} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fca67d8cc7fc3d6db4e876ea8dcef5a9300b88e5720d04b8acc2ff0ee9cbf90b
3
  size 3844788
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76d6660f41179480ea3da7519f9c44b388054b7f359c178e6cfb1b54489e7cc7
3
  size 3844788
SnowballTarget/{SnowballTarget-49936.onnx → SnowballTarget-1749960.onnx} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18e1e682f9a68e172a4cf85140596baad59dc6963864de325620a0d199a42100
3
  size 645119
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45f87a792b86aa7fc5625320232453d0d71459353fde8fc47e6d3cc3adbee6b8
3
  size 645119
SnowballTarget/{SnowballTarget-199984.pt → SnowballTarget-1749960.pt} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd9b10096e9c99f6c05d8527f28f128ba45687a13b92e24117c9e48ac3450f6f
3
  size 3844788
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6d3577c9354eedb5be3da4f4770a961432492bd6f51693aa950b9f7b86d8fd4
3
  size 3844788
SnowballTarget/SnowballTarget-1799984.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7580bb27200b65ae3a8c26bab14ed7023d66a6a9a54d39dd837592a6f1fe3ec3
3
+ size 645119
SnowballTarget/SnowballTarget-1799984.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0455f06960f461a7a0de8f59bbf9d7aa8cdbec03f1ec9f5034521cc5eed6183
3
+ size 3844788
SnowballTarget/SnowballTarget-1849984.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1adda266c50c8490c4da7c4489f4cc2097318addbd67cc79f038e56e293b4cb9
3
+ size 645119
SnowballTarget/SnowballTarget-1849984.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3cca88014eaac491218345ba618b91efd116982db883512a435c1b8f2bdcfb8
3
+ size 3844788
SnowballTarget/SnowballTarget-1899944.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34b724efc9c127cf9ad586af51783d46f1535e8f9623a933b621aef8d6094136
3
+ size 645119
SnowballTarget/SnowballTarget-1899944.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4927af5f83c62881111b940b56fe7e94c7db70802cd1416dc1764e0b7f935043
3
+ size 3844788
SnowballTarget/SnowballTarget-1949968.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e716baaef7e7b98c4ed8b0f303a1bdff26fb071a720f1a999f818503272a2e26
3
+ size 645119
SnowballTarget/SnowballTarget-1949968.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ea4e3dab9ced3cce4002f2433e8660ff1158776cf9b3d2ed0de58c37fda5c22
3
+ size 3844788
SnowballTarget/SnowballTarget-1999992.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acabf98140a6dc8ab34f9f2c68cdb740bedc5635763d13106971e2699b5c6f21
3
+ size 645119
SnowballTarget/SnowballTarget-1999992.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:379665e081fde322f42d5c7c5d8fc3d5660500fbb2a7aa213eb917e92772cf43
3
+ size 3844788
SnowballTarget/SnowballTarget-2000376.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acabf98140a6dc8ab34f9f2c68cdb740bedc5635763d13106971e2699b5c6f21
3
+ size 645119
SnowballTarget/SnowballTarget-2000376.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f042d4d81496589845355f39b74993dc72222d318356823bda4427ae06cd6416
3
+ size 3844788
SnowballTarget/SnowballTarget-99960.onnx DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:32b8e7940773ef088459ab5cfa04ed4f71bce716b79437d271cd517b982c4584
3
- size 645119
 
 
 
 
SnowballTarget/SnowballTarget-99960.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:af456fc03156ba029aa9ea30a62d2b3640e36797de4760f2078db0bf8264ddf4
3
- size 3844788
 
 
 
 
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7fbea28e9a3c8b6adef2503f2f41fa7453d1f83c0a61c87f36dd907a7ab20e2d
3
  size 3844788
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f042d4d81496589845355f39b74993dc72222d318356823bda4427ae06cd6416
3
  size 3844788
SnowballTarget/events.out.tfevents.1673800027.lamarmite.457457.0 DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:b7c18cfd2d136275800949cc0d0b7a6c0bf3aeacd483e403701f1741a3739e9c
3
- size 23957
 
 
 
 
SnowballTarget/events.out.tfevents.1673823722.lamarmite.489332.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b846d79ba1385ed590cd2eddd2fa66e17fe3aa100a4ba4ca33632a59370a843
3
+ size 170111
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "checkpoint_interval": 50000, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0002, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "checkpoint_interval": 50000, "max_steps": 2000000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget5", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -5,7 +5,7 @@ behaviors:
5
  hyperparameters:
6
  batch_size: 128
7
  buffer_size: 2048
8
- learning_rate: 0.0003
9
  beta: 0.005
10
  epsilon: 0.2
11
  lambd: 0.95
@@ -36,7 +36,7 @@ behaviors:
36
  init_path: null
37
  keep_checkpoints: 10
38
  checkpoint_interval: 50000
39
- max_steps: 200000
40
  time_horizon: 64
41
  summary_freq: 10000
42
  threaded: true
@@ -62,7 +62,7 @@ engine_settings:
62
  no_graphics: true
63
  environment_parameters: null
64
  checkpoint_settings:
65
- run_id: SnowballTarget1
66
  initialize_from: null
67
  load_model: false
68
  resume: false
 
5
  hyperparameters:
6
  batch_size: 128
7
  buffer_size: 2048
8
+ learning_rate: 0.0002
9
  beta: 0.005
10
  epsilon: 0.2
11
  lambd: 0.95
 
36
  init_path: null
37
  keep_checkpoints: 10
38
  checkpoint_interval: 50000
39
+ max_steps: 2000000
40
  time_horizon: 64
41
  summary_freq: 10000
42
  threaded: true
 
62
  no_graphics: true
63
  environment_parameters: null
64
  checkpoint_settings:
65
+ run_id: SnowballTarget5
66
  initialize_from: null
67
  load_model: false
68
  resume: false
run_logs/Player-0.log CHANGED
@@ -2,9 +2,6 @@ Mono path[0] = '/home/emmanuel/Documents/code/rl/ml-agents/training-envs-executa
2
  Mono config path = '/home/emmanuel/Documents/code/rl/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
5
- PlayerPrefs - Creating folder: /home/emmanuel/.config/unity3d/Hugging Face
6
- PlayerPrefs - Creating folder: /home/emmanuel/.config/unity3d/Hugging Face/SnowballTarget
7
- Unable to load player prefs
8
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
9
  [Subsystems] Discovering subsystems at path /home/emmanuel/Documents/code/rl/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
10
  Forcing GfxDevice: Null
@@ -14,7 +11,7 @@ NullGfxDevice:
14
  Renderer: Null Device
15
  Vendor: Unity Technologies
16
  Begin MonoManager ReloadAssembly
17
- - Completed reload, in 1.742 seconds
18
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
19
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
20
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -25,7 +22,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
25
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
26
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
27
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
28
- UnloadTime: 2.055541 ms
29
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
30
  requesting resize 84 x 84
31
  Setting up 2 worker threads for Enlighten.
@@ -33,7 +30,7 @@ Memory Statistics:
33
  [ALLOC_TEMP_TLS] TLS Allocator
34
  StackAllocators :
35
  [ALLOC_TEMP_MAIN]
36
- Peak usage frame count: [8.0 KB-16.0 KB]: 5370 frames, [16.0 KB-32.0 KB]: 91 frames, [2.0 MB-4.0 MB]: 1 frames
37
  Initial Block Size 4.0 MB
38
  Current Block Size 4.0 MB
39
  Peak Allocated Bytes 2.0 MB
@@ -159,22 +156,22 @@ Memory Statistics:
159
  Peak Allocated Bytes 0 B
160
  Overflow Count 0
161
  [ALLOC_DEFAULT] Dual Thread Allocator
162
- Peak main deferred allocation count 45
163
  [ALLOC_BUCKET]
164
  Large Block size 4.0 MB
165
  Used Block count 1
166
  Peak Allocated bytes 1.0 MB
167
  [ALLOC_DEFAULT_MAIN]
168
- Peak usage frame count: [4.0 MB-8.0 MB]: 5462 frames
169
  Requested Block Size 16.0 MB
170
- Peak Block count 1
171
- Peak Allocated memory 7.0 MB
172
  Peak Large allocation bytes 0 B
173
  [ALLOC_DEFAULT_THREAD]
174
- Peak usage frame count: [16.0 MB-32.0 MB]: 5462 frames
175
  Requested Block Size 16.0 MB
176
  Peak Block count 1
177
- Peak Allocated memory 17.4 MB
178
  Peak Large allocation bytes 16.0 MB
179
  [ALLOC_TEMP_JOB_1_FRAME]
180
  Initial Block Size 2.0 MB
@@ -203,13 +200,13 @@ Memory Statistics:
203
  Used Block count 1
204
  Peak Allocated bytes 1.0 MB
205
  [ALLOC_GFX_MAIN]
206
- Peak usage frame count: [32.0 KB-64.0 KB]: 4387 frames, [64.0 KB-128.0 KB]: 1075 frames
207
  Requested Block Size 16.0 MB
208
  Peak Block count 1
209
- Peak Allocated memory 66.8 KB
210
  Peak Large allocation bytes 0 B
211
  [ALLOC_GFX_THREAD]
212
- Peak usage frame count: [32.0 KB-64.0 KB]: 5462 frames
213
  Requested Block Size 16.0 MB
214
  Peak Block count 1
215
  Peak Allocated memory 39.6 KB
@@ -221,13 +218,13 @@ Memory Statistics:
221
  Used Block count 1
222
  Peak Allocated bytes 1.0 MB
223
  [ALLOC_CACHEOBJECTS_MAIN]
224
- Peak usage frame count: [0.5 MB-1.0 MB]: 5462 frames
225
  Requested Block Size 4.0 MB
226
  Peak Block count 1
227
  Peak Allocated memory 0.6 MB
228
  Peak Large allocation bytes 0 B
229
  [ALLOC_CACHEOBJECTS_THREAD]
230
- Peak usage frame count: [0.5 MB-1.0 MB]: 5461 frames, [2.0 MB-4.0 MB]: 1 frames
231
  Requested Block Size 4.0 MB
232
  Peak Block count 1
233
  Peak Allocated memory 2.2 MB
@@ -239,13 +236,13 @@ Memory Statistics:
239
  Used Block count 1
240
  Peak Allocated bytes 1.0 MB
241
  [ALLOC_TYPETREE_MAIN]
242
- Peak usage frame count: [0-1.0 KB]: 5462 frames
243
  Requested Block Size 2.0 MB
244
  Peak Block count 1
245
  Peak Allocated memory 1.0 KB
246
  Peak Large allocation bytes 0 B
247
  [ALLOC_TYPETREE_THREAD]
248
- Peak usage frame count: [1.0 KB-2.0 KB]: 5462 frames
249
  Requested Block Size 2.0 MB
250
  Peak Block count 1
251
  Peak Allocated memory 1.7 KB
 
2
  Mono config path = '/home/emmanuel/Documents/code/rl/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
 
 
 
5
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
6
  [Subsystems] Discovering subsystems at path /home/emmanuel/Documents/code/rl/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
7
  Forcing GfxDevice: Null
 
11
  Renderer: Null Device
12
  Vendor: Unity Technologies
13
  Begin MonoManager ReloadAssembly
14
+ - Completed reload, in 0.870 seconds
15
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
16
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
17
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
22
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
23
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
24
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
25
+ UnloadTime: 0.896450 ms
26
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
27
  requesting resize 84 x 84
28
  Setting up 2 worker threads for Enlighten.
 
30
  [ALLOC_TEMP_TLS] TLS Allocator
31
  StackAllocators :
32
  [ALLOC_TEMP_MAIN]
33
+ Peak usage frame count: [8.0 KB-16.0 KB]: 53653 frames, [16.0 KB-32.0 KB]: 909 frames, [2.0 MB-4.0 MB]: 1 frames
34
  Initial Block Size 4.0 MB
35
  Current Block Size 4.0 MB
36
  Peak Allocated Bytes 2.0 MB
 
156
  Peak Allocated Bytes 0 B
157
  Overflow Count 0
158
  [ALLOC_DEFAULT] Dual Thread Allocator
159
+ Peak main deferred allocation count 41
160
  [ALLOC_BUCKET]
161
  Large Block size 4.0 MB
162
  Used Block count 1
163
  Peak Allocated bytes 1.0 MB
164
  [ALLOC_DEFAULT_MAIN]
165
+ Peak usage frame count: [4.0 MB-8.0 MB]: 11046 frames, [8.0 MB-16.0 MB]: 27196 frames, [16.0 MB-32.0 MB]: 16321 frames
166
  Requested Block Size 16.0 MB
167
+ Peak Block count 2
168
+ Peak Allocated memory 24.4 MB
169
  Peak Large allocation bytes 0 B
170
  [ALLOC_DEFAULT_THREAD]
171
+ Peak usage frame count: [16.0 MB-32.0 MB]: 54563 frames
172
  Requested Block Size 16.0 MB
173
  Peak Block count 1
174
+ Peak Allocated memory 17.6 MB
175
  Peak Large allocation bytes 16.0 MB
176
  [ALLOC_TEMP_JOB_1_FRAME]
177
  Initial Block Size 2.0 MB
 
200
  Used Block count 1
201
  Peak Allocated bytes 1.0 MB
202
  [ALLOC_GFX_MAIN]
203
+ Peak usage frame count: [32.0 KB-64.0 KB]: 52645 frames, [64.0 KB-128.0 KB]: 1918 frames
204
  Requested Block Size 16.0 MB
205
  Peak Block count 1
206
+ Peak Allocated memory 67.7 KB
207
  Peak Large allocation bytes 0 B
208
  [ALLOC_GFX_THREAD]
209
+ Peak usage frame count: [32.0 KB-64.0 KB]: 54563 frames
210
  Requested Block Size 16.0 MB
211
  Peak Block count 1
212
  Peak Allocated memory 39.6 KB
 
218
  Used Block count 1
219
  Peak Allocated bytes 1.0 MB
220
  [ALLOC_CACHEOBJECTS_MAIN]
221
+ Peak usage frame count: [0.5 MB-1.0 MB]: 54563 frames
222
  Requested Block Size 4.0 MB
223
  Peak Block count 1
224
  Peak Allocated memory 0.6 MB
225
  Peak Large allocation bytes 0 B
226
  [ALLOC_CACHEOBJECTS_THREAD]
227
+ Peak usage frame count: [0.5 MB-1.0 MB]: 54562 frames, [2.0 MB-4.0 MB]: 1 frames
228
  Requested Block Size 4.0 MB
229
  Peak Block count 1
230
  Peak Allocated memory 2.2 MB
 
236
  Used Block count 1
237
  Peak Allocated bytes 1.0 MB
238
  [ALLOC_TYPETREE_MAIN]
239
+ Peak usage frame count: [0-1.0 KB]: 54563 frames
240
  Requested Block Size 2.0 MB
241
  Peak Block count 1
242
  Peak Allocated memory 1.0 KB
243
  Peak Large allocation bytes 0 B
244
  [ALLOC_TYPETREE_THREAD]
245
+ Peak usage frame count: [1.0 KB-2.0 KB]: 54563 frames
246
  Requested Block Size 2.0 MB
247
  Peak Block count 1
248
  Peak Allocated memory 1.7 KB
run_logs/timers.json CHANGED
@@ -2,220 +2,220 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 1.0211290121078491,
6
- "min": 1.0211290121078491,
7
- "max": 2.854013204574585,
8
- "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 9783.4375,
12
- "min": 9783.4375,
13
- "max": 29227.94921875,
14
- "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
17
- "value": 199984.0,
18
  "min": 9952.0,
19
- "max": 199984.0,
20
- "count": 20
21
  },
22
  "SnowballTarget.Step.sum": {
23
- "value": 199984.0,
24
  "min": 9952.0,
25
- "max": 199984.0,
26
- "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 12.06716251373291,
30
- "min": 0.5820059776306152,
31
- "max": 12.06716251373291,
32
- "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2353.0966796875,
36
- "min": 112.9091567993164,
37
- "max": 2436.14111328125,
38
- "count": 20
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
41
  "value": 199.0,
42
  "min": 199.0,
43
  "max": 199.0,
44
- "count": 20
45
  },
46
  "SnowballTarget.Environment.EpisodeLength.sum": {
47
  "value": 10945.0,
48
  "min": 8756.0,
49
  "max": 10945.0,
50
- "count": 20
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.06793563267887757,
54
- "min": 0.06066728261865529,
55
- "max": 0.07288110168227314,
56
- "count": 20
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.2717425307155103,
60
- "min": 0.24266913047462116,
61
- "max": 0.35947661169955286,
62
- "count": 20
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.2192965249950979,
66
- "min": 0.12452422489649526,
67
- "max": 0.3020132852389532,
68
- "count": 20
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.8771860999803917,
72
- "min": 0.498096899585981,
73
- "max": 1.3851492670455983,
74
- "count": 20
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
- "value": 8.082097306000005e-06,
78
- "min": 8.082097306000005e-06,
79
- "max": 0.000291882002706,
80
- "count": 20
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
83
- "value": 3.232838922400002e-05,
84
- "min": 3.232838922400002e-05,
85
- "max": 0.001385028038324,
86
- "count": 20
87
  },
88
  "SnowballTarget.Policy.Epsilon.mean": {
89
- "value": 0.10269400000000001,
90
- "min": 0.10269400000000001,
91
- "max": 0.19729400000000002,
92
- "count": 20
93
  },
94
  "SnowballTarget.Policy.Epsilon.sum": {
95
- "value": 0.41077600000000003,
96
- "min": 0.41077600000000003,
97
- "max": 0.9616760000000001,
98
- "count": 20
99
  },
100
  "SnowballTarget.Policy.Beta.mean": {
101
- "value": 0.0001444306000000001,
102
- "min": 0.0001444306000000001,
103
- "max": 0.0048649706,
104
- "count": 20
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
107
- "value": 0.0005777224000000004,
108
- "min": 0.0005777224000000004,
109
- "max": 0.0230876324,
110
- "count": 20
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 24.068181818181817,
114
  "min": 3.4318181818181817,
115
- "max": 24.068181818181817,
116
- "count": 20
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1059.0,
120
  "min": 151.0,
121
- "max": 1317.0,
122
- "count": 20
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 24.068181818181817,
126
  "min": 3.4318181818181817,
127
- "max": 24.068181818181817,
128
- "count": 20
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1059.0,
132
  "min": 151.0,
133
- "max": 1317.0,
134
- "count": 20
135
  },
136
  "SnowballTarget.IsTraining.mean": {
137
  "value": 1.0,
138
  "min": 1.0,
139
  "max": 1.0,
140
- "count": 20
141
  },
142
  "SnowballTarget.IsTraining.sum": {
143
  "value": 1.0,
144
  "min": 1.0,
145
  "max": 1.0,
146
- "count": 20
147
  }
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1673800023",
152
  "python_version": "3.8.16 (default, Dec 7 2022, 01:12:06) \n[GCC 11.3.0]",
153
- "command_line_arguments": "/home/emmanuel/Documents/code/rl/ml-agents/venv/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "0.29.0.dev0",
155
  "mlagents_envs_version": "0.29.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.8.1+cu102",
158
  "numpy_version": "1.20.0",
159
- "end_time_seconds": "1673800417"
160
  },
161
- "total": 394.48205814900575,
162
  "count": 1,
163
- "self": 0.3803504949901253,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.01371243700850755,
167
  "count": 1,
168
- "self": 0.01371243700850755
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 394.0879952170071,
172
  "count": 1,
173
- "self": 0.6638770639838185,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 4.369106300000567,
177
  "count": 1,
178
- "self": 4.369106300000567
179
  },
180
  "TrainerController.advance": {
181
- "total": 388.9300568620383,
182
- "count": 18206,
183
- "self": 0.33170540115679614,
184
  "children": {
185
  "env_step": {
186
- "total": 388.59835146088153,
187
- "count": 18206,
188
- "self": 279.28730860404903,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 109.011324763007,
192
- "count": 18206,
193
- "self": 1.6027072951255832,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 107.40861746788141,
197
- "count": 18206,
198
- "self": 14.976935813057935,
199
  "children": {
200
  "TorchPolicy.sample_actions": {
201
- "total": 92.43168165482348,
202
- "count": 18206,
203
- "self": 92.43168165482348
204
  }
205
  }
206
  }
207
  }
208
  },
209
  "workers": {
210
- "total": 0.2997180938255042,
211
- "count": 18206,
212
  "self": 0.0,
213
  "children": {
214
  "worker_root": {
215
- "total": 393.185588507622,
216
- "count": 18206,
217
  "is_parallel": true,
218
- "self": 192.11333968420513,
219
  "children": {
220
  "run_training.setup": {
221
  "total": 0.0,
@@ -224,48 +224,48 @@
224
  "self": 0.0,
225
  "children": {
226
  "steps_from_proto": {
227
- "total": 0.00216466499841772,
228
  "count": 1,
229
  "is_parallel": true,
230
- "self": 0.0006561369227711111,
231
  "children": {
232
  "_process_rank_one_or_two_observation": {
233
- "total": 0.001508528075646609,
234
  "count": 10,
235
  "is_parallel": true,
236
- "self": 0.001508528075646609
237
  }
238
  }
239
  },
240
  "UnityEnvironment.step": {
241
- "total": 0.04352818301413208,
242
  "count": 1,
243
  "is_parallel": true,
244
- "self": 0.00041333597619086504,
245
  "children": {
246
  "UnityEnvironment._generate_step_input": {
247
- "total": 0.0003844230086542666,
248
  "count": 1,
249
  "is_parallel": true,
250
- "self": 0.0003844230086542666
251
  },
252
  "communicator.exchange": {
253
- "total": 0.04146291600773111,
254
  "count": 1,
255
  "is_parallel": true,
256
- "self": 0.04146291600773111
257
  },
258
  "steps_from_proto": {
259
- "total": 0.001267508021555841,
260
  "count": 1,
261
  "is_parallel": true,
262
- "self": 0.0003045150078833103,
263
  "children": {
264
  "_process_rank_one_or_two_observation": {
265
- "total": 0.0009629930136725307,
266
  "count": 10,
267
  "is_parallel": true,
268
- "self": 0.0009629930136725307
269
  }
270
  }
271
  }
@@ -274,34 +274,34 @@
274
  }
275
  },
276
  "UnityEnvironment.step": {
277
- "total": 201.07224882341688,
278
- "count": 18205,
279
  "is_parallel": true,
280
- "self": 7.351680302876048,
281
  "children": {
282
  "UnityEnvironment._generate_step_input": {
283
- "total": 4.4477454441657756,
284
- "count": 18205,
285
  "is_parallel": true,
286
- "self": 4.4477454441657756
287
  },
288
  "communicator.exchange": {
289
- "total": 165.33834766363725,
290
- "count": 18205,
291
  "is_parallel": true,
292
- "self": 165.33834766363725
293
  },
294
  "steps_from_proto": {
295
- "total": 23.93447541273781,
296
- "count": 18205,
297
  "is_parallel": true,
298
- "self": 5.489679446560331,
299
  "children": {
300
  "_process_rank_one_or_two_observation": {
301
- "total": 18.44479596617748,
302
- "count": 182050,
303
  "is_parallel": true,
304
- "self": 18.44479596617748
305
  }
306
  }
307
  }
@@ -316,9 +316,9 @@
316
  }
317
  },
318
  "trainer_threads": {
319
- "total": 0.00014378997730091214,
320
  "count": 1,
321
- "self": 0.00014378997730091214,
322
  "children": {
323
  "thread_root": {
324
  "total": 0.0,
@@ -327,36 +327,36 @@
327
  "self": 0.0,
328
  "children": {
329
  "trainer_advance": {
330
- "total": 386.4299365392362,
331
- "count": 513537,
332
  "is_parallel": true,
333
- "self": 7.588924139650771,
334
  "children": {
335
  "process_trajectory": {
336
- "total": 206.30712276586564,
337
- "count": 513537,
338
  "is_parallel": true,
339
- "self": 205.5930035088968,
340
  "children": {
341
  "RLTrainer._checkpoint": {
342
- "total": 0.7141192569688428,
343
- "count": 4,
344
  "is_parallel": true,
345
- "self": 0.7141192569688428
346
  }
347
  }
348
  },
349
  "_update_policy": {
350
- "total": 172.5338896337198,
351
- "count": 90,
352
  "is_parallel": true,
353
- "self": 50.95602535252692,
354
  "children": {
355
  "TorchPPOOptimizer.update": {
356
- "total": 121.57786428119289,
357
- "count": 4578,
358
  "is_parallel": true,
359
- "self": 121.57786428119289
360
  }
361
  }
362
  }
@@ -367,14 +367,14 @@
367
  }
368
  },
369
  "TrainerController._save_models": {
370
- "total": 0.12481120100710541,
371
  "count": 1,
372
- "self": 0.001185044995509088,
373
  "children": {
374
  "RLTrainer._checkpoint": {
375
- "total": 0.12362615601159632,
376
  "count": 1,
377
- "self": 0.12362615601159632
378
  }
379
  }
380
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.5959854125976562,
6
+ "min": 0.5709379315376282,
7
+ "max": 2.8557894229888916,
8
+ "count": 200
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 6129.7099609375,
12
+ "min": 5529.92138671875,
13
+ "max": 29246.140625,
14
+ "count": 200
15
  },
16
  "SnowballTarget.Step.mean": {
17
+ "value": 1999992.0,
18
  "min": 9952.0,
19
+ "max": 1999992.0,
20
+ "count": 200
21
  },
22
  "SnowballTarget.Step.sum": {
23
+ "value": 1999992.0,
24
  "min": 9952.0,
25
+ "max": 1999992.0,
26
+ "count": 200
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 14.251872062683105,
30
+ "min": 0.47750917077064514,
31
+ "max": 14.377264022827148,
32
+ "count": 200
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 2921.6337890625,
36
+ "min": 92.63677978515625,
37
+ "max": 2939.79443359375,
38
+ "count": 200
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
41
  "value": 199.0,
42
  "min": 199.0,
43
  "max": 199.0,
44
+ "count": 200
45
  },
46
  "SnowballTarget.Environment.EpisodeLength.sum": {
47
  "value": 10945.0,
48
  "min": 8756.0,
49
  "max": 10945.0,
50
+ "count": 200
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
+ "value": 0.07214724341619706,
54
+ "min": 0.06031834367476062,
55
+ "max": 0.07803036969657043,
56
+ "count": 200
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
+ "value": 0.3607362170809853,
60
+ "min": 0.24127337469904248,
61
+ "max": 0.37994980020623326,
62
+ "count": 200
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
+ "value": 0.14957752292062723,
66
+ "min": 0.11958930644608451,
67
+ "max": 0.272374791023778,
68
+ "count": 200
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
+ "value": 0.7478876146031361,
72
+ "min": 0.47835722578433804,
73
+ "max": 1.3618739551188899,
74
+ "count": 200
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
+ "value": 4.688997656000026e-07,
78
+ "min": 4.688997656000026e-07,
79
+ "max": 0.0001994588002706,
80
+ "count": 200
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
83
+ "value": 2.344498828000013e-06,
84
+ "min": 2.344498828000013e-06,
85
+ "max": 0.000992344003828,
86
+ "count": 200
87
  },
88
  "SnowballTarget.Policy.Epsilon.mean": {
89
+ "value": 0.10023440000000002,
90
+ "min": 0.10023440000000002,
91
+ "max": 0.1997294,
92
+ "count": 200
93
  },
94
  "SnowballTarget.Policy.Epsilon.sum": {
95
+ "value": 0.5011720000000001,
96
+ "min": 0.4029176,
97
+ "max": 0.996172,
98
+ "count": 200
99
  },
100
  "SnowballTarget.Policy.Beta.mean": {
101
+ "value": 2.1696560000000067e-05,
102
+ "min": 2.1696560000000067e-05,
103
+ "max": 0.004986497059999999,
104
+ "count": 200
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
107
+ "value": 0.00010848280000000034,
108
+ "min": 0.00010848280000000034,
109
+ "max": 0.024808982800000004,
110
+ "count": 200
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
+ "value": 27.98181818181818,
114
  "min": 3.4318181818181817,
115
+ "max": 28.509090909090908,
116
+ "count": 200
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
+ "value": 1539.0,
120
  "min": 151.0,
121
+ "max": 1568.0,
122
+ "count": 200
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
+ "value": 27.98181818181818,
126
  "min": 3.4318181818181817,
127
+ "max": 28.509090909090908,
128
+ "count": 200
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
+ "value": 1539.0,
132
  "min": 151.0,
133
+ "max": 1568.0,
134
+ "count": 200
135
  },
136
  "SnowballTarget.IsTraining.mean": {
137
  "value": 1.0,
138
  "min": 1.0,
139
  "max": 1.0,
140
+ "count": 200
141
  },
142
  "SnowballTarget.IsTraining.sum": {
143
  "value": 1.0,
144
  "min": 1.0,
145
  "max": 1.0,
146
+ "count": 200
147
  }
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1673823719",
152
  "python_version": "3.8.16 (default, Dec 7 2022, 01:12:06) \n[GCC 11.3.0]",
153
+ "command_line_arguments": "/home/emmanuel/Documents/code/rl/ml-agents/venv/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget5 --no-graphics",
154
  "mlagents_version": "0.29.0.dev0",
155
  "mlagents_envs_version": "0.29.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.8.1+cu102",
158
  "numpy_version": "1.20.0",
159
+ "end_time_seconds": "1673827248"
160
  },
161
+ "total": 3528.254173344001,
162
  "count": 1,
163
+ "self": 0.32245695596793666,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.12691729102516547,
167
  "count": 1,
168
+ "self": 0.12691729102516547
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 3527.804799097008,
172
  "count": 1,
173
+ "self": 5.41858486039564,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 2.0124125689908396,
177
  "count": 1,
178
+ "self": 2.0124125689908396
179
  },
180
  "TrainerController.advance": {
181
+ "total": 3520.259165747644,
182
+ "count": 181874,
183
+ "self": 2.6112302511755843,
184
  "children": {
185
  "env_step": {
186
+ "total": 3517.6479354964686,
187
+ "count": 181874,
188
+ "self": 2569.623420906806,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 945.3249118052481,
192
+ "count": 181874,
193
+ "self": 13.851257300615543,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 931.4736545046326,
197
+ "count": 181874,
198
+ "self": 127.77290135616204,
199
  "children": {
200
  "TorchPolicy.sample_actions": {
201
+ "total": 803.7007531484705,
202
+ "count": 181874,
203
+ "self": 803.7007531484705
204
  }
205
  }
206
  }
207
  }
208
  },
209
  "workers": {
210
+ "total": 2.69960278441431,
211
+ "count": 181874,
212
  "self": 0.0,
213
  "children": {
214
  "worker_root": {
215
+ "total": 3521.14344334777,
216
+ "count": 181874,
217
  "is_parallel": true,
218
+ "self": 1719.3351771153102,
219
  "children": {
220
  "run_training.setup": {
221
  "total": 0.0,
 
224
  "self": 0.0,
225
  "children": {
226
  "steps_from_proto": {
227
+ "total": 0.0019330479844938964,
228
  "count": 1,
229
  "is_parallel": true,
230
+ "self": 0.0005646030185744166,
231
  "children": {
232
  "_process_rank_one_or_two_observation": {
233
+ "total": 0.0013684449659194797,
234
  "count": 10,
235
  "is_parallel": true,
236
+ "self": 0.0013684449659194797
237
  }
238
  }
239
  },
240
  "UnityEnvironment.step": {
241
+ "total": 0.048041015019407496,
242
  "count": 1,
243
  "is_parallel": true,
244
+ "self": 0.00035946001298725605,
245
  "children": {
246
  "UnityEnvironment._generate_step_input": {
247
+ "total": 0.0002662810147739947,
248
  "count": 1,
249
  "is_parallel": true,
250
+ "self": 0.0002662810147739947
251
  },
252
  "communicator.exchange": {
253
+ "total": 0.046154933981597424,
254
  "count": 1,
255
  "is_parallel": true,
256
+ "self": 0.046154933981597424
257
  },
258
  "steps_from_proto": {
259
+ "total": 0.0012603400100488216,
260
  "count": 1,
261
  "is_parallel": true,
262
+ "self": 0.00030130898812785745,
263
  "children": {
264
  "_process_rank_one_or_two_observation": {
265
+ "total": 0.0009590310219209641,
266
  "count": 10,
267
  "is_parallel": true,
268
+ "self": 0.0009590310219209641
269
  }
270
  }
271
  }
 
274
  }
275
  },
276
  "UnityEnvironment.step": {
277
+ "total": 1801.8082662324596,
278
+ "count": 181873,
279
  "is_parallel": true,
280
+ "self": 66.5949679738842,
281
  "children": {
282
  "UnityEnvironment._generate_step_input": {
283
+ "total": 41.4052872666507,
284
+ "count": 181873,
285
  "is_parallel": true,
286
+ "self": 41.4052872666507
287
  },
288
  "communicator.exchange": {
289
+ "total": 1474.7900642491877,
290
+ "count": 181873,
291
  "is_parallel": true,
292
+ "self": 1474.7900642491877
293
  },
294
  "steps_from_proto": {
295
+ "total": 219.01794674273697,
296
+ "count": 181873,
297
  "is_parallel": true,
298
+ "self": 50.45674358413089,
299
  "children": {
300
  "_process_rank_one_or_two_observation": {
301
+ "total": 168.56120315860608,
302
+ "count": 1818730,
303
  "is_parallel": true,
304
+ "self": 168.56120315860608
305
  }
306
  }
307
  }
 
316
  }
317
  },
318
  "trainer_threads": {
319
+ "total": 0.0003306930011603981,
320
  "count": 1,
321
+ "self": 0.0003306930011603981,
322
  "children": {
323
  "thread_root": {
324
  "total": 0.0,
 
327
  "self": 0.0,
328
  "children": {
329
  "trainer_advance": {
330
+ "total": 3502.2347593795857,
331
+ "count": 3965452,
332
  "is_parallel": true,
333
+ "self": 56.153643384197494,
334
  "children": {
335
  "process_trajectory": {
336
+ "total": 1823.7294392998156,
337
+ "count": 3965452,
338
  "is_parallel": true,
339
+ "self": 1812.3949490989326,
340
  "children": {
341
  "RLTrainer._checkpoint": {
342
+ "total": 11.33449020088301,
343
+ "count": 40,
344
  "is_parallel": true,
345
+ "self": 11.33449020088301
346
  }
347
  }
348
  },
349
  "_update_policy": {
350
+ "total": 1622.3516766955727,
351
+ "count": 909,
352
  "is_parallel": true,
353
+ "self": 472.6568366049323,
354
  "children": {
355
  "TorchPPOOptimizer.update": {
356
+ "total": 1149.6948400906404,
357
+ "count": 46257,
358
  "is_parallel": true,
359
+ "self": 1149.6948400906404
360
  }
361
  }
362
  }
 
367
  }
368
  },
369
  "TrainerController._save_models": {
370
+ "total": 0.11430522697628476,
371
  "count": 1,
372
+ "self": 0.000969816988799721,
373
  "children": {
374
  "RLTrainer._checkpoint": {
375
+ "total": 0.11333540998748504,
376
  "count": 1,
377
+ "self": 0.11333540998748504
378
  }
379
  }
380
  }
run_logs/training_status.json CHANGED
@@ -2,58 +2,103 @@
2
  "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
- "steps": 49936,
6
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 13.818181818181818,
8
- "creation_time": 1673800127.8264518,
9
  "auxillary_file_paths": [
10
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
12
  },
13
  {
14
- "steps": 99960,
15
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 20.181818181818183,
17
- "creation_time": 1673800227.5947294,
18
  "auxillary_file_paths": [
19
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
21
  },
22
  {
23
- "steps": 149984,
24
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 22.818181818181817,
26
- "creation_time": 1673800327.2283962,
27
  "auxillary_file_paths": [
28
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
30
  },
31
  {
32
- "steps": 199984,
33
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
- "reward": 25.181818181818183,
35
- "creation_time": 1673800417.1439798,
36
  "auxillary_file_paths": [
37
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
39
  },
40
  {
41
- "steps": 200112,
42
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
- "reward": 25.181818181818183,
44
- "creation_time": 1673800417.297236,
45
  "auxillary_file_paths": [
46
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  ]
48
  }
49
  ],
50
  "final_checkpoint": {
51
- "steps": 200112,
52
- "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
- "reward": 25.181818181818183,
54
- "creation_time": 1673800417.297236,
55
  "auxillary_file_paths": [
56
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]
58
  }
59
  },
 
2
  "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
+ "steps": 1599976,
6
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1599976.onnx",
7
+ "reward": 27.727272727272727,
8
+ "creation_time": 1673826592.562898,
9
  "auxillary_file_paths": [
10
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1599976.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 1649992,
15
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1649992.onnx",
16
+ "reward": 27.6,
17
+ "creation_time": 1673826675.4486237,
18
  "auxillary_file_paths": [
19
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1649992.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 1699936,
24
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1699936.onnx",
25
+ "reward": 27.545454545454547,
26
+ "creation_time": 1673826756.14448,
27
  "auxillary_file_paths": [
28
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1699936.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 1749960,
33
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1749960.onnx",
34
+ "reward": 28.181818181818183,
35
+ "creation_time": 1673826837.3941545,
36
  "auxillary_file_paths": [
37
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1749960.pt"
38
  ]
39
  },
40
  {
41
+ "steps": 1799984,
42
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1799984.onnx",
43
+ "reward": 28.0,
44
+ "creation_time": 1673826920.0776684,
45
  "auxillary_file_paths": [
46
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1799984.pt"
47
+ ]
48
+ },
49
+ {
50
+ "steps": 1849984,
51
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1849984.onnx",
52
+ "reward": 28.181818181818183,
53
+ "creation_time": 1673827000.915594,
54
+ "auxillary_file_paths": [
55
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1849984.pt"
56
+ ]
57
+ },
58
+ {
59
+ "steps": 1899944,
60
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1899944.onnx",
61
+ "reward": 28.272727272727273,
62
+ "creation_time": 1673827082.3755538,
63
+ "auxillary_file_paths": [
64
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1899944.pt"
65
+ ]
66
+ },
67
+ {
68
+ "steps": 1949968,
69
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1949968.onnx",
70
+ "reward": 27.454545454545453,
71
+ "creation_time": 1673827165.5384333,
72
+ "auxillary_file_paths": [
73
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1949968.pt"
74
+ ]
75
+ },
76
+ {
77
+ "steps": 1999992,
78
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-1999992.onnx",
79
+ "reward": 28.454545454545453,
80
+ "creation_time": 1673827247.6539536,
81
+ "auxillary_file_paths": [
82
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-1999992.pt"
83
+ ]
84
+ },
85
+ {
86
+ "steps": 2000376,
87
+ "file_path": "results/SnowballTarget5/SnowballTarget/SnowballTarget-2000376.onnx",
88
+ "reward": 28.454545454545453,
89
+ "creation_time": 1673827247.8208814,
90
+ "auxillary_file_paths": [
91
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-2000376.pt"
92
  ]
93
  }
94
  ],
95
  "final_checkpoint": {
96
+ "steps": 2000376,
97
+ "file_path": "results/SnowballTarget5/SnowballTarget.onnx",
98
+ "reward": 28.454545454545453,
99
+ "creation_time": 1673827247.8208814,
100
  "auxillary_file_paths": [
101
+ "results/SnowballTarget5/SnowballTarget/SnowballTarget-2000376.pt"
102
  ]
103
  }
104
  },