PhysHunter commited on
Commit
1583dc0
·
1 Parent(s): 00f9407

First Push

Browse files
Pyramids.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:06e2f6e7f57e8e333e5d0d2b18a2d62370957d48c80788979296aa975475236b
3
  size 1418184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6486d037f2a295d3ae456529ffddaeb17bc770b886f4814dbac8d784bcc874f9
3
  size 1418184
Pyramids/Pyramids-1499996.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edfe19b971ad80ad72d884bd5b1cdc560103ff85dcc589c02fa43a38de98b2ec
3
+ size 1418184
Pyramids/Pyramids-1499996.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c633d0a5244686462a7d64a9d837a810a98abe927b8c613ba1c50e5240f4e043
3
+ size 8650990
Pyramids/Pyramids-1999977.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c33e866a2d772adb4717757b6d82b6290f87cc298f450b63e8f0b313d447f5b
3
+ size 1418184
Pyramids/Pyramids-1999977.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee7ad1719eae92193be30ec06675e35029d3ae4aadab292a2dde4508a29196e9
3
+ size 8650990
Pyramids/Pyramids-2499973.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aad3626400c03ea37b053aac9a9f696adb3a5e3cbcd13a6dba4d1a755b16a822
3
+ size 1418184
Pyramids/Pyramids-2499973.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fb71f1860a597f76efb15f22a133b47867338214cc4b8ddc1df0bb285d3b838
3
+ size 8650990
Pyramids/Pyramids-2999969.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6486d037f2a295d3ae456529ffddaeb17bc770b886f4814dbac8d784bcc874f9
3
+ size 1418184
Pyramids/Pyramids-2999969.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df99146f7da5a233980babd54c9e15f0fdf334411353d879f3b0be5833a6e30a
3
+ size 8650990
Pyramids/Pyramids-3000097.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6486d037f2a295d3ae456529ffddaeb17bc770b886f4814dbac8d784bcc874f9
3
+ size 1418184
Pyramids/Pyramids-3000097.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bab95e81181af2db4f5c2845800930e88ea43d31b7d62820281eee3dbe0384e3
3
+ size 8650990
Pyramids/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51932df50e5b0f9d32d33f6708029c22dfe18ab682b7614f71b43a562271b48f
3
  size 8650990
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bab95e81181af2db4f5c2845800930e88ea43d31b7d62820281eee3dbe0384e3
3
  size 8650990
Pyramids/events.out.tfevents.1686376010.26c74d5d5b88.19100.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e0b8f2decced04540661cebdbaef0bd56d9cceba0966641c6916bdb5f5a0321
3
+ size 1423
Pyramids/events.out.tfevents.1686376081.26c74d5d5b88.19458.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1e6526974e527f78fb3a3397f07fbf5cb546f39d141f8dfdd72987a6ee9efef
3
+ size 397213
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"Pyramids": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "rnd": {"gamma": 0.99, "strength": 0.01, "network_settings": {"normalize": false, "hidden_units": 64, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0001, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 1000000, "time_horizon": 128, "summary_freq": 50000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/Pyramids/Pyramids", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Pyramids Training", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"Pyramids": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 500000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "rnd": {"gamma": 0.99, "strength": 0.01, "network_settings": {"normalize": false, "hidden_units": 64, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0001, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 3000000, "time_horizon": 128, "summary_freq": 50000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/Pyramids/Pyramids", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Pyramids Training", "initialize_from": null, "load_model": false, "resume": true, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -51,7 +51,7 @@ behaviors:
51
  init_path: null
52
  keep_checkpoints: 5
53
  even_checkpoints: false
54
- max_steps: 1000000
55
  time_horizon: 128
56
  summary_freq: 50000
57
  threaded: false
@@ -80,7 +80,7 @@ checkpoint_settings:
80
  run_id: Pyramids Training
81
  initialize_from: null
82
  load_model: false
83
- resume: false
84
  force: false
85
  train_model: false
86
  inference: false
 
51
  init_path: null
52
  keep_checkpoints: 5
53
  even_checkpoints: false
54
+ max_steps: 3000000
55
  time_horizon: 128
56
  summary_freq: 50000
57
  threaded: false
 
80
  run_id: Pyramids Training
81
  initialize_from: null
82
  load_model: false
83
+ resume: true
84
  force: false
85
  train_model: false
86
  inference: false
run_logs/Player-0.log CHANGED
@@ -1,12 +1,9 @@
1
  Mono path[0] = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/Managed'
2
  Mono config path = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/MonoBleedingEdge/etc'
3
  Found 1 interfaces on host : 0) 172.28.0.12
4
- Multi-casting "[IP] 172.28.0.12 [Port] 55230 [Flags] 2 [Guid] 3789030083 [EditorId] 764847374 [Version] 1048832 [Id] LinuxPlayer(13,172.28.0.12) [Debug] 0 [PackageName] LinuxPlayer [ProjectName] UnityEnvironment" to [225.0.0.222:54997]...
5
  Preloaded 'lib_burst_generated.so'
6
  Preloaded 'libgrpc_csharp_ext.x64.so'
7
- PlayerPrefs - Creating folder: /root/.config/unity3d/Unity Technologies
8
- PlayerPrefs - Creating folder: /root/.config/unity3d/Unity Technologies/UnityEnvironment
9
- Unable to load player prefs
10
  Initialize engine version: 2021.3.5f1 (40eb3a945986)
11
  [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/UnitySubsystems
12
  Forcing GfxDevice: Null
@@ -36,7 +33,7 @@ ALSA lib pcm.c:2642:(snd_pcm_open_noupdate) Unknown PCM default
36
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
37
  FMOD initialized on nosound output
38
  Begin MonoManager ReloadAssembly
39
- - Completed reload, in 0.085 seconds
40
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
42
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -68,7 +65,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
68
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
69
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
70
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
71
- UnloadTime: 0.836208 ms
72
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
73
  requesting resize 84 x 84
74
  Setting up 1 worker threads for Enlighten.
@@ -76,7 +73,7 @@ PlayerConnection::CleanupMemory Statistics:
76
  [ALLOC_TEMP_TLS] TLS Allocator
77
  StackAllocators :
78
  [ALLOC_TEMP_MAIN]
79
- Peak usage frame count: [16.0 KB-32.0 KB]: 3158 frames, [32.0 KB-64.0 KB]: 12284 frames, [64.0 KB-128.0 KB]: 3326 frames, [2.0 MB-4.0 MB]: 1 frames
80
  Initial Block Size 4.0 MB
81
  Current Block Size 4.0 MB
82
  Peak Allocated Bytes 2.0 MB
@@ -207,28 +204,28 @@ PlayerConnection::CleanupMemory Statistics:
207
  Peak Allocated Bytes 128 B
208
  Overflow Count 0
209
  [ALLOC_MEMORYPROFILER]
210
- Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 18470 frames
211
  Requested Block Size 1.0 MB
212
  Peak Block count 2
213
- Peak Allocated memory 1.6 MB
214
  Peak Large allocation bytes 0 B
215
  [ALLOC_DEFAULT] Dual Thread Allocator
216
- Peak main deferred allocation count 12590
217
  [ALLOC_BUCKET]
218
  Large Block size 4.0 MB
219
  Used Block count 1
220
- Peak Allocated bytes 1.7 MB
221
  [ALLOC_DEFAULT_MAIN]
222
- Peak usage frame count: [8.0 MB-16.0 MB]: 1 frames, [16.0 MB-32.0 MB]: 18768 frames
223
  Requested Block Size 16.0 MB
224
- Peak Block count 3
225
- Peak Allocated memory 29.7 MB
226
  Peak Large allocation bytes 0 B
227
  [ALLOC_DEFAULT_THREAD]
228
- Peak usage frame count: [16.0 MB-32.0 MB]: 18769 frames
229
  Requested Block Size 16.0 MB
230
  Peak Block count 1
231
- Peak Allocated memory 23.5 MB
232
  Peak Large allocation bytes 16.0 MB
233
  [ALLOC_TEMP_JOB_1_FRAME]
234
  Initial Block Size 2.0 MB
@@ -255,15 +252,15 @@ PlayerConnection::CleanupMemory Statistics:
255
  [ALLOC_BUCKET]
256
  Large Block size 4.0 MB
257
  Used Block count 1
258
- Peak Allocated bytes 1.7 MB
259
  [ALLOC_GFX_MAIN]
260
- Peak usage frame count: [32.0 KB-64.0 KB]: 18768 frames, [64.0 KB-128.0 KB]: 1 frames
261
  Requested Block Size 16.0 MB
262
  Peak Block count 1
263
  Peak Allocated memory 67.0 KB
264
  Peak Large allocation bytes 0 B
265
  [ALLOC_GFX_THREAD]
266
- Peak usage frame count: [64.0 KB-128.0 KB]: 18769 frames
267
  Requested Block Size 16.0 MB
268
  Peak Block count 1
269
  Peak Allocated memory 67.7 KB
@@ -273,15 +270,15 @@ PlayerConnection::CleanupMemory Statistics:
273
  [ALLOC_BUCKET]
274
  Large Block size 4.0 MB
275
  Used Block count 1
276
- Peak Allocated bytes 1.7 MB
277
  [ALLOC_CACHEOBJECTS_MAIN]
278
- Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 18470 frames
279
  Requested Block Size 4.0 MB
280
  Peak Block count 1
281
  Peak Allocated memory 1.4 MB
282
  Peak Large allocation bytes 0 B
283
  [ALLOC_CACHEOBJECTS_THREAD]
284
- Peak usage frame count: [1.0 MB-2.0 MB]: 18768 frames, [4.0 MB-8.0 MB]: 1 frames
285
  Requested Block Size 4.0 MB
286
  Peak Block count 2
287
  Peak Allocated memory 4.7 MB
@@ -291,21 +288,21 @@ PlayerConnection::CleanupMemory Statistics:
291
  [ALLOC_BUCKET]
292
  Large Block size 4.0 MB
293
  Used Block count 1
294
- Peak Allocated bytes 1.7 MB
295
  [ALLOC_TYPETREE_MAIN]
296
- Peak usage frame count: [0-1.0 KB]: 18769 frames
297
  Requested Block Size 2.0 MB
298
  Peak Block count 1
299
  Peak Allocated memory 1.0 KB
300
  Peak Large allocation bytes 0 B
301
  [ALLOC_TYPETREE_THREAD]
302
- Peak usage frame count: [1.0 KB-2.0 KB]: 18769 frames
303
  Requested Block Size 2.0 MB
304
  Peak Block count 1
305
  Peak Allocated memory 1.9 KB
306
  Peak Large allocation bytes 0 B
307
  [ALLOC_PROFILER]
308
- Peak usage frame count: [16.0 KB-32.0 KB]: 18769 frames
309
  Requested Block Size 16.0 MB
310
  Peak Block count 1
311
  Peak Allocated memory 29.2 KB
@@ -314,4 +311,4 @@ PlayerConnection::CleanupMemory Statistics:
314
  Large Block size 4.0 MB
315
  Used Block count 1
316
  Peak Allocated bytes 396 B
317
- ##utp:{"type":"MemoryLeaks","version":2,"phase":"Immediate","time":1686375302133,"processId":7249,"allocatedMemory":1970133,"memoryLabels":[{"Default":9033},{"Permanent":1264},{"NewDelete":12801},{"Thread":34460},{"Manager":10603},{"VertexData":12},{"Geometry":280},{"Texture":16},{"Shader":69173},{"Material":24},{"GfxDevice":35248},{"Animation":304},{"Audio":3976},{"Physics":288},{"Serialization":216},{"Input":9176},{"JobScheduler":200},{"Mono":40},{"ScriptingNativeRuntime":216},{"BaseObject":1609212},{"Resource":592},{"Renderer":1936},{"Transform":48},{"File":800},{"WebCam":24},{"Culling":40},{"Terrain":953},{"Wind":24},{"String":3447},{"DynamicArray":30868},{"HashMap":7680},{"Utility":1360},{"PoolAlloc":1160},{"TypeTree":1792},{"ScriptManager":80},{"RuntimeInitializeOnLoadManager":72},{"SpriteAtlas":112},{"GI":3272},{"Unet":16},{"Director":7760},{"WebRequest":720},{"VR":45473},{"SceneManager":424},{"Video":32},{"LazyScriptCache":32},{"NativeArray":384},{"Camera":25},{"Secure":1},{"SerializationCache":624},{"APIUpdating":5872},{"Subsystems":384},{"VirtualTexturing":57552},{"AssetReference":32}]}
 
1
  Mono path[0] = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/Managed'
2
  Mono config path = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/MonoBleedingEdge/etc'
3
  Found 1 interfaces on host : 0) 172.28.0.12
4
+ Multi-casting "[IP] 172.28.0.12 [Port] 55337 [Flags] 2 [Guid] 1771124001 [EditorId] 764847374 [Version] 1048832 [Id] LinuxPlayer(13,172.28.0.12) [Debug] 0 [PackageName] LinuxPlayer [ProjectName] UnityEnvironment" to [225.0.0.222:54997]...
5
  Preloaded 'lib_burst_generated.so'
6
  Preloaded 'libgrpc_csharp_ext.x64.so'
 
 
 
7
  Initialize engine version: 2021.3.5f1 (40eb3a945986)
8
  [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/UnitySubsystems
9
  Forcing GfxDevice: Null
 
33
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
34
  FMOD initialized on nosound output
35
  Begin MonoManager ReloadAssembly
36
+ - Completed reload, in 0.084 seconds
37
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
38
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
65
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
66
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
67
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
68
+ UnloadTime: 1.023544 ms
69
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
70
  requesting resize 84 x 84
71
  Setting up 1 worker threads for Enlighten.
 
73
  [ALLOC_TEMP_TLS] TLS Allocator
74
  StackAllocators :
75
  [ALLOC_TEMP_MAIN]
76
+ Peak usage frame count: [16.0 KB-32.0 KB]: 1109 frames, [32.0 KB-64.0 KB]: 23435 frames, [64.0 KB-128.0 KB]: 12972 frames, [2.0 MB-4.0 MB]: 1 frames
77
  Initial Block Size 4.0 MB
78
  Current Block Size 4.0 MB
79
  Peak Allocated Bytes 2.0 MB
 
204
  Peak Allocated Bytes 128 B
205
  Overflow Count 0
206
  [ALLOC_MEMORYPROFILER]
207
+ Peak usage frame count: [0.5 MB-1.0 MB]: 36 frames, [1.0 MB-2.0 MB]: 37481 frames
208
  Requested Block Size 1.0 MB
209
  Peak Block count 2
210
+ Peak Allocated memory 1.1 MB
211
  Peak Large allocation bytes 0 B
212
  [ALLOC_DEFAULT] Dual Thread Allocator
213
+ Peak main deferred allocation count 11671
214
  [ALLOC_BUCKET]
215
  Large Block size 4.0 MB
216
  Used Block count 1
217
+ Peak Allocated bytes 1.4 MB
218
  [ALLOC_DEFAULT_MAIN]
219
+ Peak usage frame count: [8.0 MB-16.0 MB]: 1 frames, [16.0 MB-32.0 MB]: 37516 frames
220
  Requested Block Size 16.0 MB
221
+ Peak Block count 2
222
+ Peak Allocated memory 24.6 MB
223
  Peak Large allocation bytes 0 B
224
  [ALLOC_DEFAULT_THREAD]
225
+ Peak usage frame count: [16.0 MB-32.0 MB]: 37517 frames
226
  Requested Block Size 16.0 MB
227
  Peak Block count 1
228
+ Peak Allocated memory 22.2 MB
229
  Peak Large allocation bytes 16.0 MB
230
  [ALLOC_TEMP_JOB_1_FRAME]
231
  Initial Block Size 2.0 MB
 
252
  [ALLOC_BUCKET]
253
  Large Block size 4.0 MB
254
  Used Block count 1
255
+ Peak Allocated bytes 1.4 MB
256
  [ALLOC_GFX_MAIN]
257
+ Peak usage frame count: [32.0 KB-64.0 KB]: 37516 frames, [64.0 KB-128.0 KB]: 1 frames
258
  Requested Block Size 16.0 MB
259
  Peak Block count 1
260
  Peak Allocated memory 67.0 KB
261
  Peak Large allocation bytes 0 B
262
  [ALLOC_GFX_THREAD]
263
+ Peak usage frame count: [64.0 KB-128.0 KB]: 37517 frames
264
  Requested Block Size 16.0 MB
265
  Peak Block count 1
266
  Peak Allocated memory 67.7 KB
 
270
  [ALLOC_BUCKET]
271
  Large Block size 4.0 MB
272
  Used Block count 1
273
+ Peak Allocated bytes 1.4 MB
274
  [ALLOC_CACHEOBJECTS_MAIN]
275
+ Peak usage frame count: [0.5 MB-1.0 MB]: 39 frames, [1.0 MB-2.0 MB]: 37478 frames
276
  Requested Block Size 4.0 MB
277
  Peak Block count 1
278
  Peak Allocated memory 1.4 MB
279
  Peak Large allocation bytes 0 B
280
  [ALLOC_CACHEOBJECTS_THREAD]
281
+ Peak usage frame count: [1.0 MB-2.0 MB]: 37516 frames, [4.0 MB-8.0 MB]: 1 frames
282
  Requested Block Size 4.0 MB
283
  Peak Block count 2
284
  Peak Allocated memory 4.7 MB
 
288
  [ALLOC_BUCKET]
289
  Large Block size 4.0 MB
290
  Used Block count 1
291
+ Peak Allocated bytes 1.4 MB
292
  [ALLOC_TYPETREE_MAIN]
293
+ Peak usage frame count: [0-1.0 KB]: 37517 frames
294
  Requested Block Size 2.0 MB
295
  Peak Block count 1
296
  Peak Allocated memory 1.0 KB
297
  Peak Large allocation bytes 0 B
298
  [ALLOC_TYPETREE_THREAD]
299
+ Peak usage frame count: [1.0 KB-2.0 KB]: 37517 frames
300
  Requested Block Size 2.0 MB
301
  Peak Block count 1
302
  Peak Allocated memory 1.9 KB
303
  Peak Large allocation bytes 0 B
304
  [ALLOC_PROFILER]
305
+ Peak usage frame count: [16.0 KB-32.0 KB]: 37517 frames
306
  Requested Block Size 16.0 MB
307
  Peak Block count 1
308
  Peak Allocated memory 29.2 KB
 
311
  Large Block size 4.0 MB
312
  Used Block count 1
313
  Peak Allocated bytes 396 B
314
+ ##utp:{"type":"MemoryLeaks","version":2,"phase":"Immediate","time":1686381186382,"processId":19476,"allocatedMemory":1970133,"memoryLabels":[{"Default":9033},{"Permanent":1264},{"NewDelete":12801},{"Thread":34460},{"Manager":10603},{"VertexData":12},{"Geometry":280},{"Texture":16},{"Shader":69173},{"Material":24},{"GfxDevice":35248},{"Animation":304},{"Audio":3976},{"Physics":288},{"Serialization":216},{"Input":9176},{"JobScheduler":200},{"Mono":40},{"ScriptingNativeRuntime":216},{"BaseObject":1609212},{"Resource":592},{"Renderer":1936},{"Transform":48},{"File":800},{"WebCam":24},{"Culling":40},{"Terrain":953},{"Wind":24},{"String":3447},{"DynamicArray":30868},{"HashMap":7680},{"Utility":1360},{"PoolAlloc":1160},{"TypeTree":1792},{"ScriptManager":80},{"RuntimeInitializeOnLoadManager":72},{"SpriteAtlas":112},{"GI":3272},{"Unet":16},{"Director":7760},{"WebRequest":720},{"VR":45473},{"SceneManager":424},{"Video":32},{"LazyScriptCache":32},{"NativeArray":384},{"Camera":25},{"Secure":1},{"SerializationCache":624},{"APIUpdating":5872},{"Subsystems":384},{"VirtualTexturing":57552},{"AssetReference":32}]}
run_logs/timers.json CHANGED
@@ -2,249 +2,249 @@
2
  "name": "root",
3
  "gauges": {
4
  "Pyramids.Policy.Entropy.mean": {
5
- "value": 0.35135510563850403,
6
- "min": 0.3437347114086151,
7
- "max": 1.2542389631271362,
8
- "count": 20
9
  },
10
  "Pyramids.Policy.Entropy.sum": {
11
- "value": 17573.376953125,
12
- "min": 17159.236328125,
13
- "max": 63554.796875,
14
- "count": 20
 
 
 
 
 
 
 
 
 
 
 
 
15
  },
16
  "Pyramids.Step.mean": {
17
- "value": 999999.0,
18
- "min": 49959.0,
19
- "max": 999999.0,
20
- "count": 20
21
  },
22
  "Pyramids.Step.sum": {
23
- "value": 999999.0,
24
- "min": 49959.0,
25
- "max": 999999.0,
26
- "count": 20
27
  },
28
  "Pyramids.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 0.5636146664619446,
30
- "min": -0.08218594640493393,
31
- "max": 0.5653014779090881,
32
- "count": 20
33
  },
34
  "Pyramids.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 263.2080383300781,
36
- "min": -33.03874969482422,
37
- "max": 263.2080383300781,
38
- "count": 20
39
  },
40
  "Pyramids.Policy.RndValueEstimate.mean": {
41
- "value": 0.007806922774761915,
42
- "min": 0.007806922774761915,
43
- "max": 0.4125056862831116,
44
- "count": 20
45
  },
46
  "Pyramids.Policy.RndValueEstimate.sum": {
47
- "value": 3.6458330154418945,
48
- "min": 3.6458330154418945,
49
- "max": 165.0022735595703,
50
- "count": 20
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
  },
52
  "Pyramids.Losses.PolicyLoss.mean": {
53
- "value": 0.06778426025913278,
54
- "min": 0.06652670221439246,
55
- "max": 0.07334119114734719,
56
- "count": 20
57
  },
58
  "Pyramids.Losses.PolicyLoss.sum": {
59
- "value": 1.6268222462191868,
60
- "min": 1.0267766760628607,
61
- "max": 1.6970636773187957,
62
- "count": 20
63
  },
64
  "Pyramids.Losses.ValueLoss.mean": {
65
- "value": 0.016569182275310998,
66
- "min": 0.0012711221483666876,
67
- "max": 0.01765561954096775,
68
- "count": 20
69
  },
70
  "Pyramids.Losses.ValueLoss.sum": {
71
- "value": 0.39766037460746395,
72
- "min": 0.029235809412433814,
73
- "max": 0.40607924944225826,
74
- "count": 20
75
  },
76
  "Pyramids.Policy.LearningRate.mean": {
77
- "value": 7.581272472941665e-06,
78
- "min": 7.581272472941665e-06,
79
- "max": 0.00029159603851560713,
80
- "count": 20
81
  },
82
  "Pyramids.Policy.LearningRate.sum": {
83
- "value": 0.00018195053935059997,
84
- "min": 0.00018195053935059997,
85
- "max": 0.0060378704873766,
86
- "count": 20
87
  },
88
  "Pyramids.Policy.Epsilon.mean": {
89
- "value": 0.10252705833333335,
90
- "min": 0.10252705833333335,
91
- "max": 0.19719867857142856,
92
- "count": 20
93
  },
94
  "Pyramids.Policy.Epsilon.sum": {
95
- "value": 2.4606494000000003,
96
- "min": 2.4606494000000003,
97
- "max": 4.3126234000000006,
98
- "count": 20
99
  },
100
  "Pyramids.Policy.Beta.mean": {
101
- "value": 0.0002624531275,
102
- "min": 0.0002624531275,
103
- "max": 0.009720147989285716,
104
- "count": 20
105
  },
106
  "Pyramids.Policy.Beta.sum": {
107
- "value": 0.0062988750599999995,
108
- "min": 0.0062988750599999995,
109
- "max": 0.20129107766,
110
- "count": 20
111
  },
112
  "Pyramids.Losses.RNDLoss.mean": {
113
- "value": 0.009527774527668953,
114
- "min": 0.009527774527668953,
115
- "max": 0.41762715578079224,
116
- "count": 20
117
  },
118
  "Pyramids.Losses.RNDLoss.sum": {
119
- "value": 0.22866658866405487,
120
- "min": 0.22866658866405487,
121
- "max": 5.846780300140381,
122
- "count": 20
123
- },
124
- "Pyramids.Environment.EpisodeLength.mean": {
125
- "value": 339.97972972972974,
126
- "min": 338.56643356643355,
127
- "max": 997.1458333333334,
128
- "count": 20
129
- },
130
- "Pyramids.Environment.EpisodeLength.sum": {
131
- "value": 50317.0,
132
- "min": 47051.0,
133
- "max": 52579.0,
134
- "count": 20
135
- },
136
- "Pyramids.Environment.CumulativeReward.mean": {
137
- "value": 1.592208142639423,
138
- "min": -0.9564417197058598,
139
- "max": 1.6194615204121683,
140
- "count": 20
141
- },
142
- "Pyramids.Environment.CumulativeReward.sum": {
143
- "value": 234.05459696799517,
144
- "min": -45.90920254588127,
145
- "max": 234.05459696799517,
146
- "count": 20
147
- },
148
- "Pyramids.Policy.ExtrinsicReward.mean": {
149
- "value": 1.592208142639423,
150
- "min": -0.9564417197058598,
151
- "max": 1.6194615204121683,
152
- "count": 20
153
- },
154
- "Pyramids.Policy.ExtrinsicReward.sum": {
155
- "value": 234.05459696799517,
156
- "min": -45.90920254588127,
157
- "max": 234.05459696799517,
158
- "count": 20
159
- },
160
- "Pyramids.Policy.RndReward.mean": {
161
- "value": 0.03380435873202637,
162
- "min": 0.03380435873202637,
163
- "max": 6.084363122625898,
164
- "count": 20
165
- },
166
- "Pyramids.Policy.RndReward.sum": {
167
- "value": 4.969240733607876,
168
- "min": 4.969240733607876,
169
- "max": 292.0494298860431,
170
- "count": 20
171
  },
172
  "Pyramids.IsTraining.mean": {
173
  "value": 1.0,
174
  "min": 1.0,
175
  "max": 1.0,
176
- "count": 20
177
  },
178
  "Pyramids.IsTraining.sum": {
179
  "value": 1.0,
180
  "min": 1.0,
181
  "max": 1.0,
182
- "count": 20
183
  }
184
  },
185
  "metadata": {
186
  "timer_format_version": "0.1.0",
187
- "start_time_seconds": "1686373182",
188
  "python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
189
- "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
190
  "mlagents_version": "0.31.0.dev0",
191
  "mlagents_envs_version": "0.31.0.dev0",
192
  "communication_protocol_version": "1.5.0",
193
  "pytorch_version": "1.11.0+cu102",
194
  "numpy_version": "1.21.2",
195
- "end_time_seconds": "1686375302"
196
  },
197
- "total": 2119.3034031409998,
198
  "count": 1,
199
- "self": 0.47542574499948387,
200
  "children": {
201
  "run_training.setup": {
202
- "total": 0.03698148500006937,
203
  "count": 1,
204
- "self": 0.03698148500006937
205
  },
206
  "TrainerController.start_learning": {
207
- "total": 2118.790995911,
208
  "count": 1,
209
- "self": 1.2670965880029144,
210
  "children": {
211
  "TrainerController._reset_env": {
212
- "total": 3.9749093469999934,
213
  "count": 1,
214
- "self": 3.9749093469999934
215
  },
216
  "TrainerController.advance": {
217
- "total": 2113.457703399997,
218
- "count": 63873,
219
- "self": 1.2667271999130207,
220
  "children": {
221
  "env_step": {
222
- "total": 1475.7471504740952,
223
- "count": 63873,
224
- "self": 1368.2087681150745,
225
  "children": {
226
  "SubprocessEnvManager._take_step": {
227
- "total": 106.81009274905068,
228
- "count": 63873,
229
- "self": 4.592955099020855,
230
  "children": {
231
  "TorchPolicy.evaluate": {
232
- "total": 102.21713765002983,
233
- "count": 62561,
234
- "self": 102.21713765002983
235
  }
236
  }
237
  },
238
  "workers": {
239
- "total": 0.7282896099700338,
240
- "count": 63873,
241
  "self": 0.0,
242
  "children": {
243
  "worker_root": {
244
- "total": 2114.08891636499,
245
- "count": 63873,
246
  "is_parallel": true,
247
- "self": 854.0269211289844,
248
  "children": {
249
  "run_training.setup": {
250
  "total": 0.0,
@@ -253,48 +253,48 @@
253
  "self": 0.0,
254
  "children": {
255
  "steps_from_proto": {
256
- "total": 0.0017286979998516472,
257
  "count": 1,
258
  "is_parallel": true,
259
- "self": 0.0005476689993884065,
260
  "children": {
261
  "_process_rank_one_or_two_observation": {
262
- "total": 0.0011810290004632407,
263
  "count": 8,
264
  "is_parallel": true,
265
- "self": 0.0011810290004632407
266
  }
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 0.07697923999990053,
271
  "count": 1,
272
  "is_parallel": true,
273
- "self": 0.0006119560000570345,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 0.00046551499985980627,
277
  "count": 1,
278
  "is_parallel": true,
279
- "self": 0.00046551499985980627
280
  },
281
  "communicator.exchange": {
282
- "total": 0.0740684570000667,
283
  "count": 1,
284
  "is_parallel": true,
285
- "self": 0.0740684570000667
286
  },
287
  "steps_from_proto": {
288
- "total": 0.0018333119999169867,
289
  "count": 1,
290
  "is_parallel": true,
291
- "self": 0.0003654529994037148,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 0.001467859000513272,
295
  "count": 8,
296
  "is_parallel": true,
297
- "self": 0.001467859000513272
298
  }
299
  }
300
  }
@@ -303,34 +303,34 @@
303
  }
304
  },
305
  "UnityEnvironment.step": {
306
- "total": 1260.0619952360055,
307
- "count": 63872,
308
  "is_parallel": true,
309
- "self": 32.321872371902145,
310
  "children": {
311
  "UnityEnvironment._generate_step_input": {
312
- "total": 21.65442949095427,
313
- "count": 63872,
314
  "is_parallel": true,
315
- "self": 21.65442949095427
316
  },
317
  "communicator.exchange": {
318
- "total": 1109.908463243082,
319
- "count": 63872,
320
  "is_parallel": true,
321
- "self": 1109.908463243082
322
  },
323
  "steps_from_proto": {
324
- "total": 96.17723013006707,
325
- "count": 63872,
326
  "is_parallel": true,
327
- "self": 18.477071257100533,
328
  "children": {
329
  "_process_rank_one_or_two_observation": {
330
- "total": 77.70015887296654,
331
- "count": 510976,
332
  "is_parallel": true,
333
- "self": 77.70015887296654
334
  }
335
  }
336
  }
@@ -343,31 +343,31 @@
343
  }
344
  },
345
  "trainer_advance": {
346
- "total": 636.4438257259887,
347
- "count": 63873,
348
- "self": 2.46895552301703,
349
  "children": {
350
  "process_trajectory": {
351
- "total": 106.82426519397814,
352
- "count": 63873,
353
- "self": 106.61683175897815,
354
  "children": {
355
  "RLTrainer._checkpoint": {
356
- "total": 0.20743343499998446,
357
- "count": 2,
358
- "self": 0.20743343499998446
359
  }
360
  }
361
  },
362
  "_update_policy": {
363
- "total": 527.1506050089936,
364
- "count": 454,
365
- "self": 336.0804614169881,
366
  "children": {
367
  "TorchPPOOptimizer.update": {
368
- "total": 191.07014359200548,
369
- "count": 22845,
370
- "self": 191.07014359200548
371
  }
372
  }
373
  }
@@ -376,19 +376,19 @@
376
  }
377
  },
378
  "trainer_threads": {
379
- "total": 9.089999366551638e-07,
380
  "count": 1,
381
- "self": 9.089999366551638e-07
382
  },
383
  "TrainerController._save_models": {
384
- "total": 0.09128566700019292,
385
  "count": 1,
386
- "self": 0.0013680140000360552,
387
  "children": {
388
  "RLTrainer._checkpoint": {
389
- "total": 0.08991765300015686,
390
  "count": 1,
391
- "self": 0.08991765300015686
392
  }
393
  }
394
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "Pyramids.Policy.Entropy.mean": {
5
+ "value": 0.13053016364574432,
6
+ "min": 0.13053016364574432,
7
+ "max": 0.34393319487571716,
8
+ "count": 40
9
  },
10
  "Pyramids.Policy.Entropy.sum": {
11
+ "value": 6543.2158203125,
12
+ "min": 6508.48681640625,
13
+ "max": 17504.82421875,
14
+ "count": 40
15
+ },
16
+ "Pyramids.Environment.EpisodeLength.mean": {
17
+ "value": 226.53982300884957,
18
+ "min": 217.42918454935622,
19
+ "max": 335.7635135135135,
20
+ "count": 40
21
+ },
22
+ "Pyramids.Environment.EpisodeLength.sum": {
23
+ "value": 51198.0,
24
+ "min": 46619.0,
25
+ "max": 51401.0,
26
+ "count": 40
27
  },
28
  "Pyramids.Step.mean": {
29
+ "value": 2999969.0,
30
+ "min": 1049963.0,
31
+ "max": 2999969.0,
32
+ "count": 40
33
  },
34
  "Pyramids.Step.sum": {
35
+ "value": 2999969.0,
36
+ "min": 1049963.0,
37
+ "max": 2999969.0,
38
+ "count": 40
39
  },
40
  "Pyramids.Policy.ExtrinsicValueEstimate.mean": {
41
+ "value": 0.8029366135597229,
42
+ "min": 0.5550001859664917,
43
+ "max": 0.8423917889595032,
44
+ "count": 40
45
  },
46
  "Pyramids.Policy.ExtrinsicValueEstimate.sum": {
47
+ "value": 407.0888671875,
48
+ "min": 256.965087890625,
49
+ "max": 432.9893798828125,
50
+ "count": 40
51
  },
52
  "Pyramids.Policy.RndValueEstimate.mean": {
53
+ "value": 0.016411546617746353,
54
+ "min": -0.010096848011016846,
55
+ "max": 0.0640997514128685,
56
+ "count": 40
57
  },
58
  "Pyramids.Policy.RndValueEstimate.sum": {
59
+ "value": 8.320653915405273,
60
+ "min": -4.907068252563477,
61
+ "max": 29.678184509277344,
62
+ "count": 40
63
+ },
64
+ "Pyramids.Environment.CumulativeReward.mean": {
65
+ "value": 1.7557530854457775,
66
+ "min": 1.5966486237641122,
67
+ "max": 1.7725354446606203,
68
+ "count": 40
69
+ },
70
+ "Pyramids.Environment.CumulativeReward.sum": {
71
+ "value": 396.8001973107457,
72
+ "min": 236.3039963170886,
73
+ "max": 413.1451980918646,
74
+ "count": 40
75
+ },
76
+ "Pyramids.Policy.ExtrinsicReward.mean": {
77
+ "value": 1.7557530854457775,
78
+ "min": 1.5966486237641122,
79
+ "max": 1.7725354446606203,
80
+ "count": 40
81
+ },
82
+ "Pyramids.Policy.ExtrinsicReward.sum": {
83
+ "value": 396.8001973107457,
84
+ "min": 236.3039963170886,
85
+ "max": 413.1451980918646,
86
+ "count": 40
87
+ },
88
+ "Pyramids.Policy.RndReward.mean": {
89
+ "value": 0.016251778906775368,
90
+ "min": 0.015597144257745558,
91
+ "max": 0.034564408739219156,
92
+ "count": 40
93
+ },
94
+ "Pyramids.Policy.RndReward.sum": {
95
+ "value": 3.6729020329312334,
96
+ "min": 3.4313717367040226,
97
+ "max": 5.288354537100531,
98
+ "count": 40
99
  },
100
  "Pyramids.Losses.PolicyLoss.mean": {
101
+ "value": 0.06941192193508731,
102
+ "min": 0.06467020634631303,
103
+ "max": 0.07191291106935192,
104
+ "count": 40
105
  },
106
  "Pyramids.Losses.PolicyLoss.sum": {
107
+ "value": 1.6658861264420957,
108
+ "min": 1.5020420119959725,
109
+ "max": 1.725909865664446,
110
+ "count": 40
111
  },
112
  "Pyramids.Losses.ValueLoss.mean": {
113
+ "value": 0.015384109843699404,
114
+ "min": 0.012953029207141222,
115
+ "max": 0.01637407874148907,
116
+ "count": 40
117
  },
118
  "Pyramids.Losses.ValueLoss.sum": {
119
+ "value": 0.3692186362487857,
120
+ "min": 0.2979196717642481,
121
+ "max": 0.38210905112175486,
122
+ "count": 40
123
  },
124
  "Pyramids.Policy.LearningRate.mean": {
125
+ "value": 2.429511690195832e-06,
126
+ "min": 2.429511690195832e-06,
127
+ "max": 0.00019744786896666374,
128
+ "count": 40
129
  },
130
  "Pyramids.Policy.LearningRate.sum": {
131
+ "value": 5.830828056469996e-05,
132
+ "min": 5.830828056469996e-05,
133
+ "max": 0.004619409360197167,
134
+ "count": 40
135
  },
136
  "Pyramids.Policy.Epsilon.mean": {
137
+ "value": 0.10080980416666667,
138
+ "min": 0.10080980416666667,
139
+ "max": 0.16581594492753626,
140
+ "count": 40
141
  },
142
  "Pyramids.Policy.Epsilon.sum": {
143
+ "value": 2.4194353,
144
+ "min": 2.4194353,
145
+ "max": 3.9398028333333333,
146
+ "count": 40
147
  },
148
  "Pyramids.Policy.Beta.mean": {
149
+ "value": 9.089943624999996e-05,
150
+ "min": 9.089943624999996e-05,
151
+ "max": 0.0065850128982608705,
152
+ "count": 40
153
  },
154
  "Pyramids.Policy.Beta.sum": {
155
+ "value": 0.002181586469999999,
156
+ "min": 0.002181586469999999,
157
+ "max": 0.15406630305000002,
158
+ "count": 40
159
  },
160
  "Pyramids.Losses.RNDLoss.mean": {
161
+ "value": 0.006933924276381731,
162
+ "min": 0.006791813764721155,
163
+ "max": 0.010334798134863377,
164
+ "count": 40
165
  },
166
  "Pyramids.Losses.RNDLoss.sum": {
167
+ "value": 0.16641418635845184,
168
+ "min": 0.16214092075824738,
169
+ "max": 0.23770035803318024,
170
+ "count": 40
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
171
  },
172
  "Pyramids.IsTraining.mean": {
173
  "value": 1.0,
174
  "min": 1.0,
175
  "max": 1.0,
176
+ "count": 40
177
  },
178
  "Pyramids.IsTraining.sum": {
179
  "value": 1.0,
180
  "min": 1.0,
181
  "max": 1.0,
182
+ "count": 40
183
  }
184
  },
185
  "metadata": {
186
  "timer_format_version": "0.1.0",
187
+ "start_time_seconds": "1686376080",
188
  "python_version": "3.10.12 (main, Jun 7 2023, 12:45:35) [GCC 9.4.0]",
189
+ "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics --resume",
190
  "mlagents_version": "0.31.0.dev0",
191
  "mlagents_envs_version": "0.31.0.dev0",
192
  "communication_protocol_version": "1.5.0",
193
  "pytorch_version": "1.11.0+cu102",
194
  "numpy_version": "1.21.2",
195
+ "end_time_seconds": "1686381186"
196
  },
197
+ "total": 5105.509026824,
198
  "count": 1,
199
+ "self": 0.441498130000582,
200
  "children": {
201
  "run_training.setup": {
202
+ "total": 0.03876343899992207,
203
  "count": 1,
204
+ "self": 0.03876343899992207
205
  },
206
  "TrainerController.start_learning": {
207
+ "total": 5105.028765255,
208
  "count": 1,
209
+ "self": 2.787783518168908,
210
  "children": {
211
  "TrainerController._reset_env": {
212
+ "total": 4.312965602999611,
213
  "count": 1,
214
+ "self": 4.312965602999611
215
  },
216
  "TrainerController.advance": {
217
+ "total": 5097.834866287832,
218
+ "count": 131136,
219
+ "self": 2.8199113848386332,
220
  "children": {
221
  "env_step": {
222
+ "total": 3809.999547644815,
223
+ "count": 131136,
224
+ "self": 3591.752417211447,
225
  "children": {
226
  "SubprocessEnvManager._take_step": {
227
+ "total": 216.53971534624998,
228
+ "count": 131136,
229
+ "self": 9.37691775055373,
230
  "children": {
231
  "TorchPolicy.evaluate": {
232
+ "total": 207.16279759569625,
233
+ "count": 125056,
234
+ "self": 207.16279759569625
235
  }
236
  }
237
  },
238
  "workers": {
239
+ "total": 1.7074150871176244,
240
+ "count": 131136,
241
  "self": 0.0,
242
  "children": {
243
  "worker_root": {
244
+ "total": 5093.981040424949,
245
+ "count": 131136,
246
  "is_parallel": true,
247
+ "self": 1737.400396409189,
248
  "children": {
249
  "run_training.setup": {
250
  "total": 0.0,
 
253
  "self": 0.0,
254
  "children": {
255
  "steps_from_proto": {
256
+ "total": 0.001776224000423099,
257
  "count": 1,
258
  "is_parallel": true,
259
+ "self": 0.0005706380006813561,
260
  "children": {
261
  "_process_rank_one_or_two_observation": {
262
+ "total": 0.0012055859997417429,
263
  "count": 8,
264
  "is_parallel": true,
265
+ "self": 0.0012055859997417429
266
  }
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 0.08704309699987789,
271
  "count": 1,
272
  "is_parallel": true,
273
+ "self": 0.0005678710003849119,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 0.00043306200041115517,
277
  "count": 1,
278
  "is_parallel": true,
279
+ "self": 0.00043306200041115517
280
  },
281
  "communicator.exchange": {
282
+ "total": 0.08425388799969369,
283
  "count": 1,
284
  "is_parallel": true,
285
+ "self": 0.08425388799969369
286
  },
287
  "steps_from_proto": {
288
+ "total": 0.0017882759993881336,
289
  "count": 1,
290
  "is_parallel": true,
291
+ "self": 0.0003485019988147542,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 0.0014397740005733795,
295
  "count": 8,
296
  "is_parallel": true,
297
+ "self": 0.0014397740005733795
298
  }
299
  }
300
  }
 
303
  }
304
  },
305
  "UnityEnvironment.step": {
306
+ "total": 3356.5806440157603,
307
+ "count": 131135,
308
  "is_parallel": true,
309
+ "self": 66.04875748262566,
310
  "children": {
311
  "UnityEnvironment._generate_step_input": {
312
+ "total": 44.59532830708031,
313
+ "count": 131135,
314
  "is_parallel": true,
315
+ "self": 44.59532830708031
316
  },
317
  "communicator.exchange": {
318
+ "total": 3043.2339099858636,
319
+ "count": 131135,
320
  "is_parallel": true,
321
+ "self": 3043.2339099858636
322
  },
323
  "steps_from_proto": {
324
+ "total": 202.70264824019068,
325
+ "count": 131135,
326
  "is_parallel": true,
327
+ "self": 40.38878473092791,
328
  "children": {
329
  "_process_rank_one_or_two_observation": {
330
+ "total": 162.31386350926277,
331
+ "count": 1049080,
332
  "is_parallel": true,
333
+ "self": 162.31386350926277
334
  }
335
  }
336
  }
 
343
  }
344
  },
345
  "trainer_advance": {
346
+ "total": 1285.0154072581781,
347
+ "count": 131136,
348
+ "self": 5.809300956127117,
349
  "children": {
350
  "process_trajectory": {
351
+ "total": 224.75797619306013,
352
+ "count": 131136,
353
+ "self": 224.31268628105954,
354
  "children": {
355
  "RLTrainer._checkpoint": {
356
+ "total": 0.44528991200058954,
357
+ "count": 4,
358
+ "self": 0.44528991200058954
359
  }
360
  }
361
  },
362
  "_update_policy": {
363
+ "total": 1054.4481301089909,
364
+ "count": 947,
365
+ "self": 678.76073158895,
366
  "children": {
367
  "TorchPPOOptimizer.update": {
368
+ "total": 375.6873985200409,
369
+ "count": 45561,
370
+ "self": 375.6873985200409
371
  }
372
  }
373
  }
 
376
  }
377
  },
378
  "trainer_threads": {
379
+ "total": 9.899995347950608e-07,
380
  "count": 1,
381
+ "self": 9.899995347950608e-07
382
  },
383
  "TrainerController._save_models": {
384
+ "total": 0.09314885599997069,
385
  "count": 1,
386
+ "self": 0.0017387700008839602,
387
  "children": {
388
  "RLTrainer._checkpoint": {
389
+ "total": 0.09141008599908673,
390
  "count": 1,
391
+ "self": 0.09141008599908673
392
  }
393
  }
394
  }
run_logs/training_status.json CHANGED
@@ -2,40 +2,58 @@
2
  "Pyramids": {
3
  "checkpoints": [
4
  {
5
- "steps": 499897,
6
- "file_path": "results/Pyramids Training/Pyramids/Pyramids-499897.onnx",
7
- "reward": null,
8
- "creation_time": 1686374162.5413673,
 
 
 
 
 
 
 
 
 
9
  "auxillary_file_paths": [
10
- "results/Pyramids Training/Pyramids/Pyramids-499897.pt"
11
  ]
12
  },
13
  {
14
- "steps": 999999,
15
- "file_path": "results/Pyramids Training/Pyramids/Pyramids-999999.onnx",
16
- "reward": 1.5879999846220016,
17
- "creation_time": 1686375301.5903716,
18
  "auxillary_file_paths": [
19
- "results/Pyramids Training/Pyramids/Pyramids-999999.pt"
20
  ]
21
  },
22
  {
23
- "steps": 1000120,
24
- "file_path": "results/Pyramids Training/Pyramids/Pyramids-1000120.onnx",
25
- "reward": 1.5951999843120575,
26
- "creation_time": 1686375301.6904116,
 
 
 
 
 
 
 
 
 
27
  "auxillary_file_paths": [
28
- "results/Pyramids Training/Pyramids/Pyramids-1000120.pt"
29
  ]
30
  }
31
  ],
32
  "final_checkpoint": {
33
- "steps": 1000120,
34
  "file_path": "results/Pyramids Training/Pyramids.onnx",
35
- "reward": 1.5951999843120575,
36
- "creation_time": 1686375301.6904116,
37
  "auxillary_file_paths": [
38
- "results/Pyramids Training/Pyramids/Pyramids-1000120.pt"
39
  ]
40
  }
41
  },
 
2
  "Pyramids": {
3
  "checkpoints": [
4
  {
5
+ "steps": 1499996,
6
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-1499996.onnx",
7
+ "reward": 1.7666250076144934,
8
+ "creation_time": 1686377309.325065,
9
+ "auxillary_file_paths": [
10
+ "results/Pyramids Training/Pyramids/Pyramids-1499996.pt"
11
+ ]
12
+ },
13
+ {
14
+ "steps": 1999977,
15
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-1999977.onnx",
16
+ "reward": 1.4659999534487724,
17
+ "creation_time": 1686378606.959674,
18
  "auxillary_file_paths": [
19
+ "results/Pyramids Training/Pyramids/Pyramids-1999977.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 2499973,
24
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-2499973.onnx",
25
+ "reward": 1.8709999471902847,
26
+ "creation_time": 1686379894.1080823,
27
  "auxillary_file_paths": [
28
+ "results/Pyramids Training/Pyramids/Pyramids-2499973.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 2999969,
33
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-2999969.onnx",
34
+ "reward": null,
35
+ "creation_time": 1686381185.8684719,
36
+ "auxillary_file_paths": [
37
+ "results/Pyramids Training/Pyramids/Pyramids-2999969.pt"
38
+ ]
39
+ },
40
+ {
41
+ "steps": 3000097,
42
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-3000097.onnx",
43
+ "reward": null,
44
+ "creation_time": 1686381185.972116,
45
  "auxillary_file_paths": [
46
+ "results/Pyramids Training/Pyramids/Pyramids-3000097.pt"
47
  ]
48
  }
49
  ],
50
  "final_checkpoint": {
51
+ "steps": 3000097,
52
  "file_path": "results/Pyramids Training/Pyramids.onnx",
53
+ "reward": null,
54
+ "creation_time": 1686381185.972116,
55
  "auxillary_file_paths": [
56
+ "results/Pyramids Training/Pyramids/Pyramids-3000097.pt"
57
  ]
58
  }
59
  },