talebzeghmi commited on
Commit
d58868e
·
verified ·
1 Parent(s): 6e36c7b

First Push

Browse files
Pyramids.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11b498a7286dc1cdee818015e0753fc4ca6447073b82215ee05a77a4310789de
3
+ size 1420238
Pyramids/Pyramids-269914.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d6ec9c1284ad7a0cf73024f891ce2221e96f39d69c52b1a02f18ddcd173f360
3
+ size 1420238
Pyramids/Pyramids-269914.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2455244dfebc8497ff5ac1ec07ad233b756f9ac8680c2e7309e088c82f375a57
3
+ size 8656000
Pyramids/Pyramids-279912.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7034a5348b5f7341b0c4575382486b01a4503e3050646ee4b304cdfac759a16e
3
+ size 1420238
Pyramids/Pyramids-279912.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1251669bea60092ffe56e6f813f17f215e359cc02c4a4e81873be68e223d645b
3
+ size 8656000
Pyramids/Pyramids-289923.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ead8d7ce8e7affa26c012f1e7e457b4923cc3de52473639a249e25068ad7b37
3
+ size 1420238
Pyramids/Pyramids-289923.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26dd1babb5db5c22b94e9b3dd2bbc0028c79cc9ddbd05437b41e233cbab52d95
3
+ size 8656000
Pyramids/Pyramids-299997.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cd69c8f4e18947c846ab3ebb21f0da902f70eb2d7f8ef36349fb292ae4ef060
3
+ size 1420238
Pyramids/Pyramids-299997.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bbfec6653264bc32dae9f84936f9df5421ef7808472aeb635f176beff208dd5
3
+ size 8656000
Pyramids/Pyramids-303333.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11b498a7286dc1cdee818015e0753fc4ca6447073b82215ee05a77a4310789de
3
+ size 1420238
Pyramids/Pyramids-303333.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b867a42cc40b7613f24d91dbb13b35566f9edb6172d6ba3de95221283bcc7c9
3
+ size 8656000
Pyramids/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fda7957128f27ee58c423998659c819fbea64d68627e704d1128b64aab4a5fe
3
+ size 8655605
Pyramids/events.out.tfevents.1772324510.0e9230f42a0c.9013.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a9e835eacb46d46903ad145809e889a04677729f2e38253abd5cdc4cf1c4c4b
3
+ size 85806
README.md CHANGED
@@ -1,14 +1,14 @@
1
  ---
2
  library_name: ml-agents
3
  tags:
4
- - SnowballTarget
5
  - deep-reinforcement-learning
6
  - reinforcement-learning
7
- - ML-Agents-SnowballTarget
8
  ---
9
 
10
- # **ppo** Agent playing **SnowballTarget**
11
- This is a trained model of a **ppo** agent playing **SnowballTarget**
12
  using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
 
14
  ## Usage (with ML-Agents)
 
1
  ---
2
  library_name: ml-agents
3
  tags:
4
+ - Pyramids
5
  - deep-reinforcement-learning
6
  - reinforcement-learning
7
+ - ML-Agents-Pyramids
8
  ---
9
 
10
+ # **ppo** Agent playing **Pyramids**
11
+ This is a trained model of a **ppo** agent playing **Pyramids**
12
  using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
13
 
14
  ## Usage (with ML-Agents)
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0006, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 400000, "time_horizon": 64, "summary_freq": 10000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget3", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"Pyramids": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 384, "buffer_size": 2048, "learning_rate": 0.0006, "beta": 0.01, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 10000, "network_settings": {"normalize": false, "hidden_units": 512, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}, "rnd": {"gamma": 0.99, "strength": 0.01, "network_settings": {"normalize": false, "hidden_units": 64, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "learning_rate": 0.0001, "encoding_size": null}}, "init_path": null, "keep_checkpoints": 5, "even_checkpoints": false, "max_steps": 1000000, "time_horizon": 128, "summary_freq": 30000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/Pyramids/Pyramids", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Pyramids Training1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -1,12 +1,12 @@
1
  default_settings: null
2
  behaviors:
3
- SnowballTarget:
4
  trainer_type: ppo
5
  hyperparameters:
6
- batch_size: 128
7
  buffer_size: 2048
8
  learning_rate: 0.0006
9
- beta: 0.005
10
  epsilon: 0.2
11
  lambd: 0.95
12
  num_epoch: 3
@@ -14,11 +14,11 @@ behaviors:
14
  learning_rate_schedule: linear
15
  beta_schedule: linear
16
  epsilon_schedule: linear
17
- checkpoint_interval: 50000
18
  network_settings:
19
  normalize: false
20
- hidden_units: 256
21
- num_layers: 3
22
  vis_encode_type: simple
23
  memory: null
24
  goal_conditioning_type: hyper
@@ -35,17 +35,30 @@ behaviors:
35
  memory: null
36
  goal_conditioning_type: hyper
37
  deterministic: false
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  init_path: null
39
- keep_checkpoints: 10
40
  even_checkpoints: false
41
- max_steps: 400000
42
- time_horizon: 64
43
- summary_freq: 10000
44
  threaded: false
45
  self_play: null
46
  behavioral_cloning: null
47
  env_settings:
48
- env_path: ./training-envs-executables/linux/SnowballTarget/SnowballTarget
49
  env_args: null
50
  base_port: 5005
51
  num_envs: 1
@@ -66,7 +79,7 @@ engine_settings:
66
  no_graphics_monitor: false
67
  environment_parameters: null
68
  checkpoint_settings:
69
- run_id: SnowballTarget3
70
  initialize_from: null
71
  load_model: false
72
  resume: false
 
1
  default_settings: null
2
  behaviors:
3
+ Pyramids:
4
  trainer_type: ppo
5
  hyperparameters:
6
+ batch_size: 384
7
  buffer_size: 2048
8
  learning_rate: 0.0006
9
+ beta: 0.01
10
  epsilon: 0.2
11
  lambd: 0.95
12
  num_epoch: 3
 
14
  learning_rate_schedule: linear
15
  beta_schedule: linear
16
  epsilon_schedule: linear
17
+ checkpoint_interval: 10000
18
  network_settings:
19
  normalize: false
20
+ hidden_units: 512
21
+ num_layers: 2
22
  vis_encode_type: simple
23
  memory: null
24
  goal_conditioning_type: hyper
 
35
  memory: null
36
  goal_conditioning_type: hyper
37
  deterministic: false
38
+ rnd:
39
+ gamma: 0.99
40
+ strength: 0.01
41
+ network_settings:
42
+ normalize: false
43
+ hidden_units: 64
44
+ num_layers: 3
45
+ vis_encode_type: simple
46
+ memory: null
47
+ goal_conditioning_type: hyper
48
+ deterministic: false
49
+ learning_rate: 0.0001
50
+ encoding_size: null
51
  init_path: null
52
+ keep_checkpoints: 5
53
  even_checkpoints: false
54
+ max_steps: 1000000
55
+ time_horizon: 128
56
+ summary_freq: 30000
57
  threaded: false
58
  self_play: null
59
  behavioral_cloning: null
60
  env_settings:
61
+ env_path: ./training-envs-executables/linux/Pyramids/Pyramids
62
  env_args: null
63
  base_port: 5005
64
  num_envs: 1
 
79
  no_graphics_monitor: false
80
  environment_parameters: null
81
  checkpoint_settings:
82
+ run_id: Pyramids Training1
83
  initialize_from: null
84
  load_model: false
85
  resume: false
run_logs/Player-0.log CHANGED
@@ -1,9 +1,11 @@
1
- Mono path[0] = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/Managed'
2
- Mono config path = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
 
 
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
5
- Initialize engine version: 2021.3.14f1 (eee1884e7226)
6
- [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
7
  Forcing GfxDevice: Null
8
  GfxDevice: creating device client; threaded=0; jobified=0
9
  NullGfxDevice:
@@ -31,10 +33,31 @@ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
31
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
32
  FMOD initialized on nosound output
33
  Begin MonoManager ReloadAssembly
34
- - Completed reload, in 0.078 seconds
35
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
36
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
37
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
39
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
40
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
@@ -42,85 +65,90 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
42
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
43
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
44
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
45
- UnloadTime: 0.749284 ms
46
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
47
  requesting resize 84 x 84
48
  Setting up 1 worker threads for Enlighten.
49
- Memory Statistics:
50
  [ALLOC_TEMP_TLS] TLS Allocator
51
  StackAllocators :
52
  [ALLOC_TEMP_MAIN]
53
- Peak usage frame count: [8.0 KB-16.0 KB]: 8966 frames, [16.0 KB-32.0 KB]: 151 frames, [2.0 MB-4.0 MB]: 1 frames
54
  Initial Block Size 4.0 MB
55
  Current Block Size 4.0 MB
56
  Peak Allocated Bytes 2.0 MB
57
  Overflow Count 0
58
- [ALLOC_TEMP_Loading.AsyncRead]
59
- Initial Block Size 64.0 KB
60
- Current Block Size 64.0 KB
61
- Peak Allocated Bytes 198 B
62
- Overflow Count 0
63
- [ALLOC_TEMP_Loading.PreloadManager]
64
- Initial Block Size 256.0 KB
65
- Current Block Size 304.0 KB
66
- Peak Allocated Bytes 228.9 KB
67
- Overflow Count 4
68
  [ALLOC_TEMP_Background Job.Worker 8]
69
  Initial Block Size 32.0 KB
70
  Current Block Size 32.0 KB
71
  Peak Allocated Bytes 0 B
72
  Overflow Count 0
73
- [ALLOC_TEMP_Background Job.Worker 9]
74
  Initial Block Size 32.0 KB
75
  Current Block Size 32.0 KB
76
  Peak Allocated Bytes 0 B
77
  Overflow Count 0
78
- [ALLOC_TEMP_Job.Worker 0]
79
- Initial Block Size 256.0 KB
80
- Current Block Size 256.0 KB
81
- Peak Allocated Bytes 3.6 KB
82
- Overflow Count 0
83
- [ALLOC_TEMP_Background Job.Worker 10]
84
  Initial Block Size 32.0 KB
85
  Current Block Size 32.0 KB
86
  Peak Allocated Bytes 0 B
87
  Overflow Count 0
88
- [ALLOC_TEMP_Background Job.Worker 14]
 
 
 
 
 
89
  Initial Block Size 32.0 KB
90
  Current Block Size 32.0 KB
91
  Peak Allocated Bytes 0 B
92
  Overflow Count 0
93
- [ALLOC_TEMP_Background Job.Worker 6]
94
  Initial Block Size 32.0 KB
95
  Current Block Size 32.0 KB
96
  Peak Allocated Bytes 0 B
97
  Overflow Count 0
98
- [ALLOC_TEMP_Background Job.Worker 12]
99
  Initial Block Size 32.0 KB
100
  Current Block Size 32.0 KB
101
  Peak Allocated Bytes 0 B
102
  Overflow Count 0
103
- [ALLOC_TEMP_EnlightenWorker]
104
  Initial Block Size 64.0 KB
105
  Current Block Size 64.0 KB
106
  Peak Allocated Bytes 0 B
107
  Overflow Count 0
108
- [ALLOC_TEMP_Background Job.Worker 15]
 
 
 
 
 
109
  Initial Block Size 32.0 KB
110
  Current Block Size 32.0 KB
111
  Peak Allocated Bytes 0 B
112
  Overflow Count 0
113
- [ALLOC_TEMP_Background Job.Worker 1]
114
  Initial Block Size 32.0 KB
115
  Current Block Size 32.0 KB
116
  Peak Allocated Bytes 0 B
117
  Overflow Count 0
118
- [ALLOC_TEMP_Background Job.Worker 2]
119
  Initial Block Size 32.0 KB
120
  Current Block Size 32.0 KB
121
  Peak Allocated Bytes 0 B
122
  Overflow Count 0
123
- [ALLOC_TEMP_Background Job.Worker 7]
 
 
 
 
 
 
 
 
 
 
124
  Initial Block Size 32.0 KB
125
  Current Block Size 32.0 KB
126
  Peak Allocated Bytes 0 B
@@ -140,48 +168,64 @@ Memory Statistics:
140
  Current Block Size 32.0 KB
141
  Peak Allocated Bytes 0 B
142
  Overflow Count 0
143
- [ALLOC_TEMP_Background Job.Worker 11]
144
  Initial Block Size 32.0 KB
145
  Current Block Size 32.0 KB
146
  Peak Allocated Bytes 0 B
147
  Overflow Count 0
148
- [ALLOC_TEMP_Background Job.Worker 3]
149
  Initial Block Size 32.0 KB
150
  Current Block Size 32.0 KB
151
  Peak Allocated Bytes 0 B
152
  Overflow Count 0
153
- [ALLOC_TEMP_Background Job.Worker 0]
 
 
 
 
 
154
  Initial Block Size 32.0 KB
155
  Current Block Size 32.0 KB
156
  Peak Allocated Bytes 0 B
157
  Overflow Count 0
158
- [ALLOC_TEMP_Background Job.Worker 4]
159
  Initial Block Size 32.0 KB
160
  Current Block Size 32.0 KB
161
  Peak Allocated Bytes 0 B
162
  Overflow Count 0
163
- [ALLOC_TEMP_BatchDeleteObjects]
164
  Initial Block Size 64.0 KB
165
  Current Block Size 64.0 KB
166
  Peak Allocated Bytes 0 B
167
  Overflow Count 0
 
 
 
 
 
 
 
 
 
 
 
168
  [ALLOC_DEFAULT] Dual Thread Allocator
169
- Peak main deferred allocation count 41
170
  [ALLOC_BUCKET]
171
  Large Block size 4.0 MB
172
  Used Block count 1
173
- Peak Allocated bytes 1.0 MB
174
  [ALLOC_DEFAULT_MAIN]
175
- Peak usage frame count: [4.0 MB-8.0 MB]: 9118 frames
176
  Requested Block Size 16.0 MB
177
- Peak Block count 1
178
- Peak Allocated memory 8.2 MB
179
  Peak Large allocation bytes 0 B
180
  [ALLOC_DEFAULT_THREAD]
181
- Peak usage frame count: [16.0 MB-32.0 MB]: 9118 frames
182
  Requested Block Size 16.0 MB
183
  Peak Block count 1
184
- Peak Allocated memory 17.3 MB
185
  Peak Large allocation bytes 16.0 MB
186
  [ALLOC_TEMP_JOB_1_FRAME]
187
  Initial Block Size 2.0 MB
@@ -208,52 +252,63 @@ Memory Statistics:
208
  [ALLOC_BUCKET]
209
  Large Block size 4.0 MB
210
  Used Block count 1
211
- Peak Allocated bytes 1.0 MB
212
  [ALLOC_GFX_MAIN]
213
- Peak usage frame count: [32.0 KB-64.0 KB]: 8557 frames, [64.0 KB-128.0 KB]: 561 frames
214
  Requested Block Size 16.0 MB
215
  Peak Block count 1
216
- Peak Allocated memory 66.4 KB
217
  Peak Large allocation bytes 0 B
218
  [ALLOC_GFX_THREAD]
219
- Peak usage frame count: [32.0 KB-64.0 KB]: 9118 frames
220
  Requested Block Size 16.0 MB
221
  Peak Block count 1
222
- Peak Allocated memory 39.6 KB
223
  Peak Large allocation bytes 0 B
224
  [ALLOC_CACHEOBJECTS] Dual Thread Allocator
225
- Peak main deferred allocation count 0
226
  [ALLOC_BUCKET]
227
  Large Block size 4.0 MB
228
  Used Block count 1
229
- Peak Allocated bytes 1.0 MB
230
  [ALLOC_CACHEOBJECTS_MAIN]
231
- Peak usage frame count: [0.5 MB-1.0 MB]: 9118 frames
232
  Requested Block Size 4.0 MB
233
  Peak Block count 1
234
- Peak Allocated memory 0.6 MB
235
  Peak Large allocation bytes 0 B
236
  [ALLOC_CACHEOBJECTS_THREAD]
237
- Peak usage frame count: [0.5 MB-1.0 MB]: 9117 frames, [2.0 MB-4.0 MB]: 1 frames
238
  Requested Block Size 4.0 MB
239
- Peak Block count 1
240
- Peak Allocated memory 2.2 MB
241
  Peak Large allocation bytes 0 B
242
  [ALLOC_TYPETREE] Dual Thread Allocator
243
  Peak main deferred allocation count 0
244
  [ALLOC_BUCKET]
245
  Large Block size 4.0 MB
246
  Used Block count 1
247
- Peak Allocated bytes 1.0 MB
248
  [ALLOC_TYPETREE_MAIN]
249
- Peak usage frame count: [0-1.0 KB]: 9118 frames
250
  Requested Block Size 2.0 MB
251
  Peak Block count 1
252
  Peak Allocated memory 1.0 KB
253
  Peak Large allocation bytes 0 B
254
  [ALLOC_TYPETREE_THREAD]
255
- Peak usage frame count: [1.0 KB-2.0 KB]: 9118 frames
256
  Requested Block Size 2.0 MB
257
  Peak Block count 1
258
- Peak Allocated memory 1.7 KB
259
  Peak Large allocation bytes 0 B
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Mono path[0] = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/Managed'
2
+ Mono config path = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/MonoBleedingEdge/etc'
3
+ Found 1 interfaces on host : 0) 172.28.0.12
4
+ Multi-casting "[IP] 172.28.0.12 [Port] 55329 [Flags] 2 [Guid] 1791771930 [EditorId] 764847374 [Version] 1048832 [Id] LinuxPlayer(13,172.28.0.12) [Debug] 0 [PackageName] LinuxPlayer [ProjectName] UnityEnvironment" to [225.0.0.222:54997]...
5
  Preloaded 'lib_burst_generated.so'
6
  Preloaded 'libgrpc_csharp_ext.x64.so'
7
+ Initialize engine version: 2021.3.5f1 (40eb3a945986)
8
+ [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/UnitySubsystems
9
  Forcing GfxDevice: Null
10
  GfxDevice: creating device client; threaded=0; jobified=0
11
  NullGfxDevice:
 
33
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
34
  FMOD initialized on nosound output
35
  Begin MonoManager ReloadAssembly
36
+ - Completed reload, in 0.096 seconds
37
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
38
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
+ WARNING: Shader Unsupported: 'Autodesk Interactive' - All subshaders removed
41
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
42
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
43
+ ERROR: Shader Autodesk Interactive shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
44
+ WARNING: Shader Unsupported: 'Autodesk Interactive' - All subshaders removed
45
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
46
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
47
+ WARNING: Shader Unsupported: 'Legacy Shaders/Diffuse' - All subshaders removed
48
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
49
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
50
+ ERROR: Shader Legacy Shaders/Diffuse shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
51
+ WARNING: Shader Unsupported: 'Legacy Shaders/Diffuse' - All subshaders removed
52
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
53
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
54
+ WARNING: Shader Unsupported: 'ML-Agents/GridPattern' - All subshaders removed
55
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
56
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
57
+ ERROR: Shader ML-Agents/GridPattern shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
58
+ WARNING: Shader Unsupported: 'ML-Agents/GridPattern' - All subshaders removed
59
+ WARNING: Shader Did you use #pragma only_renderers and omit this platform?
60
+ WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
61
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
62
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
63
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
 
65
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
66
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
67
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
68
+ UnloadTime: 1.113629 ms
69
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
70
  requesting resize 84 x 84
71
  Setting up 1 worker threads for Enlighten.
72
+ PlayerConnection::CleanupMemory Statistics:
73
  [ALLOC_TEMP_TLS] TLS Allocator
74
  StackAllocators :
75
  [ALLOC_TEMP_MAIN]
76
+ Peak usage frame count: [16.0 KB-32.0 KB]: 3526 frames, [32.0 KB-64.0 KB]: 1825 frames, [64.0 KB-128.0 KB]: 348 frames, [2.0 MB-4.0 MB]: 1 frames
77
  Initial Block Size 4.0 MB
78
  Current Block Size 4.0 MB
79
  Peak Allocated Bytes 2.0 MB
80
  Overflow Count 0
 
 
 
 
 
 
 
 
 
 
81
  [ALLOC_TEMP_Background Job.Worker 8]
82
  Initial Block Size 32.0 KB
83
  Current Block Size 32.0 KB
84
  Peak Allocated Bytes 0 B
85
  Overflow Count 0
86
+ [ALLOC_TEMP_Background Job.Worker 10]
87
  Initial Block Size 32.0 KB
88
  Current Block Size 32.0 KB
89
  Peak Allocated Bytes 0 B
90
  Overflow Count 0
91
+ [ALLOC_TEMP_Background Job.Worker 6]
 
 
 
 
 
92
  Initial Block Size 32.0 KB
93
  Current Block Size 32.0 KB
94
  Peak Allocated Bytes 0 B
95
  Overflow Count 0
96
+ [ALLOC_TEMP_AUDIO_FMOD mixer thread]
97
+ Initial Block Size 64.0 KB
98
+ Current Block Size 64.0 KB
99
+ Peak Allocated Bytes 0 B
100
+ Overflow Count 0
101
+ [ALLOC_TEMP_Background Job.Worker 1]
102
  Initial Block Size 32.0 KB
103
  Current Block Size 32.0 KB
104
  Peak Allocated Bytes 0 B
105
  Overflow Count 0
106
+ [ALLOC_TEMP_Background Job.Worker 11]
107
  Initial Block Size 32.0 KB
108
  Current Block Size 32.0 KB
109
  Peak Allocated Bytes 0 B
110
  Overflow Count 0
111
+ [ALLOC_TEMP_Background Job.Worker 3]
112
  Initial Block Size 32.0 KB
113
  Current Block Size 32.0 KB
114
  Peak Allocated Bytes 0 B
115
  Overflow Count 0
116
+ [ALLOC_TEMP_BatchDeleteObjects]
117
  Initial Block Size 64.0 KB
118
  Current Block Size 64.0 KB
119
  Peak Allocated Bytes 0 B
120
  Overflow Count 0
121
+ [ALLOC_TEMP_AUDIO_FMOD stream thread]
122
+ Initial Block Size 64.0 KB
123
+ Current Block Size 64.0 KB
124
+ Peak Allocated Bytes 0 B
125
+ Overflow Count 0
126
+ [ALLOC_TEMP_Background Job.Worker 2]
127
  Initial Block Size 32.0 KB
128
  Current Block Size 32.0 KB
129
  Peak Allocated Bytes 0 B
130
  Overflow Count 0
131
+ [ALLOC_TEMP_Background Job.Worker 7]
132
  Initial Block Size 32.0 KB
133
  Current Block Size 32.0 KB
134
  Peak Allocated Bytes 0 B
135
  Overflow Count 0
136
+ [ALLOC_TEMP_Background Job.Worker 15]
137
  Initial Block Size 32.0 KB
138
  Current Block Size 32.0 KB
139
  Peak Allocated Bytes 0 B
140
  Overflow Count 0
141
+ [ALLOC_TEMP_Loading.PreloadManager]
142
+ Initial Block Size 256.0 KB
143
+ Current Block Size 320.0 KB
144
+ Peak Allocated Bytes 251.5 KB
145
+ Overflow Count 4
146
+ [ALLOC_TEMP_Profiler.Dispatcher]
147
+ Initial Block Size 64.0 KB
148
+ Current Block Size 64.0 KB
149
+ Peak Allocated Bytes 240 B
150
+ Overflow Count 0
151
+ [ALLOC_TEMP_Background Job.Worker 12]
152
  Initial Block Size 32.0 KB
153
  Current Block Size 32.0 KB
154
  Peak Allocated Bytes 0 B
 
168
  Current Block Size 32.0 KB
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
+ [ALLOC_TEMP_Background Job.Worker 0]
172
  Initial Block Size 32.0 KB
173
  Current Block Size 32.0 KB
174
  Peak Allocated Bytes 0 B
175
  Overflow Count 0
176
+ [ALLOC_TEMP_Background Job.Worker 4]
177
  Initial Block Size 32.0 KB
178
  Current Block Size 32.0 KB
179
  Peak Allocated Bytes 0 B
180
  Overflow Count 0
181
+ [ALLOC_TEMP_Job.Worker 0]
182
+ Initial Block Size 256.0 KB
183
+ Current Block Size 256.0 KB
184
+ Peak Allocated Bytes 0.8 KB
185
+ Overflow Count 0
186
+ [ALLOC_TEMP_Background Job.Worker 9]
187
  Initial Block Size 32.0 KB
188
  Current Block Size 32.0 KB
189
  Peak Allocated Bytes 0 B
190
  Overflow Count 0
191
+ [ALLOC_TEMP_Background Job.Worker 14]
192
  Initial Block Size 32.0 KB
193
  Current Block Size 32.0 KB
194
  Peak Allocated Bytes 0 B
195
  Overflow Count 0
196
+ [ALLOC_TEMP_EnlightenWorker]
197
  Initial Block Size 64.0 KB
198
  Current Block Size 64.0 KB
199
  Peak Allocated Bytes 0 B
200
  Overflow Count 0
201
+ [ALLOC_TEMP_Loading.AsyncRead]
202
+ Initial Block Size 64.0 KB
203
+ Current Block Size 64.0 KB
204
+ Peak Allocated Bytes 128 B
205
+ Overflow Count 0
206
+ [ALLOC_MEMORYPROFILER]
207
+ Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 5401 frames
208
+ Requested Block Size 1.0 MB
209
+ Peak Block count 2
210
+ Peak Allocated memory 1.6 MB
211
+ Peak Large allocation bytes 0 B
212
  [ALLOC_DEFAULT] Dual Thread Allocator
213
+ Peak main deferred allocation count 11554
214
  [ALLOC_BUCKET]
215
  Large Block size 4.0 MB
216
  Used Block count 1
217
+ Peak Allocated bytes 1.7 MB
218
  [ALLOC_DEFAULT_MAIN]
219
+ Peak usage frame count: [8.0 MB-16.0 MB]: 1 frames, [16.0 MB-32.0 MB]: 5699 frames
220
  Requested Block Size 16.0 MB
221
+ Peak Block count 3
222
+ Peak Allocated memory 30.6 MB
223
  Peak Large allocation bytes 0 B
224
  [ALLOC_DEFAULT_THREAD]
225
+ Peak usage frame count: [16.0 MB-32.0 MB]: 5700 frames
226
  Requested Block Size 16.0 MB
227
  Peak Block count 1
228
+ Peak Allocated memory 23.0 MB
229
  Peak Large allocation bytes 16.0 MB
230
  [ALLOC_TEMP_JOB_1_FRAME]
231
  Initial Block Size 2.0 MB
 
252
  [ALLOC_BUCKET]
253
  Large Block size 4.0 MB
254
  Used Block count 1
255
+ Peak Allocated bytes 1.7 MB
256
  [ALLOC_GFX_MAIN]
257
+ Peak usage frame count: [32.0 KB-64.0 KB]: 5699 frames, [64.0 KB-128.0 KB]: 1 frames
258
  Requested Block Size 16.0 MB
259
  Peak Block count 1
260
+ Peak Allocated memory 67.0 KB
261
  Peak Large allocation bytes 0 B
262
  [ALLOC_GFX_THREAD]
263
+ Peak usage frame count: [64.0 KB-128.0 KB]: 5700 frames
264
  Requested Block Size 16.0 MB
265
  Peak Block count 1
266
+ Peak Allocated memory 67.7 KB
267
  Peak Large allocation bytes 0 B
268
  [ALLOC_CACHEOBJECTS] Dual Thread Allocator
269
+ Peak main deferred allocation count 1
270
  [ALLOC_BUCKET]
271
  Large Block size 4.0 MB
272
  Used Block count 1
273
+ Peak Allocated bytes 1.7 MB
274
  [ALLOC_CACHEOBJECTS_MAIN]
275
+ Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 5401 frames
276
  Requested Block Size 4.0 MB
277
  Peak Block count 1
278
+ Peak Allocated memory 1.4 MB
279
  Peak Large allocation bytes 0 B
280
  [ALLOC_CACHEOBJECTS_THREAD]
281
+ Peak usage frame count: [1.0 MB-2.0 MB]: 5699 frames, [4.0 MB-8.0 MB]: 1 frames
282
  Requested Block Size 4.0 MB
283
+ Peak Block count 2
284
+ Peak Allocated memory 4.7 MB
285
  Peak Large allocation bytes 0 B
286
  [ALLOC_TYPETREE] Dual Thread Allocator
287
  Peak main deferred allocation count 0
288
  [ALLOC_BUCKET]
289
  Large Block size 4.0 MB
290
  Used Block count 1
291
+ Peak Allocated bytes 1.7 MB
292
  [ALLOC_TYPETREE_MAIN]
293
+ Peak usage frame count: [0-1.0 KB]: 5700 frames
294
  Requested Block Size 2.0 MB
295
  Peak Block count 1
296
  Peak Allocated memory 1.0 KB
297
  Peak Large allocation bytes 0 B
298
  [ALLOC_TYPETREE_THREAD]
299
+ Peak usage frame count: [1.0 KB-2.0 KB]: 5700 frames
300
  Requested Block Size 2.0 MB
301
  Peak Block count 1
302
+ Peak Allocated memory 1.9 KB
303
  Peak Large allocation bytes 0 B
304
+ [ALLOC_PROFILER]
305
+ Peak usage frame count: [16.0 KB-32.0 KB]: 5700 frames
306
+ Requested Block Size 16.0 MB
307
+ Peak Block count 1
308
+ Peak Allocated memory 27.4 KB
309
+ Peak Large allocation bytes 0 B
310
+ [ALLOC_PROFILER_BUCKET]
311
+ Large Block size 4.0 MB
312
+ Used Block count 1
313
+ Peak Allocated bytes 396 B
314
+ ##utp:{"type":"MemoryLeaks","version":2,"phase":"Immediate","time":1772325107485,"processId":9033,"allocatedMemory":1970133,"memoryLabels":[{"Default":9033},{"Permanent":1264},{"NewDelete":12801},{"Thread":34460},{"Manager":10603},{"VertexData":12},{"Geometry":280},{"Texture":16},{"Shader":69173},{"Material":24},{"GfxDevice":35248},{"Animation":304},{"Audio":3976},{"Physics":288},{"Serialization":216},{"Input":9176},{"JobScheduler":200},{"Mono":40},{"ScriptingNativeRuntime":216},{"BaseObject":1609212},{"Resource":592},{"Renderer":1936},{"Transform":48},{"File":800},{"WebCam":24},{"Culling":40},{"Terrain":953},{"Wind":24},{"String":3447},{"DynamicArray":30868},{"HashMap":7680},{"Utility":1360},{"PoolAlloc":1160},{"TypeTree":1792},{"ScriptManager":80},{"RuntimeInitializeOnLoadManager":72},{"SpriteAtlas":112},{"GI":3272},{"Unet":16},{"Director":7760},{"WebRequest":720},{"VR":45473},{"SceneManager":424},{"Video":32},{"LazyScriptCache":32},{"NativeArray":384},{"Camera":25},{"Secure":1},{"SerializationCache":624},{"APIUpdating":5872},{"Subsystems":384},{"VirtualTexturing":57552},{"AssetReference":32}]}
run_logs/timers.json CHANGED
@@ -1,214 +1,250 @@
1
  {
2
  "name": "root",
3
  "gauges": {
4
- "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.7095248699188232,
6
- "min": 0.7095248699188232,
7
- "max": 2.8473734855651855,
8
- "count": 33
9
  },
10
- "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 6805.7626953125,
12
- "min": 6805.7626953125,
13
- "max": 29065.98828125,
14
- "count": 33
15
  },
16
- "SnowballTarget.Step.mean": {
17
- "value": 329992.0,
18
- "min": 9952.0,
19
- "max": 329992.0,
20
- "count": 33
21
  },
22
- "SnowballTarget.Step.sum": {
23
- "value": 329992.0,
24
- "min": 9952.0,
25
- "max": 329992.0,
26
- "count": 33
27
  },
28
- "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 13.541885375976562,
30
- "min": 0.3765427768230438,
31
- "max": 13.750577926635742,
32
- "count": 33
33
  },
34
- "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2762.544677734375,
36
- "min": 73.04930114746094,
37
- "max": 2814.87060546875,
38
- "count": 33
39
  },
40
- "SnowballTarget.Losses.PolicyLoss.mean": {
41
- "value": 0.07133594242052413,
42
- "min": 0.062328982485946746,
43
- "max": 0.07871535035128724,
44
- "count": 33
45
  },
46
- "SnowballTarget.Losses.PolicyLoss.sum": {
47
- "value": 0.35667971210262067,
48
- "min": 0.24931592994378698,
49
- "max": 0.3935767517564362,
50
- "count": 33
51
  },
52
- "SnowballTarget.Losses.ValueLoss.mean": {
53
- "value": 0.17541804827895818,
54
- "min": 0.11917218662611229,
55
- "max": 0.26781621604573497,
56
- "count": 33
57
  },
58
- "SnowballTarget.Losses.ValueLoss.sum": {
59
- "value": 0.8770902413947909,
60
- "min": 0.47668874650444915,
61
- "max": 1.3390810802286748,
62
- "count": 33
63
  },
64
- "SnowballTarget.Policy.LearningRate.mean": {
65
- "value": 0.00011173208137799998,
66
- "min": 0.00011173208137799998,
67
- "max": 0.0005918820013529999,
68
- "count": 33
69
  },
70
- "SnowballTarget.Policy.LearningRate.sum": {
71
- "value": 0.0005586604068899999,
72
- "min": 0.0005586604068899999,
73
- "max": 0.00288516001914,
74
- "count": 33
75
  },
76
- "SnowballTarget.Policy.Epsilon.mean": {
77
- "value": 0.118622,
78
- "min": 0.118622,
79
- "max": 0.19864700000000002,
80
- "count": 33
81
  },
82
- "SnowballTarget.Policy.Epsilon.sum": {
83
- "value": 0.59311,
84
- "min": 0.49538800000000005,
85
- "max": 0.98086,
86
- "count": 33
87
  },
88
- "SnowballTarget.Policy.Beta.mean": {
89
- "value": 0.0009392378000000002,
90
- "min": 0.0009392378000000002,
91
- "max": 0.004932485299999999,
92
- "count": 33
93
  },
94
- "SnowballTarget.Policy.Beta.sum": {
95
- "value": 0.004696189000000001,
96
- "min": 0.004696189000000001,
97
- "max": 0.024044914,
98
- "count": 33
99
  },
100
- "SnowballTarget.Environment.EpisodeLength.mean": {
101
- "value": 199.0,
102
- "min": 199.0,
103
- "max": 199.0,
104
- "count": 33
105
  },
106
- "SnowballTarget.Environment.EpisodeLength.sum": {
107
- "value": 10945.0,
108
- "min": 8756.0,
109
- "max": 10945.0,
110
- "count": 33
111
  },
112
- "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 26.796296296296298,
114
- "min": 3.477272727272727,
115
- "max": 26.963636363636365,
116
- "count": 33
117
  },
118
- "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1447.0,
120
- "min": 153.0,
121
- "max": 1483.0,
122
- "count": 33
123
  },
124
- "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 26.796296296296298,
126
- "min": 3.477272727272727,
127
- "max": 26.963636363636365,
128
- "count": 33
129
  },
130
- "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1447.0,
132
- "min": 153.0,
133
- "max": 1483.0,
134
- "count": 33
135
  },
136
- "SnowballTarget.IsTraining.mean": {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
137
  "value": 1.0,
138
  "min": 1.0,
139
  "max": 1.0,
140
- "count": 33
141
  },
142
- "SnowballTarget.IsTraining.sum": {
143
  "value": 1.0,
144
  "min": 1.0,
145
  "max": 1.0,
146
- "count": 33
147
  }
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1772314848",
152
  "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
153
- "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget3 --no-graphics",
154
  "mlagents_version": "1.2.0.dev0",
155
  "mlagents_envs_version": "1.2.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.8.0+cu128",
158
  "numpy_version": "1.23.5",
159
- "end_time_seconds": "1772315593"
160
  },
161
- "total": 744.4466958350004,
162
  "count": 1,
163
- "self": 0.42514927800039004,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.023836146000121516,
167
  "count": 1,
168
- "self": 0.023836146000121516
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 743.9977104109998,
172
  "count": 1,
173
- "self": 0.5701817649573968,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 1.9720252289998825,
177
  "count": 1,
178
- "self": 1.9720252289998825
179
  },
180
  "TrainerController.advance": {
181
- "total": 741.4484766610431,
182
- "count": 30392,
183
- "self": 0.6209287119768305,
184
  "children": {
185
  "env_step": {
186
- "total": 532.7946861069622,
187
- "count": 30392,
188
- "self": 415.44035444101155,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 117.00914793202355,
192
- "count": 30392,
193
- "self": 2.023098731980099,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 114.98604920004345,
197
- "count": 30392,
198
- "self": 114.98604920004345
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.3451837339271151,
204
- "count": 30392,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 741.0003784210057,
209
- "count": 30392,
210
  "is_parallel": true,
211
- "self": 375.93809538301775,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +253,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.0019181510001544666,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.0006041430001459958,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0013140080000084708,
227
- "count": 10,
228
  "is_parallel": true,
229
- "self": 0.0013140080000084708
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.035656288000154746,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0005850039997312706,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.00043290500025250367,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.00043290500025250367
244
  },
245
  "communicator.exchange": {
246
- "total": 0.03275899800019033,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.03275899800019033
250
  },
251
  "steps_from_proto": {
252
- "total": 0.00187938099998064,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.00036688099908133154,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.0015125000008993084,
259
- "count": 10,
260
  "is_parallel": true,
261
- "self": 0.0015125000008993084
262
  }
263
  }
264
  }
@@ -267,34 +303,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 365.062283037988,
271
- "count": 30391,
272
  "is_parallel": true,
273
- "self": 16.953038460037988,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 8.97932959200216,
277
- "count": 30391,
278
  "is_parallel": true,
279
- "self": 8.97932959200216
280
  },
281
  "communicator.exchange": {
282
- "total": 278.1169417809606,
283
- "count": 30391,
284
  "is_parallel": true,
285
- "self": 278.1169417809606
286
  },
287
  "steps_from_proto": {
288
- "total": 61.01297320498725,
289
- "count": 30391,
290
  "is_parallel": true,
291
- "self": 11.085336442878543,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 49.927636762108705,
295
- "count": 303910,
296
  "is_parallel": true,
297
- "self": 49.927636762108705
298
  }
299
  }
300
  }
@@ -307,31 +343,31 @@
307
  }
308
  },
309
  "trainer_advance": {
310
- "total": 208.032861842104,
311
- "count": 30392,
312
- "self": 0.679465280126351,
313
  "children": {
314
  "process_trajectory": {
315
- "total": 44.70426057097711,
316
- "count": 30392,
317
- "self": 44.064410302977194,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
- "total": 0.6398502679999183,
321
- "count": 6,
322
- "self": 0.6398502679999183
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
- "total": 162.64913599100055,
328
- "count": 152,
329
- "self": 64.29756426800122,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
- "total": 98.35157172299932,
333
- "count": 7698,
334
- "self": 98.35157172299932
335
  }
336
  }
337
  }
@@ -340,19 +376,19 @@
340
  }
341
  },
342
  "trainer_threads": {
343
- "total": 1.4320003174361773e-06,
344
  "count": 1,
345
- "self": 1.4320003174361773e-06
346
  },
347
  "TrainerController._save_models": {
348
- "total": 0.007025323999187094,
349
  "count": 1,
350
- "self": 2.329199924133718e-05,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
- "total": 0.007002031999945757,
354
  "count": 1,
355
- "self": 0.007002031999945757
356
  }
357
  }
358
  }
 
1
  {
2
  "name": "root",
3
  "gauges": {
4
+ "Pyramids.Policy.Entropy.mean": {
5
+ "value": 0.48718133568763733,
6
+ "min": 0.44092342257499695,
7
+ "max": 1.3449426889419556,
8
+ "count": 10
9
  },
10
+ "Pyramids.Policy.Entropy.sum": {
11
+ "value": 14771.337890625,
12
+ "min": 13051.3330078125,
13
+ "max": 40800.1796875,
14
+ "count": 10
15
  },
16
+ "Pyramids.Step.mean": {
17
+ "value": 299997.0,
18
+ "min": 29900.0,
19
+ "max": 299997.0,
20
+ "count": 10
21
  },
22
+ "Pyramids.Step.sum": {
23
+ "value": 299997.0,
24
+ "min": 29900.0,
25
+ "max": 299997.0,
26
+ "count": 10
27
  },
28
+ "Pyramids.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": -0.011249717324972153,
30
+ "min": -0.09220730513334274,
31
+ "max": 0.11397071927785873,
32
+ "count": 10
33
  },
34
+ "Pyramids.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": -2.756180763244629,
36
+ "min": -22.221960067749023,
37
+ "max": 27.01106071472168,
38
+ "count": 10
39
  },
40
+ "Pyramids.Policy.RndValueEstimate.mean": {
41
+ "value": 0.08842242509126663,
42
+ "min": 0.08842242509126663,
43
+ "max": 0.7224162220954895,
44
+ "count": 10
45
  },
46
+ "Pyramids.Policy.RndValueEstimate.sum": {
47
+ "value": 21.663494110107422,
48
+ "min": 21.663494110107422,
49
+ "max": 174.82472229003906,
50
+ "count": 10
51
  },
52
+ "Pyramids.Losses.PolicyLoss.mean": {
53
+ "value": 0.03467328250997178,
54
+ "min": 0.03386264300712175,
55
+ "max": 0.04452162305889133,
56
+ "count": 10
57
  },
58
+ "Pyramids.Losses.PolicyLoss.sum": {
59
+ "value": 0.48542595513960496,
60
+ "min": 0.35617298447113066,
61
+ "max": 0.5592456866828017,
62
+ "count": 10
63
  },
64
+ "Pyramids.Losses.ValueLoss.mean": {
65
+ "value": 0.0054510476872175265,
66
+ "min": 0.0009858210122475693,
67
+ "max": 0.019210243828085907,
68
+ "count": 10
69
  },
70
+ "Pyramids.Losses.ValueLoss.sum": {
71
+ "value": 0.07631466762104537,
72
+ "min": 0.012923946980218818,
73
+ "max": 0.15368195062468726,
74
+ "count": 10
75
  },
76
+ "Pyramids.Policy.LearningRate.mean": {
77
+ "value": 0.00042893711422477135,
78
+ "min": 0.00042893711422477135,
79
+ "max": 0.0005895222767462874,
80
+ "count": 10
81
  },
82
+ "Pyramids.Policy.LearningRate.sum": {
83
+ "value": 0.006005119599146799,
84
+ "min": 0.0047161782139702995,
85
+ "max": 0.007017206630465599,
86
+ "count": 10
87
  },
88
+ "Pyramids.Policy.Epsilon.mean": {
89
+ "value": 0.17148951428571427,
90
+ "min": 0.17148951428571427,
91
+ "max": 0.1982537125,
92
+ "count": 10
93
  },
94
+ "Pyramids.Policy.Epsilon.sum": {
95
+ "value": 2.4008532,
96
+ "min": 1.5860297,
97
+ "max": 2.5695344,
98
+ "count": 10
99
  },
100
+ "Pyramids.Policy.Beta.mean": {
101
+ "value": 0.007151802477142857,
102
+ "min": 0.007151802477142857,
103
+ "max": 0.009825545878750001,
104
+ "count": 10
105
  },
106
+ "Pyramids.Policy.Beta.sum": {
107
+ "value": 0.10012523468,
108
+ "min": 0.07860436703000001,
109
+ "max": 0.11697648656,
110
+ "count": 10
111
  },
112
+ "Pyramids.Losses.RNDLoss.mean": {
113
+ "value": 0.09631339460611343,
114
+ "min": 0.09631339460611343,
115
+ "max": 0.9355165958404541,
116
+ "count": 10
117
  },
118
+ "Pyramids.Losses.RNDLoss.sum": {
119
+ "value": 1.3483874797821045,
120
+ "min": 1.3483874797821045,
121
+ "max": 7.484132766723633,
122
+ "count": 10
123
  },
124
+ "Pyramids.Environment.EpisodeLength.mean": {
125
+ "value": 906.0571428571428,
126
+ "min": 895.5555555555555,
127
+ "max": 999.0,
128
+ "count": 10
129
  },
130
+ "Pyramids.Environment.EpisodeLength.sum": {
131
+ "value": 31712.0,
132
+ "min": 17466.0,
133
+ "max": 32713.0,
134
+ "count": 10
135
  },
136
+ "Pyramids.Environment.CumulativeReward.mean": {
137
+ "value": -0.3351657609854426,
138
+ "min": -0.9998600512742997,
139
+ "max": -0.30345930214281436,
140
+ "count": 10
141
+ },
142
+ "Pyramids.Environment.CumulativeReward.sum": {
143
+ "value": -11.73080163449049,
144
+ "min": -30.995201587677002,
145
+ "max": -8.193401157855988,
146
+ "count": 10
147
+ },
148
+ "Pyramids.Policy.ExtrinsicReward.mean": {
149
+ "value": -0.3351657609854426,
150
+ "min": -0.9998600512742997,
151
+ "max": -0.30345930214281436,
152
+ "count": 10
153
+ },
154
+ "Pyramids.Policy.ExtrinsicReward.sum": {
155
+ "value": -11.73080163449049,
156
+ "min": -30.995201587677002,
157
+ "max": -8.193401157855988,
158
+ "count": 10
159
+ },
160
+ "Pyramids.Policy.RndReward.mean": {
161
+ "value": 0.9231859935181481,
162
+ "min": 0.9231859935181481,
163
+ "max": 11.578956580824322,
164
+ "count": 10
165
+ },
166
+ "Pyramids.Policy.RndReward.sum": {
167
+ "value": 32.311509773135185,
168
+ "min": 28.336788680404425,
169
+ "max": 247.3644058406353,
170
+ "count": 10
171
+ },
172
+ "Pyramids.IsTraining.mean": {
173
  "value": 1.0,
174
  "min": 1.0,
175
  "max": 1.0,
176
+ "count": 10
177
  },
178
+ "Pyramids.IsTraining.sum": {
179
  "value": 1.0,
180
  "min": 1.0,
181
  "max": 1.0,
182
+ "count": 10
183
  }
184
  },
185
  "metadata": {
186
  "timer_format_version": "0.1.0",
187
+ "start_time_seconds": "1772324510",
188
  "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
189
+ "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training1 --no-graphics",
190
  "mlagents_version": "1.2.0.dev0",
191
  "mlagents_envs_version": "1.2.0.dev0",
192
  "communication_protocol_version": "1.5.0",
193
  "pytorch_version": "2.8.0+cu128",
194
  "numpy_version": "1.23.5",
195
+ "end_time_seconds": "1772325107"
196
  },
197
+ "total": 597.4701588340001,
198
  "count": 1,
199
+ "self": 0.6499863020001158,
200
  "children": {
201
  "run_training.setup": {
202
+ "total": 0.024338695000096777,
203
  "count": 1,
204
+ "self": 0.024338695000096777
205
  },
206
  "TrainerController.start_learning": {
207
+ "total": 596.7958338369999,
208
  "count": 1,
209
+ "self": 0.4535029420094361,
210
  "children": {
211
  "TrainerController._reset_env": {
212
+ "total": 2.920888082999909,
213
  "count": 1,
214
+ "self": 2.920888082999909
215
  },
216
  "TrainerController.advance": {
217
+ "total": 593.0715385539904,
218
+ "count": 19124,
219
+ "self": 0.46691838498054494,
220
  "children": {
221
  "env_step": {
222
+ "total": 451.1742693780134,
223
+ "count": 19124,
224
+ "self": 399.02851400802,
225
  "children": {
226
  "SubprocessEnvManager._take_step": {
227
+ "total": 51.853985225991664,
228
+ "count": 19124,
229
+ "self": 1.4880065949726031,
230
  "children": {
231
  "TorchPolicy.evaluate": {
232
+ "total": 50.36597863101906,
233
+ "count": 19000,
234
+ "self": 50.36597863101906
235
  }
236
  }
237
  },
238
  "workers": {
239
+ "total": 0.29177014400170265,
240
+ "count": 19124,
241
  "self": 0.0,
242
  "children": {
243
  "worker_root": {
244
+ "total": 593.6382533730057,
245
+ "count": 19124,
246
  "is_parallel": true,
247
+ "self": 232.91384954699197,
248
  "children": {
249
  "run_training.setup": {
250
  "total": 0.0,
 
253
  "self": 0.0,
254
  "children": {
255
  "steps_from_proto": {
256
+ "total": 0.002400254999884055,
257
  "count": 1,
258
  "is_parallel": true,
259
+ "self": 0.0006844129998171411,
260
  "children": {
261
  "_process_rank_one_or_two_observation": {
262
+ "total": 0.001715842000066914,
263
+ "count": 8,
264
  "is_parallel": true,
265
+ "self": 0.001715842000066914
266
  }
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 0.11202883600003588,
271
  "count": 1,
272
  "is_parallel": true,
273
+ "self": 0.0005162520001249504,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 0.0004278869998870505,
277
  "count": 1,
278
  "is_parallel": true,
279
+ "self": 0.0004278869998870505
280
  },
281
  "communicator.exchange": {
282
+ "total": 0.10750669499998367,
283
  "count": 1,
284
  "is_parallel": true,
285
+ "self": 0.10750669499998367
286
  },
287
  "steps_from_proto": {
288
+ "total": 0.0035780020000402146,
289
  "count": 1,
290
  "is_parallel": true,
291
+ "self": 0.0023702869998487586,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 0.001207715000191456,
295
+ "count": 8,
296
  "is_parallel": true,
297
+ "self": 0.001207715000191456
298
  }
299
  }
300
  }
 
303
  }
304
  },
305
  "UnityEnvironment.step": {
306
+ "total": 360.7244038260137,
307
+ "count": 19123,
308
  "is_parallel": true,
309
+ "self": 10.391553950986918,
310
  "children": {
311
  "UnityEnvironment._generate_step_input": {
312
+ "total": 6.915995956998813,
313
+ "count": 19123,
314
  "is_parallel": true,
315
+ "self": 6.915995956998813
316
  },
317
  "communicator.exchange": {
318
+ "total": 309.65846909999937,
319
+ "count": 19123,
320
  "is_parallel": true,
321
+ "self": 309.65846909999937
322
  },
323
  "steps_from_proto": {
324
+ "total": 33.75838481802862,
325
+ "count": 19123,
326
  "is_parallel": true,
327
+ "self": 7.371057496071444,
328
  "children": {
329
  "_process_rank_one_or_two_observation": {
330
+ "total": 26.387327321957173,
331
+ "count": 152984,
332
  "is_parallel": true,
333
+ "self": 26.387327321957173
334
  }
335
  }
336
  }
 
343
  }
344
  },
345
  "trainer_advance": {
346
+ "total": 141.4303507909965,
347
+ "count": 19124,
348
+ "self": 0.7637092439968001,
349
  "children": {
350
  "process_trajectory": {
351
+ "total": 39.89908687100046,
352
+ "count": 19124,
353
+ "self": 37.325428843000054,
354
  "children": {
355
  "RLTrainer._checkpoint": {
356
+ "total": 2.573658028000409,
357
+ "count": 30,
358
+ "self": 2.573658028000409
359
  }
360
  }
361
  },
362
  "_update_policy": {
363
+ "total": 100.76755467599924,
364
+ "count": 128,
365
+ "self": 72.64264833798234,
366
  "children": {
367
  "TorchPPOOptimizer.update": {
368
+ "total": 28.1249063380169,
369
+ "count": 2142,
370
+ "self": 28.1249063380169
371
  }
372
  }
373
  }
 
376
  }
377
  },
378
  "trainer_threads": {
379
+ "total": 1.6220001270994544e-06,
380
  "count": 1,
381
+ "self": 1.6220001270994544e-06
382
  },
383
  "TrainerController._save_models": {
384
+ "total": 0.34990263600002436,
385
  "count": 1,
386
+ "self": 0.007713586000136274,
387
  "children": {
388
  "RLTrainer._checkpoint": {
389
+ "total": 0.3421890499998881,
390
  "count": 1,
391
+ "self": 0.3421890499998881
392
  }
393
  }
394
  }
run_logs/training_status.json CHANGED
@@ -1,61 +1,61 @@
1
  {
2
- "SnowballTarget": {
3
  "checkpoints": [
4
  {
5
- "steps": 49936,
6
- "file_path": "results/SnowballTarget3/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 17.545454545454547,
8
- "creation_time": 1772314963.4103618,
9
  "auxillary_file_paths": [
10
- "results/SnowballTarget3/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
12
  },
13
  {
14
- "steps": 99960,
15
- "file_path": "results/SnowballTarget3/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 23.0,
17
- "creation_time": 1772315076.5508173,
18
  "auxillary_file_paths": [
19
- "results/SnowballTarget3/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
21
  },
22
  {
23
- "steps": 149984,
24
- "file_path": "results/SnowballTarget3/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 26.454545454545453,
26
- "creation_time": 1772315186.8693945,
27
  "auxillary_file_paths": [
28
- "results/SnowballTarget3/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
30
  },
31
  {
32
- "steps": 199984,
33
- "file_path": "results/SnowballTarget3/SnowballTarget/SnowballTarget-199984.onnx",
34
- "reward": 25.727272727272727,
35
- "creation_time": 1772315296.4198992,
36
  "auxillary_file_paths": [
37
- "results/SnowballTarget3/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
39
  },
40
  {
41
- "steps": 249944,
42
- "file_path": "results/SnowballTarget3/SnowballTarget/SnowballTarget-249944.onnx",
43
- "reward": 26.454545454545453,
44
- "creation_time": 1772315407.5301805,
45
  "auxillary_file_paths": [
46
- "results/SnowballTarget3/SnowballTarget/SnowballTarget-249944.pt"
47
- ]
48
- },
49
- {
50
- "steps": 299968,
51
- "file_path": "results/SnowballTarget3/SnowballTarget/SnowballTarget-299968.onnx",
52
- "reward": 26.545454545454547,
53
- "creation_time": 1772315518.2998154,
54
- "auxillary_file_paths": [
55
- "results/SnowballTarget3/SnowballTarget/SnowballTarget-299968.pt"
56
  ]
57
  }
58
- ]
 
 
 
 
 
 
 
 
 
59
  },
60
  "metadata": {
61
  "stats_format_version": "0.3.0",
 
1
  {
2
+ "Pyramids": {
3
  "checkpoints": [
4
  {
5
+ "steps": 269914,
6
+ "file_path": "results/Pyramids Training1/Pyramids/Pyramids-269914.onnx",
7
+ "reward": null,
8
+ "creation_time": 1772325041.706568,
9
  "auxillary_file_paths": [
10
+ "results/Pyramids Training1/Pyramids/Pyramids-269914.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 279912,
15
+ "file_path": "results/Pyramids Training1/Pyramids/Pyramids-279912.onnx",
16
+ "reward": null,
17
+ "creation_time": 1772325061.52339,
18
  "auxillary_file_paths": [
19
+ "results/Pyramids Training1/Pyramids/Pyramids-279912.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 289923,
24
+ "file_path": "results/Pyramids Training1/Pyramids/Pyramids-289923.onnx",
25
+ "reward": 0.6803999791542689,
26
+ "creation_time": 1772325081.5630383,
27
  "auxillary_file_paths": [
28
+ "results/Pyramids Training1/Pyramids/Pyramids-289923.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 299997,
33
+ "file_path": "results/Pyramids Training1/Pyramids/Pyramids-299997.onnx",
34
+ "reward": null,
35
+ "creation_time": 1772325100.941728,
36
  "auxillary_file_paths": [
37
+ "results/Pyramids Training1/Pyramids/Pyramids-299997.pt"
38
  ]
39
  },
40
  {
41
+ "steps": 303333,
42
+ "file_path": "results/Pyramids Training1/Pyramids/Pyramids-303333.onnx",
43
+ "reward": null,
44
+ "creation_time": 1772325106.9032443,
45
  "auxillary_file_paths": [
46
+ "results/Pyramids Training1/Pyramids/Pyramids-303333.pt"
 
 
 
 
 
 
 
 
 
47
  ]
48
  }
49
+ ],
50
+ "final_checkpoint": {
51
+ "steps": 303333,
52
+ "file_path": "results/Pyramids Training1/Pyramids.onnx",
53
+ "reward": null,
54
+ "creation_time": 1772325106.9032443,
55
+ "auxillary_file_paths": [
56
+ "results/Pyramids Training1/Pyramids/Pyramids-303333.pt"
57
+ ]
58
+ }
59
  },
60
  "metadata": {
61
  "stats_format_version": "0.3.0",