sgoodfriend commited on
Commit
4f3effa
·
1 Parent(s): 9dfa257

1M step training

Browse files
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "checkpoint_interval": 50000, "max_steps": 200000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "checkpoint_interval": 50000, "max_steps": 1000000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget2", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -36,7 +36,7 @@ behaviors:
36
  init_path: null
37
  keep_checkpoints: 10
38
  checkpoint_interval: 50000
39
- max_steps: 200000
40
  time_horizon: 64
41
  summary_freq: 10000
42
  threaded: true
@@ -62,7 +62,7 @@ engine_settings:
62
  no_graphics: true
63
  environment_parameters: null
64
  checkpoint_settings:
65
- run_id: SnowballTarget1
66
  initialize_from: null
67
  load_model: false
68
  resume: false
 
36
  init_path: null
37
  keep_checkpoints: 10
38
  checkpoint_interval: 50000
39
+ max_steps: 1000000
40
  time_horizon: 64
41
  summary_freq: 10000
42
  threaded: true
 
62
  no_graphics: true
63
  environment_parameters: null
64
  checkpoint_settings:
65
+ run_id: SnowballTarget2
66
  initialize_from: null
67
  load_model: false
68
  resume: false
run_logs/Player-0.log DELETED
@@ -1,292 +0,0 @@
1
- Mono path[0] = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/Managed'
2
- Mono config path = '/content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/MonoBleedingEdge/etc'
3
- Preloaded 'lib_burst_generated.so'
4
- Preloaded 'libgrpc_csharp_ext.x64.so'
5
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face
6
- PlayerPrefs - Creating folder: /root/.config/unity3d/Hugging Face/SnowballTarget
7
- Unable to load player prefs
8
- Initialize engine version: 2021.3.14f1 (eee1884e7226)
9
- [Subsystems] Discovering subsystems at path /content/ml-agents/training-envs-executables/linux/SnowballTarget/SnowballTarget_Data/UnitySubsystems
10
- Forcing GfxDevice: Null
11
- GfxDevice: creating device client; threaded=0; jobified=0
12
- NullGfxDevice:
13
- Version: NULL 1.0 [1.0]
14
- Renderer: Null Device
15
- Vendor: Unity Technologies
16
- ALSA lib confmisc.c:767:(parse_card) cannot find card '0'
17
- ALSA lib conf.c:4528:(_snd_config_evaluate) function snd_func_card_driver returned error: No such file or directory
18
- ALSA lib confmisc.c:392:(snd_func_concat) error evaluating strings
19
- ALSA lib conf.c:4528:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
20
- ALSA lib confmisc.c:1246:(snd_func_refer) error evaluating name
21
- ALSA lib conf.c:4528:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
22
- ALSA lib conf.c:5007:(snd_config_expand) Evaluate error: No such file or directory
23
- ALSA lib pcm.c:2495:(snd_pcm_open_noupdate) Unknown PCM default
24
- FMOD failed to initialize the output device.: "Error initializing output device. " (60)
25
- Forced to initialize FMOD to to the device driver's system output rate 48000, this may impact performance and/or give inconsistent experiences compared to selected sample rate 48000
26
- ALSA lib confmisc.c:767:(parse_card) cannot find card '0'
27
- ALSA lib conf.c:4528:(_snd_config_evaluate) function snd_func_card_driver returned error: No such file or directory
28
- ALSA lib confmisc.c:392:(snd_func_concat) error evaluating strings
29
- ALSA lib conf.c:4528:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
30
- ALSA lib confmisc.c:1246:(snd_func_refer) error evaluating name
31
- ALSA lib conf.c:4528:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
32
- ALSA lib conf.c:5007:(snd_config_expand) Evaluate error: No such file or directory
33
- ALSA lib pcm.c:2495:(snd_pcm_open_noupdate) Unknown PCM default
34
- FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
- FMOD initialized on nosound output
36
- Begin MonoManager ReloadAssembly
37
- - Completed reload, in 0.074 seconds
38
- ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
- ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
- ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
- WARNING: Shader Unsupported: 'Standard' - All subshaders removed
42
- WARNING: Shader Did you use #pragma only_renderers and omit this platform?
43
- WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
44
- ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
45
- WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
- WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
- WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
- UnloadTime: 0.694478 ms
49
- ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
- requesting resize 84 x 84
51
- Setting up 4 worker threads for Enlighten.
52
- Memory Statistics:
53
- [ALLOC_TEMP_TLS] TLS Allocator
54
- StackAllocators :
55
- [ALLOC_TEMP_MAIN]
56
- Peak usage frame count: [8.0 KB-16.0 KB]: 5370 frames, [16.0 KB-32.0 KB]: 91 frames, [2.0 MB-4.0 MB]: 1 frames
57
- Initial Block Size 4.0 MB
58
- Current Block Size 4.0 MB
59
- Peak Allocated Bytes 2.0 MB
60
- Overflow Count 0
61
- [ALLOC_TEMP_Loading.AsyncRead]
62
- Initial Block Size 64.0 KB
63
- Current Block Size 64.0 KB
64
- Peak Allocated Bytes 198 B
65
- Overflow Count 0
66
- [ALLOC_TEMP_Loading.PreloadManager]
67
- Initial Block Size 256.0 KB
68
- Current Block Size 304.0 KB
69
- Peak Allocated Bytes 228.9 KB
70
- Overflow Count 4
71
- [ALLOC_TEMP_Background Job.Worker 8]
72
- Initial Block Size 32.0 KB
73
- Current Block Size 32.0 KB
74
- Peak Allocated Bytes 0 B
75
- Overflow Count 0
76
- [ALLOC_TEMP_Job.Worker 6]
77
- Initial Block Size 256.0 KB
78
- Current Block Size 256.0 KB
79
- Peak Allocated Bytes 3.6 KB
80
- Overflow Count 0
81
- [ALLOC_TEMP_Job.Worker 0]
82
- Initial Block Size 256.0 KB
83
- Current Block Size 256.0 KB
84
- Peak Allocated Bytes 3.6 KB
85
- Overflow Count 0
86
- [ALLOC_TEMP_Background Job.Worker 10]
87
- Initial Block Size 32.0 KB
88
- Current Block Size 32.0 KB
89
- Peak Allocated Bytes 0 B
90
- Overflow Count 0
91
- [ALLOC_TEMP_Background Job.Worker 9]
92
- Initial Block Size 32.0 KB
93
- Current Block Size 32.0 KB
94
- Peak Allocated Bytes 0 B
95
- Overflow Count 0
96
- [ALLOC_TEMP_Job.Worker 5]
97
- Initial Block Size 256.0 KB
98
- Current Block Size 256.0 KB
99
- Peak Allocated Bytes 3.6 KB
100
- Overflow Count 0
101
- [ALLOC_TEMP_Background Job.Worker 14]
102
- Initial Block Size 32.0 KB
103
- Current Block Size 32.0 KB
104
- Peak Allocated Bytes 0 B
105
- Overflow Count 0
106
- [ALLOC_TEMP_Background Job.Worker 6]
107
- Initial Block Size 32.0 KB
108
- Current Block Size 32.0 KB
109
- Peak Allocated Bytes 0 B
110
- Overflow Count 0
111
- [ALLOC_TEMP_Job.Worker 4]
112
- Initial Block Size 256.0 KB
113
- Current Block Size 256.0 KB
114
- Peak Allocated Bytes 3.6 KB
115
- Overflow Count 0
116
- [ALLOC_TEMP_Job.Worker 3]
117
- Initial Block Size 256.0 KB
118
- Current Block Size 256.0 KB
119
- Peak Allocated Bytes 3.6 KB
120
- Overflow Count 0
121
- [ALLOC_TEMP_Background Job.Worker 12]
122
- Initial Block Size 32.0 KB
123
- Current Block Size 32.0 KB
124
- Peak Allocated Bytes 0 B
125
- Overflow Count 0
126
- [ALLOC_TEMP_EnlightenWorker] x 4
127
- Initial Block Size 64.0 KB
128
- Current Block Size 64.0 KB
129
- Peak Allocated Bytes 0 B
130
- Overflow Count 0
131
- [ALLOC_TEMP_Background Job.Worker 15]
132
- Initial Block Size 32.0 KB
133
- Current Block Size 32.0 KB
134
- Peak Allocated Bytes 0 B
135
- Overflow Count 0
136
- [ALLOC_TEMP_Background Job.Worker 1]
137
- Initial Block Size 32.0 KB
138
- Current Block Size 32.0 KB
139
- Peak Allocated Bytes 0 B
140
- Overflow Count 0
141
- [ALLOC_TEMP_Background Job.Worker 2]
142
- Initial Block Size 32.0 KB
143
- Current Block Size 32.0 KB
144
- Peak Allocated Bytes 0 B
145
- Overflow Count 0
146
- [ALLOC_TEMP_Background Job.Worker 7]
147
- Initial Block Size 32.0 KB
148
- Current Block Size 32.0 KB
149
- Peak Allocated Bytes 0 B
150
- Overflow Count 0
151
- [ALLOC_TEMP_AssetGarbageCollectorHelper] x 7
152
- Initial Block Size 64.0 KB
153
- Current Block Size 64.0 KB
154
- Peak Allocated Bytes 0 B
155
- Overflow Count 0
156
- [ALLOC_TEMP_Background Job.Worker 5]
157
- Initial Block Size 32.0 KB
158
- Current Block Size 32.0 KB
159
- Peak Allocated Bytes 0 B
160
- Overflow Count 0
161
- [ALLOC_TEMP_Background Job.Worker 13]
162
- Initial Block Size 32.0 KB
163
- Current Block Size 32.0 KB
164
- Peak Allocated Bytes 0 B
165
- Overflow Count 0
166
- [ALLOC_TEMP_Job.Worker 1]
167
- Initial Block Size 256.0 KB
168
- Current Block Size 256.0 KB
169
- Peak Allocated Bytes 3.6 KB
170
- Overflow Count 0
171
- [ALLOC_TEMP_Job.Worker 2]
172
- Initial Block Size 256.0 KB
173
- Current Block Size 256.0 KB
174
- Peak Allocated Bytes 3.6 KB
175
- Overflow Count 0
176
- [ALLOC_TEMP_Background Job.Worker 3]
177
- Initial Block Size 32.0 KB
178
- Current Block Size 32.0 KB
179
- Peak Allocated Bytes 0 B
180
- Overflow Count 0
181
- [ALLOC_TEMP_Background Job.Worker 11]
182
- Initial Block Size 32.0 KB
183
- Current Block Size 32.0 KB
184
- Peak Allocated Bytes 0 B
185
- Overflow Count 0
186
- [ALLOC_TEMP_Background Job.Worker 0]
187
- Initial Block Size 32.0 KB
188
- Current Block Size 32.0 KB
189
- Peak Allocated Bytes 0 B
190
- Overflow Count 0
191
- [ALLOC_TEMP_Background Job.Worker 4]
192
- Initial Block Size 32.0 KB
193
- Current Block Size 32.0 KB
194
- Peak Allocated Bytes 0 B
195
- Overflow Count 0
196
- [ALLOC_TEMP_BatchDeleteObjects]
197
- Initial Block Size 64.0 KB
198
- Current Block Size 64.0 KB
199
- Peak Allocated Bytes 0 B
200
- Overflow Count 0
201
- [ALLOC_DEFAULT] Dual Thread Allocator
202
- Peak main deferred allocation count 41
203
- [ALLOC_BUCKET]
204
- Large Block size 4.0 MB
205
- Used Block count 1
206
- Peak Allocated bytes 0.9 MB
207
- [ALLOC_DEFAULT_MAIN]
208
- Peak usage frame count: [4.0 MB-8.0 MB]: 5462 frames
209
- Requested Block Size 16.0 MB
210
- Peak Block count 1
211
- Peak Allocated memory 6.8 MB
212
- Peak Large allocation bytes 0 B
213
- [ALLOC_DEFAULT_THREAD]
214
- Peak usage frame count: [16.0 MB-32.0 MB]: 5462 frames
215
- Requested Block Size 16.0 MB
216
- Peak Block count 1
217
- Peak Allocated memory 17.6 MB
218
- Peak Large allocation bytes 16.0 MB
219
- [ALLOC_TEMP_JOB_1_FRAME]
220
- Initial Block Size 2.0 MB
221
- Used Block Count 1
222
- Overflow Count (too large) 0
223
- Overflow Count (full) 0
224
- [ALLOC_TEMP_JOB_2_FRAMES]
225
- Initial Block Size 2.0 MB
226
- Used Block Count 1
227
- Overflow Count (too large) 0
228
- Overflow Count (full) 0
229
- [ALLOC_TEMP_JOB_4_FRAMES (JobTemp)]
230
- Initial Block Size 2.0 MB
231
- Used Block Count 2
232
- Overflow Count (too large) 0
233
- Overflow Count (full) 0
234
- [ALLOC_TEMP_JOB_ASYNC (Background)]
235
- Initial Block Size 1.0 MB
236
- Used Block Count 3
237
- Overflow Count (too large) 0
238
- Overflow Count (full) 0
239
- [ALLOC_GFX] Dual Thread Allocator
240
- Peak main deferred allocation count 0
241
- [ALLOC_BUCKET]
242
- Large Block size 4.0 MB
243
- Used Block count 1
244
- Peak Allocated bytes 0.9 MB
245
- [ALLOC_GFX_MAIN]
246
- Peak usage frame count: [32.0 KB-64.0 KB]: 4763 frames, [64.0 KB-128.0 KB]: 699 frames
247
- Requested Block Size 16.0 MB
248
- Peak Block count 1
249
- Peak Allocated memory 66.4 KB
250
- Peak Large allocation bytes 0 B
251
- [ALLOC_GFX_THREAD]
252
- Peak usage frame count: [32.0 KB-64.0 KB]: 5462 frames
253
- Requested Block Size 16.0 MB
254
- Peak Block count 1
255
- Peak Allocated memory 39.6 KB
256
- Peak Large allocation bytes 0 B
257
- [ALLOC_CACHEOBJECTS] Dual Thread Allocator
258
- Peak main deferred allocation count 0
259
- [ALLOC_BUCKET]
260
- Large Block size 4.0 MB
261
- Used Block count 1
262
- Peak Allocated bytes 0.9 MB
263
- [ALLOC_CACHEOBJECTS_MAIN]
264
- Peak usage frame count: [0.5 MB-1.0 MB]: 5462 frames
265
- Requested Block Size 4.0 MB
266
- Peak Block count 1
267
- Peak Allocated memory 0.6 MB
268
- Peak Large allocation bytes 0 B
269
- [ALLOC_CACHEOBJECTS_THREAD]
270
- Peak usage frame count: [0.5 MB-1.0 MB]: 5461 frames, [2.0 MB-4.0 MB]: 1 frames
271
- Requested Block Size 4.0 MB
272
- Peak Block count 1
273
- Peak Allocated memory 2.2 MB
274
- Peak Large allocation bytes 0 B
275
- [ALLOC_TYPETREE] Dual Thread Allocator
276
- Peak main deferred allocation count 0
277
- [ALLOC_BUCKET]
278
- Large Block size 4.0 MB
279
- Used Block count 1
280
- Peak Allocated bytes 0.9 MB
281
- [ALLOC_TYPETREE_MAIN]
282
- Peak usage frame count: [0-1.0 KB]: 5462 frames
283
- Requested Block Size 2.0 MB
284
- Peak Block count 1
285
- Peak Allocated memory 1.0 KB
286
- Peak Large allocation bytes 0 B
287
- [ALLOC_TYPETREE_THREAD]
288
- Peak usage frame count: [1.0 KB-2.0 KB]: 5462 frames
289
- Requested Block Size 2.0 MB
290
- Peak Block count 1
291
- Peak Allocated memory 1.7 KB
292
- Peak Large allocation bytes 0 B
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_logs/timers.json CHANGED
@@ -1,382 +1,45 @@
1
  {
2
  "name": "root",
3
- "gauges": {
4
- "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.9843306541442871,
6
- "min": 0.9843306541442871,
7
- "max": 2.8748831748962402,
8
- "count": 20
9
- },
10
- "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 9430.8720703125,
12
- "min": 9430.8720703125,
13
- "max": 29536.55078125,
14
- "count": 20
15
- },
16
- "SnowballTarget.Step.mean": {
17
- "value": 199984.0,
18
- "min": 9952.0,
19
- "max": 199984.0,
20
- "count": 20
21
- },
22
- "SnowballTarget.Step.sum": {
23
- "value": 199984.0,
24
- "min": 9952.0,
25
- "max": 199984.0,
26
- "count": 20
27
- },
28
- "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 12.789898872375488,
30
- "min": 0.41111764311790466,
31
- "max": 12.789898872375488,
32
- "count": 20
33
- },
34
- "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2494.0302734375,
36
- "min": 79.75682067871094,
37
- "max": 2591.443603515625,
38
- "count": 20
39
- },
40
- "SnowballTarget.Environment.EpisodeLength.mean": {
41
- "value": 199.0,
42
- "min": 199.0,
43
- "max": 199.0,
44
- "count": 20
45
- },
46
- "SnowballTarget.Environment.EpisodeLength.sum": {
47
- "value": 8756.0,
48
- "min": 8756.0,
49
- "max": 10945.0,
50
- "count": 20
51
- },
52
- "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.06996092077986081,
54
- "min": 0.06280759313370914,
55
- "max": 0.07853127599639051,
56
- "count": 20
57
- },
58
- "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.27984368311944324,
60
- "min": 0.2540403184853038,
61
- "max": 0.3926563799819526,
62
- "count": 20
63
- },
64
- "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.19596380433615515,
66
- "min": 0.12684643339119195,
67
- "max": 0.3017532670030407,
68
- "count": 20
69
- },
70
- "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.7838552173446206,
72
- "min": 0.5073857335647678,
73
- "max": 1.5087663350152036,
74
- "count": 20
75
- },
76
- "SnowballTarget.Policy.LearningRate.mean": {
77
- "value": 8.082097306000005e-06,
78
- "min": 8.082097306000005e-06,
79
- "max": 0.000291882002706,
80
- "count": 20
81
- },
82
- "SnowballTarget.Policy.LearningRate.sum": {
83
- "value": 3.232838922400002e-05,
84
- "min": 3.232838922400002e-05,
85
- "max": 0.00138516003828,
86
- "count": 20
87
- },
88
- "SnowballTarget.Policy.Epsilon.mean": {
89
- "value": 0.10269400000000001,
90
- "min": 0.10269400000000001,
91
- "max": 0.19729400000000002,
92
- "count": 20
93
- },
94
- "SnowballTarget.Policy.Epsilon.sum": {
95
- "value": 0.41077600000000003,
96
- "min": 0.41077600000000003,
97
- "max": 0.96172,
98
- "count": 20
99
- },
100
- "SnowballTarget.Policy.Beta.mean": {
101
- "value": 0.0001444306000000001,
102
- "min": 0.0001444306000000001,
103
- "max": 0.0048649706,
104
- "count": 20
105
- },
106
- "SnowballTarget.Policy.Beta.sum": {
107
- "value": 0.0005777224000000004,
108
- "min": 0.0005777224000000004,
109
- "max": 0.023089828,
110
- "count": 20
111
- },
112
- "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 25.386363636363637,
114
- "min": 3.3863636363636362,
115
- "max": 25.386363636363637,
116
- "count": 20
117
- },
118
- "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1117.0,
120
- "min": 149.0,
121
- "max": 1379.0,
122
- "count": 20
123
- },
124
- "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 25.386363636363637,
126
- "min": 3.3863636363636362,
127
- "max": 25.386363636363637,
128
- "count": 20
129
- },
130
- "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1117.0,
132
- "min": 149.0,
133
- "max": 1379.0,
134
- "count": 20
135
- },
136
- "SnowballTarget.IsTraining.mean": {
137
- "value": 1.0,
138
- "min": 1.0,
139
- "max": 1.0,
140
- "count": 20
141
- },
142
- "SnowballTarget.IsTraining.sum": {
143
- "value": 1.0,
144
- "min": 1.0,
145
- "max": 1.0,
146
- "count": 20
147
- }
148
- },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1673486342",
152
  "python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
153
- "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "0.29.0.dev0",
155
  "mlagents_envs_version": "0.29.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.8.1+cu102",
158
  "numpy_version": "1.21.6",
159
- "end_time_seconds": "1673486703"
160
  },
161
- "total": 360.75774344399997,
162
  "count": 1,
163
- "self": 0.27468581599987374,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.1040496779996829,
167
  "count": 1,
168
- "self": 0.1040496779996829
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 360.3790079500004,
172
  "count": 1,
173
- "self": 0.4669498790040052,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 6.234684719999677,
177
  "count": 1,
178
- "self": 6.234684719999677
179
- },
180
- "TrainerController.advance": {
181
- "total": 353.56549876599684,
182
- "count": 18206,
183
- "self": 0.22864415497087975,
184
- "children": {
185
- "env_step": {
186
- "total": 353.33685461102596,
187
- "count": 18206,
188
- "self": 240.38119864901228,
189
- "children": {
190
- "SubprocessEnvManager._take_step": {
191
- "total": 112.72627681500899,
192
- "count": 18206,
193
- "self": 1.2353198649680053,
194
- "children": {
195
- "TorchPolicy.evaluate": {
196
- "total": 111.49095695004098,
197
- "count": 18206,
198
- "self": 20.966767412053287,
199
- "children": {
200
- "TorchPolicy.sample_actions": {
201
- "total": 90.5241895379877,
202
- "count": 18206,
203
- "self": 90.5241895379877
204
- }
205
- }
206
- }
207
- }
208
- },
209
- "workers": {
210
- "total": 0.22937914700469264,
211
- "count": 18206,
212
- "self": 0.0,
213
- "children": {
214
- "worker_root": {
215
- "total": 359.6540712289948,
216
- "count": 18206,
217
- "is_parallel": true,
218
- "self": 181.04659040699016,
219
- "children": {
220
- "run_training.setup": {
221
- "total": 0.0,
222
- "count": 0,
223
- "is_parallel": true,
224
- "self": 0.0,
225
- "children": {
226
- "steps_from_proto": {
227
- "total": 0.005211338000208343,
228
- "count": 1,
229
- "is_parallel": true,
230
- "self": 0.003955740000947117,
231
- "children": {
232
- "_process_rank_one_or_two_observation": {
233
- "total": 0.0012555979992612265,
234
- "count": 10,
235
- "is_parallel": true,
236
- "self": 0.0012555979992612265
237
- }
238
- }
239
- },
240
- "UnityEnvironment.step": {
241
- "total": 0.02575548799995886,
242
- "count": 1,
243
- "is_parallel": true,
244
- "self": 0.0005079609995846113,
245
- "children": {
246
- "UnityEnvironment._generate_step_input": {
247
- "total": 0.0003724839998540119,
248
- "count": 1,
249
- "is_parallel": true,
250
- "self": 0.0003724839998540119
251
- },
252
- "communicator.exchange": {
253
- "total": 0.02315245900035734,
254
- "count": 1,
255
- "is_parallel": true,
256
- "self": 0.02315245900035734
257
- },
258
- "steps_from_proto": {
259
- "total": 0.0017225840001628967,
260
- "count": 1,
261
- "is_parallel": true,
262
- "self": 0.0003883700001097168,
263
- "children": {
264
- "_process_rank_one_or_two_observation": {
265
- "total": 0.00133421400005318,
266
- "count": 10,
267
- "is_parallel": true,
268
- "self": 0.00133421400005318
269
- }
270
- }
271
- }
272
- }
273
- }
274
- }
275
- },
276
- "UnityEnvironment.step": {
277
- "total": 178.60748082200462,
278
- "count": 18205,
279
- "is_parallel": true,
280
- "self": 5.730856564036003,
281
- "children": {
282
- "UnityEnvironment._generate_step_input": {
283
- "total": 3.4736756610036537,
284
- "count": 18205,
285
- "is_parallel": true,
286
- "self": 3.4736756610036537
287
- },
288
- "communicator.exchange": {
289
- "total": 149.53251031099444,
290
- "count": 18205,
291
- "is_parallel": true,
292
- "self": 149.53251031099444
293
- },
294
- "steps_from_proto": {
295
- "total": 19.870438285970522,
296
- "count": 18205,
297
- "is_parallel": true,
298
- "self": 4.547301782989052,
299
- "children": {
300
- "_process_rank_one_or_two_observation": {
301
- "total": 15.32313650298147,
302
- "count": 182050,
303
- "is_parallel": true,
304
- "self": 15.32313650298147
305
- }
306
- }
307
- }
308
- }
309
- }
310
- }
311
- }
312
- }
313
- }
314
- }
315
- }
316
- }
317
  },
318
  "trainer_threads": {
319
- "total": 4.24899999416084e-05,
320
  "count": 1,
321
- "self": 4.24899999416084e-05,
322
- "children": {
323
- "thread_root": {
324
- "total": 0.0,
325
- "count": 0,
326
- "is_parallel": true,
327
- "self": 0.0,
328
- "children": {
329
- "trainer_advance": {
330
- "total": 352.7372576001676,
331
- "count": 215900,
332
- "is_parallel": true,
333
- "self": 3.3001920450492435,
334
- "children": {
335
- "process_trajectory": {
336
- "total": 198.2483639781185,
337
- "count": 215900,
338
- "is_parallel": true,
339
- "self": 197.4899532981185,
340
- "children": {
341
- "RLTrainer._checkpoint": {
342
- "total": 0.7584106799999972,
343
- "count": 4,
344
- "is_parallel": true,
345
- "self": 0.7584106799999972
346
- }
347
- }
348
- },
349
- "_update_policy": {
350
- "total": 151.18870157699985,
351
- "count": 90,
352
- "is_parallel": true,
353
- "self": 40.51049466801169,
354
- "children": {
355
- "TorchPPOOptimizer.update": {
356
- "total": 110.67820690898816,
357
- "count": 4587,
358
- "is_parallel": true,
359
- "self": 110.67820690898816
360
- }
361
- }
362
- }
363
- }
364
- }
365
- }
366
- }
367
- }
368
  },
369
  "TrainerController._save_models": {
370
- "total": 0.11183209499995428,
371
  "count": 1,
372
- "self": 0.000804690000222763,
373
- "children": {
374
- "RLTrainer._checkpoint": {
375
- "total": 0.11102740499973152,
376
- "count": 1,
377
- "self": 0.11102740499973152
378
- }
379
- }
380
  }
381
  }
382
  }
 
1
  {
2
  "name": "root",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "metadata": {
4
  "timer_format_version": "0.1.0",
5
+ "start_time_seconds": "1673507950",
6
  "python_version": "3.8.16 (default, Dec 7 2022, 01:12:13) \n[GCC 7.5.0]",
7
+ "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget2 --no-graphics",
8
  "mlagents_version": "0.29.0.dev0",
9
  "mlagents_envs_version": "0.29.0.dev0",
10
  "communication_protocol_version": "1.5.0",
11
  "pytorch_version": "1.8.1+cu102",
12
  "numpy_version": "1.21.6",
13
+ "end_time_seconds": "1673507950"
14
  },
15
+ "total": 0.11813009299976329,
16
  "count": 1,
17
+ "self": 0.005423961999895255,
18
  "children": {
19
  "run_training.setup": {
20
+ "total": 0.10168198399969697,
21
  "count": 1,
22
+ "self": 0.10168198399969697
23
  },
24
  "TrainerController.start_learning": {
25
+ "total": 0.01102414700017107,
26
  "count": 1,
27
+ "self": 0.0003675170000860817,
28
  "children": {
29
  "TrainerController._reset_env": {
30
+ "total": 0.010646658000041498,
31
  "count": 1,
32
+ "self": 0.010646658000041498
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  },
34
  "trainer_threads": {
35
+ "total": 1.5589998838549946e-06,
36
  "count": 1,
37
+ "self": 1.5589998838549946e-06
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  },
39
  "TrainerController._save_models": {
40
+ "total": 8.413000159634976e-06,
41
  "count": 1,
42
+ "self": 8.413000159634976e-06
 
 
 
 
 
 
 
43
  }
44
  }
45
  }
run_logs/training_status.json CHANGED
@@ -1,62 +1,4 @@
1
  {
2
- "SnowballTarget": {
3
- "checkpoints": [
4
- {
5
- "steps": 49936,
6
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 12.090909090909092,
8
- "creation_time": 1673486437.4956174,
9
- "auxillary_file_paths": [
10
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
- ]
12
- },
13
- {
14
- "steps": 99960,
15
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 20.636363636363637,
17
- "creation_time": 1673486526.106195,
18
- "auxillary_file_paths": [
19
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
- ]
21
- },
22
- {
23
- "steps": 149984,
24
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 23.363636363636363,
26
- "creation_time": 1673486614.9631987,
27
- "auxillary_file_paths": [
28
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
- ]
30
- },
31
- {
32
- "steps": 199984,
33
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
- "reward": 25.454545454545453,
35
- "creation_time": 1673486702.9010992,
36
- "auxillary_file_paths": [
37
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
- ]
39
- },
40
- {
41
- "steps": 200112,
42
- "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
- "reward": 25.454545454545453,
44
- "creation_time": 1673486703.0392766,
45
- "auxillary_file_paths": [
46
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
- ]
48
- }
49
- ],
50
- "final_checkpoint": {
51
- "steps": 200112,
52
- "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
- "reward": 25.454545454545453,
54
- "creation_time": 1673486703.0392766,
55
- "auxillary_file_paths": [
56
- "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
- ]
58
- }
59
- },
60
  "metadata": {
61
  "stats_format_version": "0.3.0",
62
  "mlagents_version": "0.29.0.dev0",
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "metadata": {
3
  "stats_format_version": "0.3.0",
4
  "mlagents_version": "0.29.0.dev0",