sanjay-906 commited on
Commit
6ef819f
·
verified ·
1 Parent(s): 0c5cb3e

First Push

Browse files
Pyramids.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f18ca53840cea461c362739e97ab1f6786eeef7743cf477429e8f7eb9f79c0cc
3
+ size 1420238
Pyramids/Pyramids-1000021.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f18ca53840cea461c362739e97ab1f6786eeef7743cf477429e8f7eb9f79c0cc
3
+ size 1420238
Pyramids/Pyramids-1000021.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d280f519f03d2ac4ab848ba7fc3946c97dc2656d4cbb1e7e3f3cc9ef04e195db
3
+ size 8656252
Pyramids/Pyramids-499946.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d677a96e38579e72069ab8fa7dab4e7f11a877290420813800cbc36d36a7685
3
+ size 1420238
Pyramids/Pyramids-499946.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02bfd3e614a370e7bd5778797c93d40c300736d99eb20d9b42019d112d0e76d0
3
+ size 8656173
Pyramids/Pyramids-999893.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f18ca53840cea461c362739e97ab1f6786eeef7743cf477429e8f7eb9f79c0cc
3
+ size 1420238
Pyramids/Pyramids-999893.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c80001a9ed23aa47c5d39feaab0493fcf690f940932fe8d9cfd0ff50d480b522
3
+ size 8656173
Pyramids/checkpoint.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7440e8df15e0c677d25e37455e9af7b0686ce153766b289b1dc4ef23693331e8
3
+ size 8655778
Pyramids/events.out.tfevents.1739076644.2697cd083b0b.7038.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:98a87e496fc189b20efa32e51d9136b9cb33e27cb3afbc7de24f35115d7fcd27
3
+ size 323825
run_logs/Player-0.log CHANGED
@@ -1,7 +1,7 @@
1
  Mono path[0] = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/Managed'
2
  Mono config path = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/MonoBleedingEdge/etc'
3
  Found 1 interfaces on host : 0) 172.28.0.12
4
- Multi-casting "[IP] 172.28.0.12 [Port] 55417 [Flags] 2 [Guid] 2020767234 [EditorId] 764847374 [Version] 1048832 [Id] LinuxPlayer(13,172.28.0.12) [Debug] 0 [PackageName] LinuxPlayer [ProjectName] UnityEnvironment" to [225.0.0.222:54997]...
5
  Preloaded 'lib_burst_generated.so'
6
  Preloaded 'libgrpc_csharp_ext.x64.so'
7
  PlayerPrefs - Creating folder: /root/.config/unity3d/Unity Technologies
@@ -36,7 +36,7 @@ ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
36
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
37
  FMOD initialized on nosound output
38
  Begin MonoManager ReloadAssembly
39
- - Completed reload, in 0.102 seconds
40
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
42
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -68,7 +68,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
68
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
69
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
70
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
71
- UnloadTime: 1.049950 ms
72
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
73
  requesting resize 84 x 84
74
  Setting up 1 worker threads for Enlighten.
@@ -76,7 +76,7 @@ PlayerConnection::CleanupMemory Statistics:
76
  [ALLOC_TEMP_TLS] TLS Allocator
77
  StackAllocators :
78
  [ALLOC_TEMP_MAIN]
79
- Peak usage frame count: [16.0 KB-32.0 KB]: 2653 frames, [32.0 KB-64.0 KB]: 3247 frames, [64.0 KB-128.0 KB]: 377 frames, [2.0 MB-4.0 MB]: 1 frames
80
  Initial Block Size 4.0 MB
81
  Current Block Size 4.0 MB
82
  Peak Allocated Bytes 2.0 MB
@@ -207,28 +207,28 @@ PlayerConnection::CleanupMemory Statistics:
207
  Peak Allocated Bytes 128 B
208
  Overflow Count 0
209
  [ALLOC_MEMORYPROFILER]
210
- Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 5979 frames
211
  Requested Block Size 1.0 MB
212
  Peak Block count 2
213
  Peak Allocated memory 1.6 MB
214
  Peak Large allocation bytes 0 B
215
  [ALLOC_DEFAULT] Dual Thread Allocator
216
- Peak main deferred allocation count 11909
217
  [ALLOC_BUCKET]
218
  Large Block size 4.0 MB
219
  Used Block count 1
220
  Peak Allocated bytes 1.7 MB
221
  [ALLOC_DEFAULT_MAIN]
222
- Peak usage frame count: [8.0 MB-16.0 MB]: 1 frames, [16.0 MB-32.0 MB]: 6277 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 3
225
- Peak Allocated memory 29.6 MB
226
  Peak Large allocation bytes 0 B
227
  [ALLOC_DEFAULT_THREAD]
228
- Peak usage frame count: [16.0 MB-32.0 MB]: 6278 frames
229
  Requested Block Size 16.0 MB
230
  Peak Block count 1
231
- Peak Allocated memory 23.9 MB
232
  Peak Large allocation bytes 16.0 MB
233
  [ALLOC_TEMP_JOB_1_FRAME]
234
  Initial Block Size 2.0 MB
@@ -257,13 +257,13 @@ PlayerConnection::CleanupMemory Statistics:
257
  Used Block count 1
258
  Peak Allocated bytes 1.7 MB
259
  [ALLOC_GFX_MAIN]
260
- Peak usage frame count: [32.0 KB-64.0 KB]: 6277 frames, [64.0 KB-128.0 KB]: 1 frames
261
  Requested Block Size 16.0 MB
262
  Peak Block count 1
263
  Peak Allocated memory 67.0 KB
264
  Peak Large allocation bytes 0 B
265
  [ALLOC_GFX_THREAD]
266
- Peak usage frame count: [64.0 KB-128.0 KB]: 6278 frames
267
  Requested Block Size 16.0 MB
268
  Peak Block count 1
269
  Peak Allocated memory 67.7 KB
@@ -275,13 +275,13 @@ PlayerConnection::CleanupMemory Statistics:
275
  Used Block count 1
276
  Peak Allocated bytes 1.7 MB
277
  [ALLOC_CACHEOBJECTS_MAIN]
278
- Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 5979 frames
279
  Requested Block Size 4.0 MB
280
  Peak Block count 1
281
  Peak Allocated memory 1.4 MB
282
  Peak Large allocation bytes 0 B
283
  [ALLOC_CACHEOBJECTS_THREAD]
284
- Peak usage frame count: [1.0 MB-2.0 MB]: 6277 frames, [4.0 MB-8.0 MB]: 1 frames
285
  Requested Block Size 4.0 MB
286
  Peak Block count 2
287
  Peak Allocated memory 4.7 MB
@@ -293,25 +293,25 @@ PlayerConnection::CleanupMemory Statistics:
293
  Used Block count 1
294
  Peak Allocated bytes 1.7 MB
295
  [ALLOC_TYPETREE_MAIN]
296
- Peak usage frame count: [0-1.0 KB]: 6278 frames
297
  Requested Block Size 2.0 MB
298
  Peak Block count 1
299
  Peak Allocated memory 1.0 KB
300
  Peak Large allocation bytes 0 B
301
  [ALLOC_TYPETREE_THREAD]
302
- Peak usage frame count: [1.0 KB-2.0 KB]: 6278 frames
303
  Requested Block Size 2.0 MB
304
  Peak Block count 1
305
  Peak Allocated memory 1.9 KB
306
  Peak Large allocation bytes 0 B
307
  [ALLOC_PROFILER]
308
- Peak usage frame count: [16.0 KB-32.0 KB]: 6278 frames
309
  Requested Block Size 16.0 MB
310
  Peak Block count 1
311
- Peak Allocated memory 27.4 KB
312
  Peak Large allocation bytes 0 B
313
  [ALLOC_PROFILER_BUCKET]
314
  Large Block size 4.0 MB
315
  Used Block count 1
316
  Peak Allocated bytes 396 B
317
- ##utp:{"type":"MemoryLeaks","version":2,"phase":"Immediate","time":1738486016039,"processId":15196,"allocatedMemory":1970133,"memoryLabels":[{"Default":9033},{"Permanent":1264},{"NewDelete":12801},{"Thread":34460},{"Manager":10603},{"VertexData":12},{"Geometry":280},{"Texture":16},{"Shader":69173},{"Material":24},{"GfxDevice":35248},{"Animation":304},{"Audio":3976},{"Physics":288},{"Serialization":216},{"Input":9176},{"JobScheduler":200},{"Mono":40},{"ScriptingNativeRuntime":216},{"BaseObject":1609212},{"Resource":592},{"Renderer":1936},{"Transform":48},{"File":800},{"WebCam":24},{"Culling":40},{"Terrain":953},{"Wind":24},{"String":3447},{"DynamicArray":30868},{"HashMap":7680},{"Utility":1360},{"PoolAlloc":1160},{"TypeTree":1792},{"ScriptManager":80},{"RuntimeInitializeOnLoadManager":72},{"SpriteAtlas":112},{"GI":3272},{"Unet":16},{"Director":7760},{"WebRequest":720},{"VR":45473},{"SceneManager":424},{"Video":32},{"LazyScriptCache":32},{"NativeArray":384},{"Camera":25},{"Secure":1},{"SerializationCache":624},{"APIUpdating":5872},{"Subsystems":384},{"VirtualTexturing":57552},{"AssetReference":32}]}
 
1
  Mono path[0] = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/Managed'
2
  Mono config path = '/content/ml-agents/training-envs-executables/linux/Pyramids/Pyramids_Data/MonoBleedingEdge/etc'
3
  Found 1 interfaces on host : 0) 172.28.0.12
4
+ Multi-casting "[IP] 172.28.0.12 [Port] 55126 [Flags] 2 [Guid] 653271550 [EditorId] 764847374 [Version] 1048832 [Id] LinuxPlayer(13,172.28.0.12) [Debug] 0 [PackageName] LinuxPlayer [ProjectName] UnityEnvironment" to [225.0.0.222:54997]...
5
  Preloaded 'lib_burst_generated.so'
6
  Preloaded 'libgrpc_csharp_ext.x64.so'
7
  PlayerPrefs - Creating folder: /root/.config/unity3d/Unity Technologies
 
36
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
37
  FMOD initialized on nosound output
38
  Begin MonoManager ReloadAssembly
39
+ - Completed reload, in 0.095 seconds
40
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
41
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
42
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
68
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
69
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
70
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
71
+ UnloadTime: 0.886455 ms
72
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
73
  requesting resize 84 x 84
74
  Setting up 1 worker threads for Enlighten.
 
76
  [ALLOC_TEMP_TLS] TLS Allocator
77
  StackAllocators :
78
  [ALLOC_TEMP_MAIN]
79
+ Peak usage frame count: [16.0 KB-32.0 KB]: 3540 frames, [32.0 KB-64.0 KB]: 12098 frames, [64.0 KB-128.0 KB]: 3127 frames, [2.0 MB-4.0 MB]: 1 frames
80
  Initial Block Size 4.0 MB
81
  Current Block Size 4.0 MB
82
  Peak Allocated Bytes 2.0 MB
 
207
  Peak Allocated Bytes 128 B
208
  Overflow Count 0
209
  [ALLOC_MEMORYPROFILER]
210
+ Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 18467 frames
211
  Requested Block Size 1.0 MB
212
  Peak Block count 2
213
  Peak Allocated memory 1.6 MB
214
  Peak Large allocation bytes 0 B
215
  [ALLOC_DEFAULT] Dual Thread Allocator
216
+ Peak main deferred allocation count 11678
217
  [ALLOC_BUCKET]
218
  Large Block size 4.0 MB
219
  Used Block count 1
220
  Peak Allocated bytes 1.7 MB
221
  [ALLOC_DEFAULT_MAIN]
222
+ Peak usage frame count: [8.0 MB-16.0 MB]: 1 frames, [16.0 MB-32.0 MB]: 18765 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 3
225
+ Peak Allocated memory 31.6 MB
226
  Peak Large allocation bytes 0 B
227
  [ALLOC_DEFAULT_THREAD]
228
+ Peak usage frame count: [16.0 MB-32.0 MB]: 18766 frames
229
  Requested Block Size 16.0 MB
230
  Peak Block count 1
231
+ Peak Allocated memory 22.7 MB
232
  Peak Large allocation bytes 16.0 MB
233
  [ALLOC_TEMP_JOB_1_FRAME]
234
  Initial Block Size 2.0 MB
 
257
  Used Block count 1
258
  Peak Allocated bytes 1.7 MB
259
  [ALLOC_GFX_MAIN]
260
+ Peak usage frame count: [32.0 KB-64.0 KB]: 18765 frames, [64.0 KB-128.0 KB]: 1 frames
261
  Requested Block Size 16.0 MB
262
  Peak Block count 1
263
  Peak Allocated memory 67.0 KB
264
  Peak Large allocation bytes 0 B
265
  [ALLOC_GFX_THREAD]
266
+ Peak usage frame count: [64.0 KB-128.0 KB]: 18766 frames
267
  Requested Block Size 16.0 MB
268
  Peak Block count 1
269
  Peak Allocated memory 67.7 KB
 
275
  Used Block count 1
276
  Peak Allocated bytes 1.7 MB
277
  [ALLOC_CACHEOBJECTS_MAIN]
278
+ Peak usage frame count: [0.5 MB-1.0 MB]: 299 frames, [1.0 MB-2.0 MB]: 18467 frames
279
  Requested Block Size 4.0 MB
280
  Peak Block count 1
281
  Peak Allocated memory 1.4 MB
282
  Peak Large allocation bytes 0 B
283
  [ALLOC_CACHEOBJECTS_THREAD]
284
+ Peak usage frame count: [1.0 MB-2.0 MB]: 18765 frames, [4.0 MB-8.0 MB]: 1 frames
285
  Requested Block Size 4.0 MB
286
  Peak Block count 2
287
  Peak Allocated memory 4.7 MB
 
293
  Used Block count 1
294
  Peak Allocated bytes 1.7 MB
295
  [ALLOC_TYPETREE_MAIN]
296
+ Peak usage frame count: [0-1.0 KB]: 18766 frames
297
  Requested Block Size 2.0 MB
298
  Peak Block count 1
299
  Peak Allocated memory 1.0 KB
300
  Peak Large allocation bytes 0 B
301
  [ALLOC_TYPETREE_THREAD]
302
+ Peak usage frame count: [1.0 KB-2.0 KB]: 18766 frames
303
  Requested Block Size 2.0 MB
304
  Peak Block count 1
305
  Peak Allocated memory 1.9 KB
306
  Peak Large allocation bytes 0 B
307
  [ALLOC_PROFILER]
308
+ Peak usage frame count: [16.0 KB-32.0 KB]: 18766 frames
309
  Requested Block Size 16.0 MB
310
  Peak Block count 1
311
+ Peak Allocated memory 29.2 KB
312
  Peak Large allocation bytes 0 B
313
  [ALLOC_PROFILER_BUCKET]
314
  Large Block size 4.0 MB
315
  Used Block count 1
316
  Peak Allocated bytes 396 B
317
+ ##utp:{"type":"MemoryLeaks","version":2,"phase":"Immediate","time":1739078916063,"processId":7059,"allocatedMemory":1970133,"memoryLabels":[{"Default":9033},{"Permanent":1264},{"NewDelete":12801},{"Thread":34460},{"Manager":10603},{"VertexData":12},{"Geometry":280},{"Texture":16},{"Shader":69173},{"Material":24},{"GfxDevice":35248},{"Animation":304},{"Audio":3976},{"Physics":288},{"Serialization":216},{"Input":9176},{"JobScheduler":200},{"Mono":40},{"ScriptingNativeRuntime":216},{"BaseObject":1609212},{"Resource":592},{"Renderer":1936},{"Transform":48},{"File":800},{"WebCam":24},{"Culling":40},{"Terrain":953},{"Wind":24},{"String":3447},{"DynamicArray":30868},{"HashMap":7680},{"Utility":1360},{"PoolAlloc":1160},{"TypeTree":1792},{"ScriptManager":80},{"RuntimeInitializeOnLoadManager":72},{"SpriteAtlas":112},{"GI":3272},{"Unet":16},{"Director":7760},{"WebRequest":720},{"VR":45473},{"SceneManager":424},{"Video":32},{"LazyScriptCache":32},{"NativeArray":384},{"Camera":25},{"Secure":1},{"SerializationCache":624},{"APIUpdating":5872},{"Subsystems":384},{"VirtualTexturing":57552},{"AssetReference":32}]}
run_logs/timers.json CHANGED
@@ -2,189 +2,189 @@
2
  "name": "root",
3
  "gauges": {
4
  "Pyramids.Policy.Entropy.mean": {
5
- "value": 1.179793119430542,
6
- "min": 0.9126117825508118,
7
- "max": 1.4288430213928223,
8
- "count": 11
9
  },
10
  "Pyramids.Policy.Entropy.sum": {
11
- "value": 35337.1640625,
12
- "min": 27334.548828125,
13
- "max": 43345.3828125,
14
- "count": 11
15
  },
16
  "Pyramids.Step.mean": {
17
- "value": 329971.0,
18
  "min": 29952.0,
19
- "max": 329971.0,
20
- "count": 11
21
  },
22
  "Pyramids.Step.sum": {
23
- "value": 329971.0,
24
  "min": 29952.0,
25
- "max": 329971.0,
26
- "count": 11
27
  },
28
  "Pyramids.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": -0.06974944472312927,
30
- "min": -0.11756870150566101,
31
- "max": -0.05136850103735924,
32
- "count": 11
33
  },
34
  "Pyramids.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": -16.809616088867188,
36
- "min": -28.334056854248047,
37
- "max": -12.37980842590332,
38
- "count": 11
39
  },
40
  "Pyramids.Policy.RndValueEstimate.mean": {
41
- "value": 0.028045665472745895,
42
- "min": 0.028045665472745895,
43
- "max": 0.4137362837791443,
44
- "count": 11
45
  },
46
  "Pyramids.Policy.RndValueEstimate.sum": {
47
- "value": 6.759005546569824,
48
- "min": 6.759005546569824,
49
- "max": 98.05549621582031,
50
- "count": 11
51
  },
52
  "Pyramids.Losses.PolicyLoss.mean": {
53
- "value": 0.07004367525921738,
54
- "min": 0.06672222652882145,
55
- "max": 0.07416302564273448,
56
- "count": 11
57
  },
58
  "Pyramids.Losses.PolicyLoss.sum": {
59
- "value": 0.9806114536290435,
60
- "min": 0.4970513874134343,
61
- "max": 0.9806114536290435,
62
- "count": 11
63
  },
64
  "Pyramids.Losses.ValueLoss.mean": {
65
- "value": 0.001212557160188258,
66
- "min": 0.0002448085269279316,
67
- "max": 0.017204903678673854,
68
- "count": 11
69
  },
70
  "Pyramids.Losses.ValueLoss.sum": {
71
- "value": 0.016975800242635612,
72
- "min": 0.0031825108500631106,
73
- "max": 0.12043432575071697,
74
- "count": 11
75
  },
76
  "Pyramids.Policy.LearningRate.mean": {
77
- "value": 0.00020550693149769999,
78
- "min": 0.00020550693149769999,
79
  "max": 0.00029515063018788575,
80
- "count": 11
81
  },
82
  "Pyramids.Policy.LearningRate.sum": {
83
- "value": 0.0028770970409678,
84
- "min": 0.0020660544113152,
85
- "max": 0.0036087543970818995,
86
- "count": 11
87
  },
88
  "Pyramids.Policy.Epsilon.mean": {
89
- "value": 0.16850229999999997,
90
- "min": 0.16850229999999997,
91
  "max": 0.19838354285714285,
92
- "count": 11
93
  },
94
  "Pyramids.Policy.Epsilon.sum": {
95
- "value": 2.3590321999999997,
96
  "min": 1.3886848,
97
- "max": 2.5029180999999996,
98
- "count": 11
99
  },
100
  "Pyramids.Policy.Beta.mean": {
101
- "value": 0.006853379769999999,
102
- "min": 0.006853379769999999,
103
  "max": 0.00983851593142857,
104
- "count": 11
105
  },
106
  "Pyramids.Policy.Beta.sum": {
107
- "value": 0.09594731677999999,
108
- "min": 0.06886961152,
109
- "max": 0.12030151818999998,
110
- "count": 11
111
  },
112
  "Pyramids.Losses.RNDLoss.mean": {
113
- "value": 0.02431432716548443,
114
- "min": 0.02431432716548443,
115
- "max": 0.5971757769584656,
116
- "count": 11
117
  },
118
  "Pyramids.Losses.RNDLoss.sum": {
119
- "value": 0.3404005765914917,
120
- "min": 0.3404005765914917,
121
- "max": 4.180230617523193,
122
- "count": 11
123
  },
124
  "Pyramids.Environment.EpisodeLength.mean": {
125
- "value": 955.7741935483871,
126
- "min": 940.0588235294117,
127
  "max": 999.0,
128
- "count": 11
129
  },
130
  "Pyramids.Environment.EpisodeLength.sum": {
131
- "value": 29629.0,
132
  "min": 15984.0,
133
- "max": 32082.0,
134
- "count": 11
135
  },
136
  "Pyramids.Environment.CumulativeReward.mean": {
137
- "value": -0.6982452095516266,
138
  "min": -1.0000000521540642,
139
- "max": -0.5876471067176146,
140
- "count": 11
141
  },
142
  "Pyramids.Environment.CumulativeReward.sum": {
143
- "value": -21.645601496100426,
144
- "min": -31.995201662182808,
145
- "max": -16.000000834465027,
146
- "count": 11
147
  },
148
  "Pyramids.Policy.ExtrinsicReward.mean": {
149
- "value": -0.6982452095516266,
150
  "min": -1.0000000521540642,
151
- "max": -0.5876471067176146,
152
- "count": 11
153
  },
154
  "Pyramids.Policy.ExtrinsicReward.sum": {
155
- "value": -21.645601496100426,
156
- "min": -31.995201662182808,
157
- "max": -16.000000834465027,
158
- "count": 11
159
  },
160
  "Pyramids.Policy.RndReward.mean": {
161
- "value": 0.23877150024617871,
162
- "min": 0.23877150024617871,
163
- "max": 13.660724414512515,
164
- "count": 11
165
  },
166
  "Pyramids.Policy.RndReward.sum": {
167
- "value": 7.40191650763154,
168
- "min": 7.40191650763154,
169
- "max": 218.57159063220024,
170
- "count": 11
171
  },
172
  "Pyramids.IsTraining.mean": {
173
  "value": 1.0,
174
  "min": 1.0,
175
  "max": 1.0,
176
- "count": 11
177
  },
178
  "Pyramids.IsTraining.sum": {
179
  "value": 1.0,
180
  "min": 1.0,
181
  "max": 1.0,
182
- "count": 11
183
  }
184
  },
185
  "metadata": {
186
  "timer_format_version": "0.1.0",
187
- "start_time_seconds": "1738485355",
188
  "python_version": "3.10.12 (main, Jan 17 2025, 14:35:34) [GCC 11.4.0]",
189
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
190
  "mlagents_version": "1.2.0.dev0",
@@ -192,59 +192,59 @@
192
  "communication_protocol_version": "1.5.0",
193
  "pytorch_version": "2.6.0+cu124",
194
  "numpy_version": "1.23.5",
195
- "end_time_seconds": "1738486016"
196
  },
197
- "total": 660.7835122830002,
198
  "count": 1,
199
- "self": 0.47981536499992217,
200
  "children": {
201
  "run_training.setup": {
202
- "total": 0.024658480000198324,
203
  "count": 1,
204
- "self": 0.024658480000198324
205
  },
206
  "TrainerController.start_learning": {
207
- "total": 660.279038438,
208
  "count": 1,
209
- "self": 0.3969741110081486,
210
  "children": {
211
  "TrainerController._reset_env": {
212
- "total": 2.2280916830000024,
213
  "count": 1,
214
- "self": 2.2280916830000024
215
  },
216
  "TrainerController.advance": {
217
- "total": 657.6535209349909,
218
- "count": 21063,
219
- "self": 0.4425503390020822,
220
  "children": {
221
  "env_step": {
222
- "total": 431.66776145894255,
223
- "count": 21063,
224
- "self": 381.88409126698343,
225
  "children": {
226
  "SubprocessEnvManager._take_step": {
227
- "total": 49.540611742976125,
228
- "count": 21063,
229
- "self": 1.4700073779204104,
230
  "children": {
231
  "TorchPolicy.evaluate": {
232
- "total": 48.070604365055715,
233
- "count": 20926,
234
- "self": 48.070604365055715
235
  }
236
  }
237
  },
238
  "workers": {
239
- "total": 0.24305844898299256,
240
- "count": 21062,
241
  "self": 0.0,
242
  "children": {
243
  "worker_root": {
244
- "total": 658.823163729,
245
- "count": 21062,
246
  "is_parallel": true,
247
- "self": 312.0380708139742,
248
  "children": {
249
  "run_training.setup": {
250
  "total": 0.0,
@@ -253,48 +253,48 @@
253
  "self": 0.0,
254
  "children": {
255
  "steps_from_proto": {
256
- "total": 0.001981133000299451,
257
  "count": 1,
258
  "is_parallel": true,
259
- "self": 0.0006336440010272781,
260
  "children": {
261
  "_process_rank_one_or_two_observation": {
262
- "total": 0.0013474889992721728,
263
  "count": 8,
264
  "is_parallel": true,
265
- "self": 0.0013474889992721728
266
  }
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 0.04955500999994911,
271
  "count": 1,
272
  "is_parallel": true,
273
- "self": 0.0005343220000213478,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 0.0004600269999173179,
277
  "count": 1,
278
  "is_parallel": true,
279
- "self": 0.0004600269999173179
280
  },
281
  "communicator.exchange": {
282
- "total": 0.04684041500013336,
283
  "count": 1,
284
  "is_parallel": true,
285
- "self": 0.04684041500013336
286
  },
287
  "steps_from_proto": {
288
- "total": 0.0017202459998770792,
289
  "count": 1,
290
  "is_parallel": true,
291
- "self": 0.0004296430006434093,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 0.0012906029992336698,
295
  "count": 8,
296
  "is_parallel": true,
297
- "self": 0.0012906029992336698
298
  }
299
  }
300
  }
@@ -303,34 +303,34 @@
303
  }
304
  },
305
  "UnityEnvironment.step": {
306
- "total": 346.78509291502587,
307
- "count": 21061,
308
  "is_parallel": true,
309
- "self": 10.04582730004131,
310
  "children": {
311
  "UnityEnvironment._generate_step_input": {
312
- "total": 7.525484949988368,
313
- "count": 21061,
314
  "is_parallel": true,
315
- "self": 7.525484949988368
316
  },
317
  "communicator.exchange": {
318
- "total": 298.43514139397985,
319
- "count": 21061,
320
  "is_parallel": true,
321
- "self": 298.43514139397985
322
  },
323
  "steps_from_proto": {
324
- "total": 30.778639271016345,
325
- "count": 21061,
326
  "is_parallel": true,
327
- "self": 6.015530061970367,
328
  "children": {
329
  "_process_rank_one_or_two_observation": {
330
- "total": 24.76310920904598,
331
- "count": 168488,
332
  "is_parallel": true,
333
- "self": 24.76310920904598
334
  }
335
  }
336
  }
@@ -343,24 +343,31 @@
343
  }
344
  },
345
  "trainer_advance": {
346
- "total": 225.54320913704623,
347
- "count": 21062,
348
- "self": 0.6303866070538788,
349
  "children": {
350
  "process_trajectory": {
351
- "total": 39.77186068799392,
352
- "count": 21062,
353
- "self": 39.77186068799392
 
 
 
 
 
 
 
354
  },
355
  "_update_policy": {
356
- "total": 185.14096184199843,
357
- "count": 140,
358
- "self": 102.63071788501293,
359
  "children": {
360
  "TorchPPOOptimizer.update": {
361
- "total": 82.5102439569855,
362
- "count": 7659,
363
- "self": 82.5102439569855
364
  }
365
  }
366
  }
@@ -369,19 +376,19 @@
369
  }
370
  },
371
  "trainer_threads": {
372
- "total": 9.160000445262995e-06,
373
  "count": 1,
374
- "self": 9.160000445262995e-06
375
  },
376
  "TrainerController._save_models": {
377
- "total": 0.0004425490005814936,
378
  "count": 1,
379
- "self": 2.3487000362365507e-05,
380
  "children": {
381
  "RLTrainer._checkpoint": {
382
- "total": 0.0004190620002191281,
383
  "count": 1,
384
- "self": 0.0004190620002191281
385
  }
386
  }
387
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "Pyramids.Policy.Entropy.mean": {
5
+ "value": 0.38364672660827637,
6
+ "min": 0.38364672660827637,
7
+ "max": 1.459175944328308,
8
+ "count": 33
9
  },
10
  "Pyramids.Policy.Entropy.sum": {
11
+ "value": 11490.986328125,
12
+ "min": 11490.986328125,
13
+ "max": 44265.5625,
14
+ "count": 33
15
  },
16
  "Pyramids.Step.mean": {
17
+ "value": 989911.0,
18
  "min": 29952.0,
19
+ "max": 989911.0,
20
+ "count": 33
21
  },
22
  "Pyramids.Step.sum": {
23
+ "value": 989911.0,
24
  "min": 29952.0,
25
+ "max": 989911.0,
26
+ "count": 33
27
  },
28
  "Pyramids.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 0.6232269406318665,
30
+ "min": -0.12185036391019821,
31
+ "max": 0.6492074131965637,
32
+ "count": 33
33
  },
34
  "Pyramids.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 175.75,
36
+ "min": -29.365938186645508,
37
+ "max": 181.778076171875,
38
+ "count": 33
39
  },
40
  "Pyramids.Policy.RndValueEstimate.mean": {
41
+ "value": -0.006430040113627911,
42
+ "min": -0.014972793869674206,
43
+ "max": 0.5385123491287231,
44
+ "count": 33
45
  },
46
  "Pyramids.Policy.RndValueEstimate.sum": {
47
+ "value": -1.8132712841033936,
48
+ "min": -3.9078991413116455,
49
+ "max": 127.62742614746094,
50
+ "count": 33
51
  },
52
  "Pyramids.Losses.PolicyLoss.mean": {
53
+ "value": 0.06903137952446871,
54
+ "min": 0.06533447390318876,
55
+ "max": 0.07313739010307739,
56
+ "count": 33
57
  },
58
  "Pyramids.Losses.PolicyLoss.sum": {
59
+ "value": 0.9664393133425618,
60
+ "min": 0.5037453645270253,
61
+ "max": 1.0635975119749121,
62
+ "count": 33
63
  },
64
  "Pyramids.Losses.ValueLoss.mean": {
65
+ "value": 0.014940131267177916,
66
+ "min": 0.0013474361503178396,
67
+ "max": 0.01710394765692241,
68
+ "count": 33
69
  },
70
  "Pyramids.Losses.ValueLoss.sum": {
71
+ "value": 0.20916183774049082,
72
+ "min": 0.018425317096706535,
73
+ "max": 0.24555822061180757,
74
+ "count": 33
75
  },
76
  "Pyramids.Policy.LearningRate.mean": {
77
+ "value": 7.51511892356429e-06,
78
+ "min": 7.51511892356429e-06,
79
  "max": 0.00029515063018788575,
80
+ "count": 33
81
  },
82
  "Pyramids.Policy.LearningRate.sum": {
83
+ "value": 0.00010521166492990006,
84
+ "min": 0.00010521166492990006,
85
+ "max": 0.0035087264304246,
86
+ "count": 33
87
  },
88
  "Pyramids.Policy.Epsilon.mean": {
89
+ "value": 0.10250500714285715,
90
+ "min": 0.10250500714285715,
91
  "max": 0.19838354285714285,
92
+ "count": 33
93
  },
94
  "Pyramids.Policy.Epsilon.sum": {
95
+ "value": 1.4350701000000001,
96
  "min": 1.3886848,
97
+ "max": 2.5695754,
98
+ "count": 33
99
  },
100
  "Pyramids.Policy.Beta.mean": {
101
+ "value": 0.00026025021357142873,
102
+ "min": 0.00026025021357142873,
103
  "max": 0.00983851593142857,
104
+ "count": 33
105
  },
106
  "Pyramids.Policy.Beta.sum": {
107
+ "value": 0.003643502990000002,
108
+ "min": 0.003643502990000002,
109
+ "max": 0.11698058246000001,
110
+ "count": 33
111
  },
112
  "Pyramids.Losses.RNDLoss.mean": {
113
+ "value": 0.012131420895457268,
114
+ "min": 0.012131420895457268,
115
+ "max": 0.720535397529602,
116
+ "count": 33
117
  },
118
  "Pyramids.Losses.RNDLoss.sum": {
119
+ "value": 0.16983988881111145,
120
+ "min": 0.16983988881111145,
121
+ "max": 5.043747901916504,
122
+ "count": 33
123
  },
124
  "Pyramids.Environment.EpisodeLength.mean": {
125
+ "value": 322.9789473684211,
126
+ "min": 287.25,
127
  "max": 999.0,
128
+ "count": 33
129
  },
130
  "Pyramids.Environment.EpisodeLength.sum": {
131
+ "value": 30683.0,
132
  "min": 15984.0,
133
+ "max": 32190.0,
134
+ "count": 33
135
  },
136
  "Pyramids.Environment.CumulativeReward.mean": {
137
+ "value": 1.613842091748589,
138
  "min": -1.0000000521540642,
139
+ "max": 1.7145463741317237,
140
+ "count": 33
141
  },
142
  "Pyramids.Environment.CumulativeReward.sum": {
143
+ "value": 153.31499871611595,
144
+ "min": -30.22120165079832,
145
+ "max": 171.56659818440676,
146
+ "count": 33
147
  },
148
  "Pyramids.Policy.ExtrinsicReward.mean": {
149
+ "value": 1.613842091748589,
150
  "min": -1.0000000521540642,
151
+ "max": 1.7145463741317237,
152
+ "count": 33
153
  },
154
  "Pyramids.Policy.ExtrinsicReward.sum": {
155
+ "value": 153.31499871611595,
156
+ "min": -30.22120165079832,
157
+ "max": 171.56659818440676,
158
+ "count": 33
159
  },
160
  "Pyramids.Policy.RndReward.mean": {
161
+ "value": 0.040745640572430075,
162
+ "min": 0.03720512080979029,
163
+ "max": 14.474275099113584,
164
+ "count": 33
165
  },
166
  "Pyramids.Policy.RndReward.sum": {
167
+ "value": 3.870835854380857,
168
+ "min": 3.6088967185496585,
169
+ "max": 231.58840158581734,
170
+ "count": 33
171
  },
172
  "Pyramids.IsTraining.mean": {
173
  "value": 1.0,
174
  "min": 1.0,
175
  "max": 1.0,
176
+ "count": 33
177
  },
178
  "Pyramids.IsTraining.sum": {
179
  "value": 1.0,
180
  "min": 1.0,
181
  "max": 1.0,
182
+ "count": 33
183
  }
184
  },
185
  "metadata": {
186
  "timer_format_version": "0.1.0",
187
+ "start_time_seconds": "1739076643",
188
  "python_version": "3.10.12 (main, Jan 17 2025, 14:35:34) [GCC 11.4.0]",
189
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=Pyramids Training --no-graphics",
190
  "mlagents_version": "1.2.0.dev0",
 
192
  "communication_protocol_version": "1.5.0",
193
  "pytorch_version": "2.6.0+cu124",
194
  "numpy_version": "1.23.5",
195
+ "end_time_seconds": "1739078916"
196
  },
197
+ "total": 2272.8826874469996,
198
  "count": 1,
199
+ "self": 0.5268317149998438,
200
  "children": {
201
  "run_training.setup": {
202
+ "total": 0.018968414000028133,
203
  "count": 1,
204
+ "self": 0.018968414000028133
205
  },
206
  "TrainerController.start_learning": {
207
+ "total": 2272.3368873179998,
208
  "count": 1,
209
+ "self": 1.4268662989629775,
210
  "children": {
211
  "TrainerController._reset_env": {
212
+ "total": 2.554789932999938,
213
  "count": 1,
214
+ "self": 2.554789932999938
215
  },
216
  "TrainerController.advance": {
217
+ "total": 2268.266272111037,
218
+ "count": 63936,
219
+ "self": 1.4746662989577999,
220
  "children": {
221
  "env_step": {
222
+ "total": 1573.5335474690655,
223
+ "count": 63936,
224
+ "self": 1414.928939595101,
225
  "children": {
226
  "SubprocessEnvManager._take_step": {
227
+ "total": 157.7747828600218,
228
+ "count": 63936,
229
+ "self": 4.770811132075778,
230
  "children": {
231
  "TorchPolicy.evaluate": {
232
+ "total": 153.003971727946,
233
+ "count": 62551,
234
+ "self": 153.003971727946
235
  }
236
  }
237
  },
238
  "workers": {
239
+ "total": 0.8298250139428092,
240
+ "count": 63936,
241
  "self": 0.0,
242
  "children": {
243
  "worker_root": {
244
+ "total": 2267.1156505549857,
245
+ "count": 63936,
246
  "is_parallel": true,
247
+ "self": 967.3725625719514,
248
  "children": {
249
  "run_training.setup": {
250
  "total": 0.0,
 
253
  "self": 0.0,
254
  "children": {
255
  "steps_from_proto": {
256
+ "total": 0.002021134999949936,
257
  "count": 1,
258
  "is_parallel": true,
259
+ "self": 0.0006381190003139636,
260
  "children": {
261
  "_process_rank_one_or_two_observation": {
262
+ "total": 0.0013830159996359725,
263
  "count": 8,
264
  "is_parallel": true,
265
+ "self": 0.0013830159996359725
266
  }
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 0.08994533300005969,
271
  "count": 1,
272
  "is_parallel": true,
273
+ "self": 0.0005137139999078499,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 0.0004639890000817104,
277
  "count": 1,
278
  "is_parallel": true,
279
+ "self": 0.0004639890000817104
280
  },
281
  "communicator.exchange": {
282
+ "total": 0.08731925100005355,
283
  "count": 1,
284
  "is_parallel": true,
285
+ "self": 0.08731925100005355
286
  },
287
  "steps_from_proto": {
288
+ "total": 0.0016483790000165754,
289
  "count": 1,
290
  "is_parallel": true,
291
+ "self": 0.00035842799979945994,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 0.0012899510002171155,
295
  "count": 8,
296
  "is_parallel": true,
297
+ "self": 0.0012899510002171155
298
  }
299
  }
300
  }
 
303
  }
304
  },
305
  "UnityEnvironment.step": {
306
+ "total": 1299.7430879830342,
307
+ "count": 63935,
308
  "is_parallel": true,
309
+ "self": 31.532362687061095,
310
  "children": {
311
  "UnityEnvironment._generate_step_input": {
312
+ "total": 22.92380315401124,
313
+ "count": 63935,
314
  "is_parallel": true,
315
+ "self": 22.92380315401124
316
  },
317
  "communicator.exchange": {
318
+ "total": 1147.5995340329728,
319
+ "count": 63935,
320
  "is_parallel": true,
321
+ "self": 1147.5995340329728
322
  },
323
  "steps_from_proto": {
324
+ "total": 97.68738810898913,
325
+ "count": 63935,
326
  "is_parallel": true,
327
+ "self": 19.866832524058736,
328
  "children": {
329
  "_process_rank_one_or_two_observation": {
330
+ "total": 77.82055558493039,
331
+ "count": 511480,
332
  "is_parallel": true,
333
+ "self": 77.82055558493039
334
  }
335
  }
336
  }
 
343
  }
344
  },
345
  "trainer_advance": {
346
+ "total": 693.2580583430135,
347
+ "count": 63936,
348
+ "self": 2.6799836619782127,
349
  "children": {
350
  "process_trajectory": {
351
+ "total": 132.26090501903354,
352
+ "count": 63936,
353
+ "self": 131.86275718103343,
354
+ "children": {
355
+ "RLTrainer._checkpoint": {
356
+ "total": 0.39814783800011355,
357
+ "count": 2,
358
+ "self": 0.39814783800011355
359
+ }
360
+ }
361
  },
362
  "_update_policy": {
363
+ "total": 558.3171696620018,
364
+ "count": 453,
365
+ "self": 309.53836960001513,
366
  "children": {
367
  "TorchPPOOptimizer.update": {
368
+ "total": 248.77880006198666,
369
+ "count": 22773,
370
+ "self": 248.77880006198666
371
  }
372
  }
373
  }
 
376
  }
377
  },
378
  "trainer_threads": {
379
+ "total": 9.54999904934084e-07,
380
  "count": 1,
381
+ "self": 9.54999904934084e-07
382
  },
383
  "TrainerController._save_models": {
384
+ "total": 0.0889580200000637,
385
  "count": 1,
386
+ "self": 0.0015613450000273588,
387
  "children": {
388
  "RLTrainer._checkpoint": {
389
+ "total": 0.08739667500003634,
390
  "count": 1,
391
+ "self": 0.08739667500003634
392
  }
393
  }
394
  }
run_logs/training_status.json CHANGED
@@ -1,4 +1,44 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "metadata": {
3
  "stats_format_version": "0.3.0",
4
  "mlagents_version": "1.2.0.dev0",
 
1
  {
2
+ "Pyramids": {
3
+ "checkpoints": [
4
+ {
5
+ "steps": 499946,
6
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-499946.onnx",
7
+ "reward": -0.16386672109365463,
8
+ "creation_time": 1739077694.2716844,
9
+ "auxillary_file_paths": [
10
+ "results/Pyramids Training/Pyramids/Pyramids-499946.pt"
11
+ ]
12
+ },
13
+ {
14
+ "steps": 999893,
15
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-999893.onnx",
16
+ "reward": 1.7355000004172325,
17
+ "creation_time": 1739078915.5107608,
18
+ "auxillary_file_paths": [
19
+ "results/Pyramids Training/Pyramids/Pyramids-999893.pt"
20
+ ]
21
+ },
22
+ {
23
+ "steps": 1000021,
24
+ "file_path": "results/Pyramids Training/Pyramids/Pyramids-1000021.onnx",
25
+ "reward": 1.7355000004172325,
26
+ "creation_time": 1739078915.611193,
27
+ "auxillary_file_paths": [
28
+ "results/Pyramids Training/Pyramids/Pyramids-1000021.pt"
29
+ ]
30
+ }
31
+ ],
32
+ "final_checkpoint": {
33
+ "steps": 1000021,
34
+ "file_path": "results/Pyramids Training/Pyramids.onnx",
35
+ "reward": 1.7355000004172325,
36
+ "creation_time": 1739078915.611193,
37
+ "auxillary_file_paths": [
38
+ "results/Pyramids Training/Pyramids/Pyramids-1000021.pt"
39
+ ]
40
+ }
41
+ },
42
  "metadata": {
43
  "stats_format_version": "0.3.0",
44
  "mlagents_version": "1.2.0.dev0",