marcus07 commited on
Commit
ff00502
·
verified ·
1 Parent(s): 930e6d8

First Push

Browse files
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3653f7fab889c920e28ccc34d44002937d93808d36ed054f7804996b74950a55
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d12b37aff962cdfacad8f5f34057a067f5ab2342a02ad632c5d9787d931a4836
3
  size 650646
SnowballTarget/SnowballTarget-134968.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e38ac7eb504106c77435485725b0d3219a4a1a14288b7223bd868402ed0d40ef
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e272dd94b2b4323a21cae9b5787776808458620793af1f465ee7a462fc0b0094
3
  size 650646
SnowballTarget/SnowballTarget-134968.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e87fc2a38453f0576daee4c94087e41cb2d060dbfa0a9e738cd682c7d425a5a
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e80ce904d910975285384a1c85fcd345d03b851c380d572536a64008fccdc18e
3
  size 3850075
SnowballTarget/SnowballTarget-179992.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b6f733a5edb169fe0ebb24dd29dfbb164fcad001f91142b68625bda22d66801
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:433d47f42efd2b2fca3e8d23cc9b5503ba88c06843840551cccf1a031e336bb1
3
  size 650646
SnowballTarget/SnowballTarget-179992.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea020f7a3054b6718868ad8f7bedefc2dcecc95beeefc043e6a516b9f86d3727
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a37db718d35f14d327fe83b3c6b00061e99f8064141821ffb0e5a1683e89d81
3
  size 3850075
SnowballTarget/SnowballTarget-224976.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa22120307bc62c0230746867a92df6e43559802c5d0a4d43348382d99f42d1b
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef3a940c4c6d7f8e9cce9c60d6966ec1a8b1eba3b4e0017da3d5067799b50926
3
  size 650646
SnowballTarget/SnowballTarget-224976.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:beaf73fccebf6b99fc32159eb1b465d4db9a36f21700f1a95954883a589c13e2
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ed5d46d3c69f5fee37611cb13b0ae3ce94c76337949a9b907cd9f8fe54f8d2c
3
  size 3850075
SnowballTarget/SnowballTarget-269936.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54e4d5e641b13c368616402f2f5ebce2547a7890f70eb9a07118b5725369362a
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be80904b425c5c0edb6149fa325914a6a8115dcb3e7a10a51dd47b3f4b183b7d
3
  size 650646
SnowballTarget/SnowballTarget-269936.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66a0eecf0fdf5a99322aeb08d57954521cdc94d92175ea83a503648f6ed0e750
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c55600bee664b2b392c822bb071fc804d728b0b7d132f1143434622d2c4030ae
3
  size 3850075
SnowballTarget/SnowballTarget-314984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ec9178e61c3c0f0ceb206d78e2077376e90b3e2a29bc701548fb26e242a0a8d1
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c9c083d73a3e1f6b3d42aba4cee0d5abf1f7eaf72e1793ac1b436c786d7bea2
3
  size 650646
SnowballTarget/SnowballTarget-314984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d78cb32485a610d1dd81fe1acf7dcccb43151e20f43f09ed654fadb623b2aff6
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02cc8b232f78b89f5dbdc286871cbf83844d5f52384bb387590e198a4af78b71
3
  size 3850075
SnowballTarget/SnowballTarget-359944.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:269615da8b86a40df118d9cb6ce754dfba26239f0861055d0e2d987ab401a166
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25a0cfe145d2453028e16d235b6fc8e169fd5357cca9f5e4612c8d0ba5b9bf01
3
  size 650646
SnowballTarget/SnowballTarget-359944.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:632e5ddfae59d03ad384665613ca6bd0f9407cc5545ac3a0cab42df75a4efe5f
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f58fd1287196ec68292cfaf8e1d81047bd3fa3db40decc98ac915db8529ee649
3
  size 3850075
SnowballTarget/SnowballTarget-404992.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5fabfde31977bae6875484c89cf9ed94bfd872b8100c4308762b29e93fb8fd9
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fce01043004348630832137a8b51b2fbaedf4e181d23529f8f73e94ef4715986
3
  size 650646
SnowballTarget/SnowballTarget-404992.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c82b1fa602f0dabe8545b45f23835e0faa1f25b12b13db7e39daecff5eec758d
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db8e1c04b320d08caebf338b88e89b7c261beb5331969793b0d465f23a327603
3
  size 3850075
SnowballTarget/SnowballTarget-449952.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e5e478803f002e2a103d845e322b0a39485c80a14b1409c926c33811729069f
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c1758f37505eec83397f7540275342e65aecff3b48ef2e4a6ba5cd375dc8742
3
  size 650646
SnowballTarget/SnowballTarget-449952.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48cf4cbe73cf849ca5b1e50ac1c4cd3a658cac045060213501e982c73786e390
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7ce00e0572c8550d35c5cf098d8c5979f17443e8dcf17ed04777db6ac372bfe
3
  size 3850075
SnowballTarget/SnowballTarget-494992.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:273ce1c7123d04c5c3d37037e70eaabd99459bfd318a36f9615e80186851b668
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b351ceaa745dc71281312884c81eb5bfb7f0b92afe2d449460f8292e52d995b
3
  size 650646
SnowballTarget/SnowballTarget-494992.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8a14f46d24ffa87353b8a6a37a4df74099528a3dddca6dcae5b150f38d3da41a
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:432ff02a1f10f29348f87daa4eb9e355bd3218db53783fce355f9094398ffe20
3
  size 3850075
SnowballTarget/SnowballTarget-500104.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3653f7fab889c920e28ccc34d44002937d93808d36ed054f7804996b74950a55
3
  size 650646
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d12b37aff962cdfacad8f5f34057a067f5ab2342a02ad632c5d9787d931a4836
3
  size 650646
SnowballTarget/SnowballTarget-500104.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fac9edea51a6ffbca9c06396b98b2472864069fca4a32d384f8ec77084bb1540
3
  size 3850075
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a7bf322007dcb130833821df591acf30bb4131bd816a6d787d4fd7892ffd035
3
  size 3850075
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a267e31e2ac3856c2d332fc83bae028e2f0b90dab6baad216d09657f6128e49
3
  size 3849250
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d685c74ff567cd65a6d9d488f1dc4d5b28bc7ea2fbc4850b9d637dbcf83eff39
3
  size 3849250
SnowballTarget/events.out.tfevents.1714562245.b99f21e3afc2.3104.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c604971259ae79fbe9afa9a6a53ab0564daf8332cb744f7f6011cb21e22665b2
3
+ size 48860
run_logs/Player-0.log CHANGED
@@ -45,7 +45,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
- UnloadTime: 0.687360 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 1 worker threads for Enlighten.
@@ -169,16 +169,16 @@ Memory Statistics:
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
- Peak main deferred allocation count 44
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
- Peak usage frame count: [4.0 MB-8.0 MB]: 12524 frames, [8.0 MB-16.0 MB]: 1118 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
- Peak Allocated memory 9.6 MB
182
  Peak Large allocation bytes 0 B
183
  [ALLOC_DEFAULT_THREAD]
184
  Peak usage frame count: [16.0 MB-32.0 MB]: 13642 frames
@@ -213,10 +213,10 @@ Memory Statistics:
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
- Peak usage frame count: [32.0 KB-64.0 KB]: 12810 frames, [64.0 KB-128.0 KB]: 832 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
- Peak Allocated memory 66.8 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
  Peak usage frame count: [32.0 KB-64.0 KB]: 13642 frames
 
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
+ UnloadTime: 0.676908 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 1 worker threads for Enlighten.
 
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
+ Peak main deferred allocation count 37
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
+ Peak usage frame count: [4.0 MB-8.0 MB]: 12065 frames, [8.0 MB-16.0 MB]: 1577 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
+ Peak Allocated memory 9.5 MB
182
  Peak Large allocation bytes 0 B
183
  [ALLOC_DEFAULT_THREAD]
184
  Peak usage frame count: [16.0 MB-32.0 MB]: 13642 frames
 
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
+ Peak usage frame count: [32.0 KB-64.0 KB]: 13046 frames, [64.0 KB-128.0 KB]: 596 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
+ Peak Allocated memory 66.4 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
  Peak usage frame count: [32.0 KB-64.0 KB]: 13642 frames
run_logs/timers.json CHANGED
@@ -2,15 +2,15 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.6708531379699707,
6
- "min": 0.6675255298614502,
7
- "max": 2.86983585357666,
8
  "count": 50
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 6479.099609375,
12
- "min": 6344.16259765625,
13
- "max": 29389.98828125,
14
  "count": 50
15
  },
16
  "SnowballTarget.Step.mean": {
@@ -26,15 +26,15 @@
26
  "count": 50
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 13.789155006408691,
30
- "min": 0.3778553605079651,
31
- "max": 13.972408294677734,
32
  "count": 50
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2826.77685546875,
36
- "min": 73.30393981933594,
37
- "max": 2852.46044921875,
38
  "count": 50
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
@@ -50,87 +50,87 @@
50
  "count": 50
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.0648688686200687,
54
- "min": 0.05976627333243681,
55
- "max": 0.07717936485149789,
56
  "count": 50
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.3243443431003435,
60
- "min": 0.23906509332974724,
61
- "max": 0.37300145060276807,
62
  "count": 50
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.17004113148021346,
66
- "min": 0.10590756960971026,
67
- "max": 0.2807742419829166,
68
  "count": 50
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.8502056574010673,
72
- "min": 0.42363027843884105,
73
- "max": 1.400986860780155,
74
  "count": 50
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
- "value": 3.0605789798400035e-06,
78
- "min": 3.0605789798400035e-06,
79
  "max": 0.00029675280108239997,
80
  "count": 50
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
83
- "value": 1.5302894899200018e-05,
84
- "min": 1.5302894899200018e-05,
85
  "max": 0.001454064015312,
86
  "count": 50
87
  },
88
  "SnowballTarget.Policy.Epsilon.mean": {
89
- "value": 0.10102016,
90
- "min": 0.10102016,
91
  "max": 0.19891759999999997,
92
  "count": 50
93
  },
94
  "SnowballTarget.Policy.Epsilon.sum": {
95
- "value": 0.5051008,
96
  "min": 0.41199040000000003,
97
  "max": 0.984688,
98
  "count": 50
99
  },
100
  "SnowballTarget.Policy.Beta.mean": {
101
- "value": 6.090598400000005e-05,
102
- "min": 6.090598400000005e-05,
103
  "max": 0.00494598824,
104
  "count": 50
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
107
- "value": 0.0003045299200000003,
108
- "min": 0.0003045299200000003,
109
  "max": 0.0242359312,
110
  "count": 50
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 26.381818181818183,
114
- "min": 3.25,
115
- "max": 27.613636363636363,
116
  "count": 50
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1451.0,
120
- "min": 143.0,
121
- "max": 1497.0,
122
  "count": 50
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 26.381818181818183,
126
- "min": 3.25,
127
- "max": 27.613636363636363,
128
  "count": 50
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1451.0,
132
- "min": 143.0,
133
- "max": 1497.0,
134
  "count": 50
135
  },
136
  "SnowballTarget.IsTraining.mean": {
@@ -148,7 +148,7 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1713500670",
152
  "python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "1.1.0.dev0",
@@ -156,59 +156,59 @@
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.2.1+cu121",
158
  "numpy_version": "1.23.5",
159
- "end_time_seconds": "1713501901"
160
  },
161
- "total": 1231.0941466509998,
162
  "count": 1,
163
- "self": 0.496644072999743,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.09866140200006157,
167
  "count": 1,
168
- "self": 0.09866140200006157
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 1230.498841176,
172
  "count": 1,
173
- "self": 1.6710581750110123,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 2.8385041750000255,
177
  "count": 1,
178
- "self": 2.8385041750000255
179
  },
180
  "TrainerController.advance": {
181
- "total": 1225.895372972989,
182
- "count": 45472,
183
- "self": 0.8477747759650356,
184
  "children": {
185
  "env_step": {
186
- "total": 1225.047598197024,
187
- "count": 45472,
188
- "self": 800.8792869380989,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 423.3421535989627,
192
- "count": 45472,
193
- "self": 4.093057696959363,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 419.2490959020033,
197
- "count": 45472,
198
- "self": 419.2490959020033
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.826157659962405,
204
- "count": 45472,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 1227.131502106949,
209
- "count": 45472,
210
  "is_parallel": true,
211
- "self": 606.1230222958941,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +217,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.005343885999991471,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.003563574000054359,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.001780311999937112,
227
  "count": 10,
228
  "is_parallel": true,
229
- "self": 0.001780311999937112
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.046734046999972634,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0007847460000220963,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.00045010199994521827,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.00045010199994521827
244
  },
245
  "communicator.exchange": {
246
- "total": 0.0432126010000502,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.0432126010000502
250
  },
251
  "steps_from_proto": {
252
- "total": 0.00228659799995512,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.00043842299987773004,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.00184817500007739,
259
  "count": 10,
260
  "is_parallel": true,
261
- "self": 0.00184817500007739
262
  }
263
  }
264
  }
@@ -267,34 +267,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 621.0084798110549,
271
- "count": 45471,
272
  "is_parallel": true,
273
- "self": 27.92437867812282,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 15.227463315993646,
277
- "count": 45471,
278
  "is_parallel": true,
279
- "self": 15.227463315993646
280
  },
281
  "communicator.exchange": {
282
- "total": 482.87709400796905,
283
- "count": 45471,
284
  "is_parallel": true,
285
- "self": 482.87709400796905
286
  },
287
  "steps_from_proto": {
288
- "total": 94.97954380896942,
289
- "count": 45471,
290
  "is_parallel": true,
291
- "self": 18.201260577935955,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 76.77828323103347,
295
- "count": 454710,
296
  "is_parallel": true,
297
- "self": 76.77828323103347
298
  }
299
  }
300
  }
@@ -309,9 +309,9 @@
309
  }
310
  },
311
  "trainer_threads": {
312
- "total": 0.00016404999996666447,
313
  "count": 1,
314
- "self": 0.00016404999996666447,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
@@ -320,36 +320,36 @@
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
- "total": 1210.07174479809,
324
- "count": 1839513,
325
  "is_parallel": true,
326
- "self": 40.65298581313732,
327
  "children": {
328
  "process_trajectory": {
329
- "total": 666.971226824954,
330
- "count": 1839513,
331
  "is_parallel": true,
332
- "self": 664.9093375149539,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
- "total": 2.0618893100000832,
336
  "count": 11,
337
  "is_parallel": true,
338
- "self": 2.0618893100000832
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
- "total": 502.4475321599988,
344
  "count": 227,
345
  "is_parallel": true,
346
- "self": 148.18481411099356,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
- "total": 354.26271804900523,
350
- "count": 11571,
351
  "is_parallel": true,
352
- "self": 354.26271804900523
353
  }
354
  }
355
  }
@@ -360,14 +360,14 @@
360
  }
361
  },
362
  "TrainerController._save_models": {
363
- "total": 0.09374180300005719,
364
  "count": 1,
365
- "self": 0.0010508670000035636,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
- "total": 0.09269093600005363,
369
  "count": 1,
370
- "self": 0.09269093600005363
371
  }
372
  }
373
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.7754722833633423,
6
+ "min": 0.7657526135444641,
7
+ "max": 2.8711438179016113,
8
  "count": 50
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 7506.57177734375,
12
+ "min": 7277.712890625,
13
+ "max": 29403.3828125,
14
  "count": 50
15
  },
16
  "SnowballTarget.Step.mean": {
 
26
  "count": 50
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 13.606134414672852,
30
+ "min": 0.3890088200569153,
31
+ "max": 13.722529411315918,
32
  "count": 50
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 2789.257568359375,
36
+ "min": 75.46771240234375,
37
+ "max": 2799.39599609375,
38
  "count": 50
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
 
50
  "count": 50
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
+ "value": 0.06538191351494022,
54
+ "min": 0.06220048081555504,
55
+ "max": 0.07803483618886722,
56
  "count": 50
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
+ "value": 0.3269095675747011,
60
+ "min": 0.24880192326222017,
61
+ "max": 0.36883155694322695,
62
  "count": 50
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
+ "value": 0.18201290512201834,
66
+ "min": 0.11943581322203482,
67
+ "max": 0.2981918522101991,
68
  "count": 50
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
+ "value": 0.9100645256100917,
72
+ "min": 0.4777432528881393,
73
+ "max": 1.3902910865989386,
74
  "count": 50
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
+ "value": 3.0528989824000028e-06,
78
+ "min": 3.0528989824000028e-06,
79
  "max": 0.00029675280108239997,
80
  "count": 50
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
83
+ "value": 1.5264494912000015e-05,
84
+ "min": 1.5264494912000015e-05,
85
  "max": 0.001454064015312,
86
  "count": 50
87
  },
88
  "SnowballTarget.Policy.Epsilon.mean": {
89
+ "value": 0.10101760000000001,
90
+ "min": 0.10101760000000001,
91
  "max": 0.19891759999999997,
92
  "count": 50
93
  },
94
  "SnowballTarget.Policy.Epsilon.sum": {
95
+ "value": 0.5050880000000001,
96
  "min": 0.41199040000000003,
97
  "max": 0.984688,
98
  "count": 50
99
  },
100
  "SnowballTarget.Policy.Beta.mean": {
101
+ "value": 6.0778240000000044e-05,
102
+ "min": 6.0778240000000044e-05,
103
  "max": 0.00494598824,
104
  "count": 50
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
107
+ "value": 0.0003038912000000002,
108
+ "min": 0.0003038912000000002,
109
  "max": 0.0242359312,
110
  "count": 50
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
+ "value": 26.254545454545454,
114
+ "min": 3.4318181818181817,
115
+ "max": 27.145454545454545,
116
  "count": 50
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
+ "value": 1444.0,
120
+ "min": 151.0,
121
+ "max": 1493.0,
122
  "count": 50
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
+ "value": 26.254545454545454,
126
+ "min": 3.4318181818181817,
127
+ "max": 27.145454545454545,
128
  "count": 50
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
+ "value": 1444.0,
132
+ "min": 151.0,
133
+ "max": 1493.0,
134
  "count": 50
135
  },
136
  "SnowballTarget.IsTraining.mean": {
 
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1714562244",
152
  "python_version": "3.10.12 (main, Nov 20 2023, 15:14:05) [GCC 11.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "1.1.0.dev0",
 
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "2.2.1+cu121",
158
  "numpy_version": "1.23.5",
159
+ "end_time_seconds": "1714563411"
160
  },
161
+ "total": 1166.6966391229998,
162
  "count": 1,
163
+ "self": 0.44238212100003693,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.05772473499996522,
167
  "count": 1,
168
+ "self": 0.05772473499996522
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 1166.1965322669998,
172
  "count": 1,
173
+ "self": 1.394253988980381,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 2.7503450049999856,
177
  "count": 1,
178
+ "self": 2.7503450049999856
179
  },
180
  "TrainerController.advance": {
181
+ "total": 1161.9619085960192,
182
+ "count": 45473,
183
+ "self": 0.7203130979917205,
184
  "children": {
185
  "env_step": {
186
+ "total": 1161.2415954980274,
187
+ "count": 45473,
188
+ "self": 752.3691589549815,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 408.1283050980237,
192
+ "count": 45473,
193
+ "self": 4.090393263988972,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 404.03791183403473,
197
+ "count": 45473,
198
+ "self": 404.03791183403473
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 0.7441314450222762,
204
+ "count": 45473,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 1163.14613647302,
209
+ "count": 45473,
210
  "is_parallel": true,
211
+ "self": 579.8086245110172,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.005242626000040218,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.003626690000146482,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.0016159359998937362,
227
  "count": 10,
228
  "is_parallel": true,
229
+ "self": 0.0016159359998937362
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.04481413300004533,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.0021169469999904322,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 0.000408078000077694,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 0.000408078000077694
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.04011978699998053,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.04011978699998053
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.002169320999996671,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 0.00043197400009375997,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.0017373469999029112,
259
  "count": 10,
260
  "is_parallel": true,
261
+ "self": 0.0017373469999029112
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 583.3375119620028,
271
+ "count": 45472,
272
  "is_parallel": true,
273
+ "self": 26.57459214304015,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 14.553109399001073,
277
+ "count": 45472,
278
  "is_parallel": true,
279
+ "self": 14.553109399001073
280
  },
281
  "communicator.exchange": {
282
+ "total": 451.7761595199755,
283
+ "count": 45472,
284
  "is_parallel": true,
285
+ "self": 451.7761595199755
286
  },
287
  "steps_from_proto": {
288
+ "total": 90.4336508999861,
289
+ "count": 45472,
290
  "is_parallel": true,
291
+ "self": 16.97052263493765,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 73.46312826504845,
295
+ "count": 454720,
296
  "is_parallel": true,
297
+ "self": 73.46312826504845
298
  }
299
  }
300
  }
 
309
  }
310
  },
311
  "trainer_threads": {
312
+ "total": 0.00014010500012773264,
313
  "count": 1,
314
+ "self": 0.00014010500012773264,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
 
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
+ "total": 1146.897774661027,
324
+ "count": 1760735,
325
  "is_parallel": true,
326
+ "self": 38.388327626931186,
327
  "children": {
328
  "process_trajectory": {
329
+ "total": 632.6697963490964,
330
+ "count": 1760735,
331
  "is_parallel": true,
332
+ "self": 631.2122672960967,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
+ "total": 1.4575290529996892,
336
  "count": 11,
337
  "is_parallel": true,
338
+ "self": 1.4575290529996892
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
+ "total": 475.83965068499936,
344
  "count": 227,
345
  "is_parallel": true,
346
+ "self": 139.78062287798855,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
+ "total": 336.0590278070108,
350
+ "count": 11574,
351
  "is_parallel": true,
352
+ "self": 336.0590278070108
353
  }
354
  }
355
  }
 
360
  }
361
  },
362
  "TrainerController._save_models": {
363
+ "total": 0.0898845720000736,
364
  "count": 1,
365
+ "self": 0.0009937120000813593,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
+ "total": 0.08889085999999224,
369
  "count": 1,
370
+ "self": 0.08889085999999224
371
  }
372
  }
373
  }
run_logs/training_status.json CHANGED
@@ -4,8 +4,8 @@
4
  {
5
  "steps": 134968,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-134968.onnx",
7
- "reward": 23.818181818181817,
8
- "creation_time": 1713501015.800465,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-134968.pt"
11
  ]
@@ -13,8 +13,8 @@
13
  {
14
  "steps": 179992,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-179992.onnx",
16
- "reward": 25.272727272727273,
17
- "creation_time": 1713501123.9698143,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-179992.pt"
20
  ]
@@ -22,8 +22,8 @@
22
  {
23
  "steps": 224976,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-224976.onnx",
25
- "reward": 26.181818181818183,
26
- "creation_time": 1713501234.9495454,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-224976.pt"
29
  ]
@@ -31,8 +31,8 @@
31
  {
32
  "steps": 269936,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-269936.onnx",
34
- "reward": 26.363636363636363,
35
- "creation_time": 1713501343.62578,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-269936.pt"
38
  ]
@@ -40,8 +40,8 @@
40
  {
41
  "steps": 314984,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-314984.onnx",
43
- "reward": 26.09090909090909,
44
- "creation_time": 1713501453.1664703,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-314984.pt"
47
  ]
@@ -49,8 +49,8 @@
49
  {
50
  "steps": 359944,
51
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-359944.onnx",
52
- "reward": 26.636363636363637,
53
- "creation_time": 1713501558.2743406,
54
  "auxillary_file_paths": [
55
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-359944.pt"
56
  ]
@@ -58,8 +58,8 @@
58
  {
59
  "steps": 404992,
60
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-404992.onnx",
61
- "reward": 27.0,
62
- "creation_time": 1713501670.4487193,
63
  "auxillary_file_paths": [
64
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-404992.pt"
65
  ]
@@ -67,8 +67,8 @@
67
  {
68
  "steps": 449952,
69
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.onnx",
70
- "reward": 27.363636363636363,
71
- "creation_time": 1713501777.8827279,
72
  "auxillary_file_paths": [
73
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.pt"
74
  ]
@@ -76,8 +76,8 @@
76
  {
77
  "steps": 494992,
78
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-494992.onnx",
79
- "reward": 26.2,
80
- "creation_time": 1713501889.9007506,
81
  "auxillary_file_paths": [
82
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-494992.pt"
83
  ]
@@ -85,8 +85,8 @@
85
  {
86
  "steps": 500104,
87
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.onnx",
88
- "reward": 25.727272727272727,
89
- "creation_time": 1713501900.8817897,
90
  "auxillary_file_paths": [
91
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
92
  ]
@@ -95,8 +95,8 @@
95
  "final_checkpoint": {
96
  "steps": 500104,
97
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
98
- "reward": 25.727272727272727,
99
- "creation_time": 1713501900.8817897,
100
  "auxillary_file_paths": [
101
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
102
  ]
 
4
  {
5
  "steps": 134968,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-134968.onnx",
7
+ "reward": 24.727272727272727,
8
+ "creation_time": 1714562565.9412103,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-134968.pt"
11
  ]
 
13
  {
14
  "steps": 179992,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-179992.onnx",
16
+ "reward": 26.272727272727273,
17
+ "creation_time": 1714562670.2212827,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-179992.pt"
20
  ]
 
22
  {
23
  "steps": 224976,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-224976.onnx",
25
+ "reward": 28.0,
26
+ "creation_time": 1714562774.1750045,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-224976.pt"
29
  ]
 
31
  {
32
  "steps": 269936,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-269936.onnx",
34
+ "reward": 26.181818181818183,
35
+ "creation_time": 1714562878.4955747,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-269936.pt"
38
  ]
 
40
  {
41
  "steps": 314984,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-314984.onnx",
43
+ "reward": 25.818181818181817,
44
+ "creation_time": 1714562983.0768056,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-314984.pt"
47
  ]
 
49
  {
50
  "steps": 359944,
51
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-359944.onnx",
52
+ "reward": 27.363636363636363,
53
+ "creation_time": 1714563085.2067554,
54
  "auxillary_file_paths": [
55
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-359944.pt"
56
  ]
 
58
  {
59
  "steps": 404992,
60
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-404992.onnx",
61
+ "reward": 26.636363636363637,
62
+ "creation_time": 1714563190.7871082,
63
  "auxillary_file_paths": [
64
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-404992.pt"
65
  ]
 
67
  {
68
  "steps": 449952,
69
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.onnx",
70
+ "reward": 26.636363636363637,
71
+ "creation_time": 1714563293.107314,
72
  "auxillary_file_paths": [
73
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.pt"
74
  ]
 
76
  {
77
  "steps": 494992,
78
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-494992.onnx",
79
+ "reward": 25.8,
80
+ "creation_time": 1714563399.821591,
81
  "auxillary_file_paths": [
82
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-494992.pt"
83
  ]
 
85
  {
86
  "steps": 500104,
87
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.onnx",
88
+ "reward": 25.636363636363637,
89
+ "creation_time": 1714563410.6735628,
90
  "auxillary_file_paths": [
91
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
92
  ]
 
95
  "final_checkpoint": {
96
  "steps": 500104,
97
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
98
+ "reward": 25.636363636363637,
99
+ "creation_time": 1714563410.6735628,
100
  "auxillary_file_paths": [
101
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
102
  ]