kavindumit commited on
Commit
bb93611
·
verified ·
1 Parent(s): 2666f40

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -8,13 +8,22 @@ tags:
8
  ---
9
 
10
  # **ppo** Agent playing **SnowballTarget**
11
- This is a trained model of a **ppo** agent playing **SnowballTarget** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
12
 
 
 
 
 
 
 
 
 
 
13
  ### Watch your Agent play
14
  You can watch your agent **playing directly in your browser:**.
15
 
16
- 1. Go to https://singularite.itch.io/snowballtarget
17
- 2. Step 1: Find your model_id: Mikepool117/ML-Agents-SnowballTarget
18
  3. Step 2: Select your *.nn /*.onnx file
19
  4. Click on Watch the agent play 👀
20
 
 
8
  ---
9
 
10
  # **ppo** Agent playing **SnowballTarget**
11
+ This is a trained model of a **ppo** agent playing **SnowballTarget** using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
12
 
13
+ ## Usage (with ML-Agents)
14
+ The Documentation: https://github.com/huggingface/ml-agents#get-started
15
+ We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
16
+
17
+
18
+ ### Resume the training
19
+ ```
20
+ mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
21
+ ```
22
  ### Watch your Agent play
23
  You can watch your agent **playing directly in your browser:**.
24
 
25
+ 1. Go to https://huggingface.co/spaces/unity/ML-Agents-SnowballTarget
26
+ 2. Step 1: Find your model_id: azetaaa/ppo-ML-Agents-SnowballTarget
27
  3. Step 2: Select your *.nn /*.onnx file
28
  4. Click on Watch the agent play 👀
29
 
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19f302df594438d1e73e37f3b26af3d4aad130a13d91c2403c3de448e41b4bd4
3
  size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41a42a6cca3b7a37202046fc0916b440d0fbb0a1024428825b39009e8a22bfe8
3
  size 646532
SnowballTarget/SnowballTarget-149984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12e87cf9e0578f0d459a120c0cf3650f9aea1178b6c8a60b0518df7e51c4dd68
3
  size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e18815bc26375f04944a981d6fbfd070c60537f1ff40aada4e2f6008e793b8e
3
  size 646532
SnowballTarget/SnowballTarget-149984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d2bfd0358bc283b1a790dbc144d4c22f7bb103c35a2329b8dd4aa93f9d623a3
3
  size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27d2d593292f1295102ea2324ce77e55b977fc0e9b709b9b29795d41d6217b19
3
  size 3843968
SnowballTarget/SnowballTarget-199984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19f302df594438d1e73e37f3b26af3d4aad130a13d91c2403c3de448e41b4bd4
3
  size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41a42a6cca3b7a37202046fc0916b440d0fbb0a1024428825b39009e8a22bfe8
3
  size 646532
SnowballTarget/SnowballTarget-199984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1754461556406f77735b2538fb79dabd273a28a287a0950de1289b0ee98a8f9
3
  size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef28be86220c1dd5ca1acd76f8a1f69ef03d7568938a2b797b6438c58060f522
3
  size 3843968
SnowballTarget/SnowballTarget-200112.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19f302df594438d1e73e37f3b26af3d4aad130a13d91c2403c3de448e41b4bd4
3
  size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41a42a6cca3b7a37202046fc0916b440d0fbb0a1024428825b39009e8a22bfe8
3
  size 646532
SnowballTarget/SnowballTarget-200112.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a88a69e3a61c88d2302c3ffb456b52454501f08d1dff41cfcba43dcb408b6c2
3
  size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d96509001e6dbfba9ba5c221e18b1a18f0e69dfed05fcff14c5fc0e1cbe994da
3
  size 3843968
SnowballTarget/SnowballTarget-49936.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9777695108c54a342cedc32bd00b6baadbbc5d914610b8f4f0128c8926d4ad6e
3
  size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c1de8bbacfbd88d21ed29f3e35f698191d9e999c3e66ab855e7dd773b5d4051
3
  size 646532
SnowballTarget/SnowballTarget-49936.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa14035541b64d2007dc28238444459f32ceaa155b66449c4f15e623dae50522
3
  size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef920958d091c2619be5db1d0ebfec7ea0e0808d7d539383791b2b2cd65390c
3
  size 3843968
SnowballTarget/SnowballTarget-99960.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:424e95110136b37331dd3fe2d5ae3b6dd3d4ec2bd5b67afdea4998303ae0686e
3
  size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2404ab0d51f51beb8b07642e38f6ce6ba25b460a8e8fe1a2f9c65475dbb6a2a
3
  size 646532
SnowballTarget/SnowballTarget-99960.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3f44f4b24fd37a243ce078022ed4813502e307965081712fe3b4cf43f2cc978c
3
  size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa6dc19862338ae3dbafb23e51b22cd14faa9532c6aff5919ef0ee03abdfd706
3
  size 3843968
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4a88a69e3a61c88d2302c3ffb456b52454501f08d1dff41cfcba43dcb408b6c2
3
  size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d96509001e6dbfba9ba5c221e18b1a18f0e69dfed05fcff14c5fc0e1cbe994da
3
  size 3843968
SnowballTarget/events.out.tfevents.1684870816.6d4d96298542.1875.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a795f9f28d3d4c96082efda15dbc069e7b13d6eaa8d1f75ad7a6a6c00faff38
3
+ size 28575
run_logs/Player-0.log CHANGED
@@ -34,7 +34,7 @@ ALSA lib pcm.c:2642:(snd_pcm_open_noupdate) Unknown PCM default
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
- - Completed reload, in 0.085 seconds
38
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -45,7 +45,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
- UnloadTime: 0.649668 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 1 worker threads for Enlighten.
@@ -53,7 +53,7 @@ Memory Statistics:
53
  [ALLOC_TEMP_TLS] TLS Allocator
54
  StackAllocators :
55
  [ALLOC_TEMP_MAIN]
56
- Peak usage frame count: [8.0 KB-16.0 KB]: 5368 frames, [16.0 KB-32.0 KB]: 91 frames, [2.0 MB-4.0 MB]: 1 frames
57
  Initial Block Size 4.0 MB
58
  Current Block Size 4.0 MB
59
  Peak Allocated Bytes 2.0 MB
@@ -169,19 +169,19 @@ Memory Statistics:
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
- Peak main deferred allocation count 41
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
- Peak usage frame count: [4.0 MB-8.0 MB]: 5460 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
- Peak Allocated memory 6.6 MB
182
  Peak Large allocation bytes 0 B
183
  [ALLOC_DEFAULT_THREAD]
184
- Peak usage frame count: [16.0 MB-32.0 MB]: 5460 frames
185
  Requested Block Size 16.0 MB
186
  Peak Block count 1
187
  Peak Allocated memory 17.8 MB
@@ -213,13 +213,13 @@ Memory Statistics:
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
- Peak usage frame count: [32.0 KB-64.0 KB]: 4181 frames, [64.0 KB-128.0 KB]: 1279 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
- Peak Allocated memory 67.3 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
- Peak usage frame count: [32.0 KB-64.0 KB]: 5460 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 1
225
  Peak Allocated memory 39.6 KB
@@ -231,13 +231,13 @@ Memory Statistics:
231
  Used Block count 1
232
  Peak Allocated bytes 1.0 MB
233
  [ALLOC_CACHEOBJECTS_MAIN]
234
- Peak usage frame count: [0.5 MB-1.0 MB]: 5460 frames
235
  Requested Block Size 4.0 MB
236
  Peak Block count 1
237
  Peak Allocated memory 0.6 MB
238
  Peak Large allocation bytes 0 B
239
  [ALLOC_CACHEOBJECTS_THREAD]
240
- Peak usage frame count: [0.5 MB-1.0 MB]: 5459 frames, [2.0 MB-4.0 MB]: 1 frames
241
  Requested Block Size 4.0 MB
242
  Peak Block count 1
243
  Peak Allocated memory 2.2 MB
@@ -249,13 +249,13 @@ Memory Statistics:
249
  Used Block count 1
250
  Peak Allocated bytes 1.0 MB
251
  [ALLOC_TYPETREE_MAIN]
252
- Peak usage frame count: [0-1.0 KB]: 5460 frames
253
  Requested Block Size 2.0 MB
254
  Peak Block count 1
255
  Peak Allocated memory 1.0 KB
256
  Peak Large allocation bytes 0 B
257
  [ALLOC_TYPETREE_THREAD]
258
- Peak usage frame count: [1.0 KB-2.0 KB]: 5460 frames
259
  Requested Block Size 2.0 MB
260
  Peak Block count 1
261
  Peak Allocated memory 1.7 KB
 
34
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
35
  FMOD initialized on nosound output
36
  Begin MonoManager ReloadAssembly
37
+ - Completed reload, in 0.082 seconds
38
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
39
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
40
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
45
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
46
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
47
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
48
+ UnloadTime: 0.661584 ms
49
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
50
  requesting resize 84 x 84
51
  Setting up 1 worker threads for Enlighten.
 
53
  [ALLOC_TEMP_TLS] TLS Allocator
54
  StackAllocators :
55
  [ALLOC_TEMP_MAIN]
56
+ Peak usage frame count: [8.0 KB-16.0 KB]: 5374 frames, [16.0 KB-32.0 KB]: 91 frames, [2.0 MB-4.0 MB]: 1 frames
57
  Initial Block Size 4.0 MB
58
  Current Block Size 4.0 MB
59
  Peak Allocated Bytes 2.0 MB
 
169
  Peak Allocated Bytes 0 B
170
  Overflow Count 0
171
  [ALLOC_DEFAULT] Dual Thread Allocator
172
+ Peak main deferred allocation count 40
173
  [ALLOC_BUCKET]
174
  Large Block size 4.0 MB
175
  Used Block count 1
176
  Peak Allocated bytes 1.0 MB
177
  [ALLOC_DEFAULT_MAIN]
178
+ Peak usage frame count: [4.0 MB-8.0 MB]: 5466 frames
179
  Requested Block Size 16.0 MB
180
  Peak Block count 1
181
+ Peak Allocated memory 6.8 MB
182
  Peak Large allocation bytes 0 B
183
  [ALLOC_DEFAULT_THREAD]
184
+ Peak usage frame count: [16.0 MB-32.0 MB]: 5466 frames
185
  Requested Block Size 16.0 MB
186
  Peak Block count 1
187
  Peak Allocated memory 17.8 MB
 
213
  Used Block count 1
214
  Peak Allocated bytes 1.0 MB
215
  [ALLOC_GFX_MAIN]
216
+ Peak usage frame count: [32.0 KB-64.0 KB]: 4650 frames, [64.0 KB-128.0 KB]: 816 frames
217
  Requested Block Size 16.0 MB
218
  Peak Block count 1
219
+ Peak Allocated memory 67.7 KB
220
  Peak Large allocation bytes 0 B
221
  [ALLOC_GFX_THREAD]
222
+ Peak usage frame count: [32.0 KB-64.0 KB]: 5466 frames
223
  Requested Block Size 16.0 MB
224
  Peak Block count 1
225
  Peak Allocated memory 39.6 KB
 
231
  Used Block count 1
232
  Peak Allocated bytes 1.0 MB
233
  [ALLOC_CACHEOBJECTS_MAIN]
234
+ Peak usage frame count: [0.5 MB-1.0 MB]: 5466 frames
235
  Requested Block Size 4.0 MB
236
  Peak Block count 1
237
  Peak Allocated memory 0.6 MB
238
  Peak Large allocation bytes 0 B
239
  [ALLOC_CACHEOBJECTS_THREAD]
240
+ Peak usage frame count: [0.5 MB-1.0 MB]: 5465 frames, [2.0 MB-4.0 MB]: 1 frames
241
  Requested Block Size 4.0 MB
242
  Peak Block count 1
243
  Peak Allocated memory 2.2 MB
 
249
  Used Block count 1
250
  Peak Allocated bytes 1.0 MB
251
  [ALLOC_TYPETREE_MAIN]
252
+ Peak usage frame count: [0-1.0 KB]: 5466 frames
253
  Requested Block Size 2.0 MB
254
  Peak Block count 1
255
  Peak Allocated memory 1.0 KB
256
  Peak Large allocation bytes 0 B
257
  [ALLOC_TYPETREE_THREAD]
258
+ Peak usage frame count: [1.0 KB-2.0 KB]: 5466 frames
259
  Requested Block Size 2.0 MB
260
  Peak Block count 1
261
  Peak Allocated memory 1.7 KB
run_logs/timers.json CHANGED
@@ -2,15 +2,15 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 1.1146881580352783,
6
- "min": 1.1124016046524048,
7
- "max": 2.8580334186553955,
8
  "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 10643.0419921875,
12
- "min": 10643.0419921875,
13
- "max": 29269.12109375,
14
  "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
@@ -26,15 +26,15 @@
26
  "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 12.643514633178711,
30
- "min": 0.3894132673740387,
31
- "max": 12.643514633178711,
32
  "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2465.4853515625,
36
- "min": 75.54617309570312,
37
- "max": 2534.64599609375,
38
  "count": 20
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
@@ -50,27 +50,27 @@
50
  "count": 20
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.06619036073600143,
54
- "min": 0.06052785818796039,
55
- "max": 0.0764143095554127,
56
  "count": 20
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.2647614429440057,
60
- "min": 0.24211143275184155,
61
- "max": 0.38207154777706354,
62
  "count": 20
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.20453253808413066,
66
- "min": 0.11950370855172915,
67
- "max": 0.27864555617173514,
68
  "count": 20
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.8181301523365226,
72
- "min": 0.4780148342069166,
73
- "max": 1.3932277808586757,
74
  "count": 20
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
@@ -110,27 +110,27 @@
110
  "count": 20
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 25.045454545454547,
114
- "min": 3.3863636363636362,
115
- "max": 25.045454545454547,
116
  "count": 20
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1102.0,
120
- "min": 149.0,
121
- "max": 1370.0,
122
  "count": 20
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 25.045454545454547,
126
- "min": 3.3863636363636362,
127
- "max": 25.045454545454547,
128
  "count": 20
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1102.0,
132
- "min": 149.0,
133
- "max": 1370.0,
134
  "count": 20
135
  },
136
  "SnowballTarget.IsTraining.mean": {
@@ -148,7 +148,7 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1684300069",
152
  "python_version": "3.10.11 (main, Apr 5 2023, 14:15:10) [GCC 9.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "0.31.0.dev0",
@@ -156,59 +156,59 @@
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.11.0+cu102",
158
  "numpy_version": "1.21.2",
159
- "end_time_seconds": "1684300517"
160
  },
161
- "total": 448.22644002499993,
162
  "count": 1,
163
- "self": 0.7966314149999221,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.042010804000028656,
167
  "count": 1,
168
- "self": 0.042010804000028656
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 447.387797806,
172
  "count": 1,
173
- "self": 0.5210835430021916,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 3.7907317779999516,
177
  "count": 1,
178
- "self": 3.7907317779999516
179
  },
180
  "TrainerController.advance": {
181
- "total": 442.9101492269978,
182
- "count": 18200,
183
- "self": 0.25095092499623206,
184
  "children": {
185
  "env_step": {
186
- "total": 442.6591983020016,
187
- "count": 18200,
188
- "self": 323.35721580101284,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 119.0468231949938,
192
- "count": 18200,
193
- "self": 1.8031842269994058,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 117.2436389679944,
197
- "count": 18200,
198
- "self": 117.2436389679944
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.25515930599493686,
204
- "count": 18200,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 445.91512152999235,
209
- "count": 18200,
210
  "is_parallel": true,
211
- "self": 207.7106967119982,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +217,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.005656584999996994,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.004313371999955962,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0013432130000410325,
227
  "count": 10,
228
  "is_parallel": true,
229
- "self": 0.0013432130000410325
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.044221881999988,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0005668419999551588,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.00031216100001074665,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.00031216100001074665
244
  },
245
  "communicator.exchange": {
246
- "total": 0.0414147660000026,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.0414147660000026
250
  },
251
  "steps_from_proto": {
252
- "total": 0.0019281130000194935,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.0003374430001485962,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.0015906699998708973,
259
  "count": 10,
260
  "is_parallel": true,
261
- "self": 0.0015906699998708973
262
  }
263
  }
264
  }
@@ -267,34 +267,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 238.20442481799415,
271
- "count": 18199,
272
  "is_parallel": true,
273
- "self": 9.762043927981097,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 5.056561690001104,
277
- "count": 18199,
278
  "is_parallel": true,
279
- "self": 5.056561690001104
280
  },
281
  "communicator.exchange": {
282
- "total": 190.71931240100588,
283
- "count": 18199,
284
  "is_parallel": true,
285
- "self": 190.71931240100588
286
  },
287
  "steps_from_proto": {
288
- "total": 32.66650679900607,
289
- "count": 18199,
290
  "is_parallel": true,
291
- "self": 5.837936313986461,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 26.82857048501961,
295
- "count": 181990,
296
  "is_parallel": true,
297
- "self": 26.82857048501961
298
  }
299
  }
300
  }
@@ -309,9 +309,9 @@
309
  }
310
  },
311
  "trainer_threads": {
312
- "total": 4.2783000026247464e-05,
313
  "count": 1,
314
- "self": 4.2783000026247464e-05,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
@@ -320,36 +320,36 @@
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
- "total": 439.6015811119803,
324
- "count": 411394,
325
  "is_parallel": true,
326
- "self": 9.027218850983218,
327
  "children": {
328
  "process_trajectory": {
329
- "total": 239.9014348819971,
330
- "count": 411394,
331
  "is_parallel": true,
332
- "self": 238.93572218399714,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
- "total": 0.9657126979999475,
336
  "count": 4,
337
  "is_parallel": true,
338
- "self": 0.9657126979999475
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
- "total": 190.672927379,
344
  "count": 90,
345
  "is_parallel": true,
346
- "self": 72.33409942399987,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
- "total": 118.33882795500011,
350
- "count": 4584,
351
  "is_parallel": true,
352
- "self": 118.33882795500011
353
  }
354
  }
355
  }
@@ -360,14 +360,14 @@
360
  }
361
  },
362
  "TrainerController._save_models": {
363
- "total": 0.16579047499999433,
364
  "count": 1,
365
- "self": 0.0011322039999868139,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
- "total": 0.16465827100000752,
369
  "count": 1,
370
- "self": 0.16465827100000752
371
  }
372
  }
373
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.9898553490638733,
6
+ "min": 0.9872123599052429,
7
+ "max": 2.8721883296966553,
8
  "count": 20
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 9451.138671875,
12
+ "min": 9451.138671875,
13
+ "max": 29445.67578125,
14
  "count": 20
15
  },
16
  "SnowballTarget.Step.mean": {
 
26
  "count": 20
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 12.536450386047363,
30
+ "min": 0.29756423830986023,
31
+ "max": 12.538382530212402,
32
  "count": 20
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 2444.60791015625,
36
+ "min": 57.72746276855469,
37
+ "max": 2557.830078125,
38
  "count": 20
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
 
50
  "count": 20
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
+ "value": 0.07195115317510642,
54
+ "min": 0.0613162228997395,
55
+ "max": 0.07617167744733616,
56
  "count": 20
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
+ "value": 0.2878046127004257,
60
+ "min": 0.256827344703625,
61
+ "max": 0.3594081038928719,
62
  "count": 20
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
+ "value": 0.24117908371137636,
66
+ "min": 0.08387925713802413,
67
+ "max": 0.26983621847980166,
68
  "count": 20
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
+ "value": 0.9647163348455055,
72
+ "min": 0.33551702855209653,
73
+ "max": 1.2757801127200032,
74
  "count": 20
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
 
110
  "count": 20
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
+ "value": 24.204545454545453,
114
+ "min": 2.7954545454545454,
115
+ "max": 24.795454545454547,
116
  "count": 20
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
+ "value": 1065.0,
120
+ "min": 123.0,
121
+ "max": 1347.0,
122
  "count": 20
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
+ "value": 24.204545454545453,
126
+ "min": 2.7954545454545454,
127
+ "max": 24.795454545454547,
128
  "count": 20
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
+ "value": 1065.0,
132
+ "min": 123.0,
133
+ "max": 1347.0,
134
  "count": 20
135
  },
136
  "SnowballTarget.IsTraining.mean": {
 
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1684870815",
152
  "python_version": "3.10.11 (main, Apr 5 2023, 14:15:10) [GCC 9.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics",
154
  "mlagents_version": "0.31.0.dev0",
 
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.11.0+cu102",
158
  "numpy_version": "1.21.2",
159
+ "end_time_seconds": "1684871285"
160
  },
161
+ "total": 469.56133938,
162
  "count": 1,
163
+ "self": 0.842734503000031,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.042406071000016254,
167
  "count": 1,
168
+ "self": 0.042406071000016254
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 468.67619880599995,
172
  "count": 1,
173
+ "self": 0.609915781990992,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 3.816429191999987,
177
  "count": 1,
178
+ "self": 3.816429191999987
179
  },
180
  "TrainerController.advance": {
181
+ "total": 464.0328816840089,
182
+ "count": 18217,
183
+ "self": 0.28893696700981764,
184
  "children": {
185
  "env_step": {
186
+ "total": 463.74394471699907,
187
+ "count": 18217,
188
+ "self": 340.63498641598903,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 122.8177015530024,
192
+ "count": 18217,
193
+ "self": 1.795141740006727,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 121.02255981299567,
197
+ "count": 18217,
198
+ "self": 121.02255981299567
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 0.29125674800764045,
204
+ "count": 18217,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 466.7312000590057,
209
+ "count": 18217,
210
  "is_parallel": true,
211
+ "self": 217.19802221699774,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.005602129999999761,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.0041608370000858486,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.0014412929999139124,
227
  "count": 10,
228
  "is_parallel": true,
229
+ "self": 0.0014412929999139124
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.04976658599997563,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.0005963019999057906,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 0.0003819240000098034,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 0.0003819240000098034
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.04680640200001562,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.04680640200001562
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.001981958000044415,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 0.00038292199997158605,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.0015990360000728288,
259
  "count": 10,
260
  "is_parallel": true,
261
+ "self": 0.0015990360000728288
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 249.53317784200794,
271
+ "count": 18216,
272
  "is_parallel": true,
273
+ "self": 9.766430027006322,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 5.369825343998855,
277
+ "count": 18216,
278
  "is_parallel": true,
279
+ "self": 5.369825343998855
280
  },
281
  "communicator.exchange": {
282
+ "total": 201.42829633700143,
283
+ "count": 18216,
284
  "is_parallel": true,
285
+ "self": 201.42829633700143
286
  },
287
  "steps_from_proto": {
288
+ "total": 32.96862613400134,
289
+ "count": 18216,
290
  "is_parallel": true,
291
+ "self": 6.361346101003164,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 26.607280032998176,
295
+ "count": 182160,
296
  "is_parallel": true,
297
+ "self": 26.607280032998176
298
  }
299
  }
300
  }
 
309
  }
310
  },
311
  "trainer_threads": {
312
+ "total": 0.00014625900007558812,
313
  "count": 1,
314
+ "self": 0.00014625900007558812,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
 
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
+ "total": 460.60320343997773,
324
+ "count": 447862,
325
  "is_parallel": true,
326
+ "self": 9.63815720195862,
327
  "children": {
328
  "process_trajectory": {
329
+ "total": 250.9966665210183,
330
+ "count": 447862,
331
  "is_parallel": true,
332
+ "self": 249.35790389901825,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
+ "total": 1.6387626220000584,
336
  "count": 4,
337
  "is_parallel": true,
338
+ "self": 1.6387626220000584
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
+ "total": 199.9683797170008,
344
  "count": 90,
345
  "is_parallel": true,
346
+ "self": 78.98294489099652,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
+ "total": 120.98543482600428,
350
+ "count": 4587,
351
  "is_parallel": true,
352
+ "self": 120.98543482600428
353
  }
354
  }
355
  }
 
360
  }
361
  },
362
  "TrainerController._save_models": {
363
+ "total": 0.2168258890000061,
364
  "count": 1,
365
+ "self": 0.0011585900000454785,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
+ "total": 0.21566729899996062,
369
  "count": 1,
370
+ "self": 0.21566729899996062
371
  }
372
  }
373
  }
run_logs/training_status.json CHANGED
@@ -4,8 +4,8 @@
4
  {
5
  "steps": 49936,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
- "reward": 13.545454545454545,
8
- "creation_time": 1684300185.4667082,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
@@ -13,8 +13,8 @@
13
  {
14
  "steps": 99960,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
- "reward": 19.181818181818183,
17
- "creation_time": 1684300296.6563413,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
@@ -22,8 +22,8 @@
22
  {
23
  "steps": 149984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
- "reward": 23.454545454545453,
26
- "creation_time": 1684300407.7618053,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
@@ -32,7 +32,7 @@
32
  "steps": 199984,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
  "reward": 25.09090909090909,
35
- "creation_time": 1684300516.6304414,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
@@ -41,7 +41,7 @@
41
  "steps": 200112,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
  "reward": 25.09090909090909,
44
- "creation_time": 1684300516.8216317,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
  ]
@@ -51,7 +51,7 @@
51
  "steps": 200112,
52
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
  "reward": 25.09090909090909,
54
- "creation_time": 1684300516.8216317,
55
  "auxillary_file_paths": [
56
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]
 
4
  {
5
  "steps": 49936,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.onnx",
7
+ "reward": 12.0,
8
+ "creation_time": 1684870936.8293312,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-49936.pt"
11
  ]
 
13
  {
14
  "steps": 99960,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
16
+ "reward": 20.636363636363637,
17
+ "creation_time": 1684871053.1772025,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
20
  ]
 
22
  {
23
  "steps": 149984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
25
+ "reward": 24.90909090909091,
26
+ "creation_time": 1684871168.907012,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
29
  ]
 
32
  "steps": 199984,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
34
  "reward": 25.09090909090909,
35
+ "creation_time": 1684871284.4315462,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
38
  ]
 
41
  "steps": 200112,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.onnx",
43
  "reward": 25.09090909090909,
44
+ "creation_time": 1684871284.6668,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
47
  ]
 
51
  "steps": 200112,
52
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
53
  "reward": 25.09090909090909,
54
+ "creation_time": 1684871284.6668,
55
  "auxillary_file_paths": [
56
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-200112.pt"
57
  ]