arbts commited on
Commit
70fc61d
·
1 Parent(s): a5b43af
SnowballTarget.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd1e1fffdceb466face92cbe366d64a85094217c78e97ff1dd77c6c6d96cfd2c
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e8ac4c2cdef9a859e97da6a396fc7f43f5f4721549ffc9817d46446e261543f
3
+ size 326963
SnowballTarget/SnowballTarget-149984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a5435c447311075879c2d1a36651053369af3a962db23faf5b11b18d4776851
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd21b971b6334bfb2273d3cb4954514aecc25f7fd891872a8440dc1563db327e
3
+ size 326963
SnowballTarget/SnowballTarget-149984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a102800bfc9c386900c79d64bd6bb7f8f03bb367f4dfc7ebf4333da214878951
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6d400bc8f4c6a99624402e0cebdb8ae8243d1b650a4f390fbaeb672caf6bbbb
3
+ size 1938524
SnowballTarget/SnowballTarget-199984.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cde3586498c099a49193ee80f54c93260a7309e3a14d241cc9677f77591a849c
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e04099a3b1bec183694255fae0957297da26921dd6deb76ab87746ea3e4e10f
3
+ size 326963
SnowballTarget/SnowballTarget-199984.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d87bd88f8c97fbe7b2fbddd00d5117a9db641bd0f7241f150d9770fd543ff854
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73cd14430ad13e8723c8f78753451f1be96698f2188d36f4b42e358653cd09bd
3
+ size 1938524
SnowballTarget/SnowballTarget-249944.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e01cd74e4bf35d9b45e06f79a0461d19d95427e0cd471bedbafbbeb023fdff77
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f906cf747a64f5943fb0e268ef70285ded3e779a306dd81d43184eaf0112d16f
3
+ size 326963
SnowballTarget/SnowballTarget-249944.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fa311a6b26572add11ca9f9a3846ed22e03680ab0a69bcf2679ff0735795ddfd
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8894381fbf40036babb3be65d261043031bab3368c6bcf68a6d47684287a6222
3
+ size 1938524
SnowballTarget/SnowballTarget-299968.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39db50fd6cf3f943e02c9b6065906822365ea8a644a1f9be03754529c84bc724
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cb54d5914865095eecdc5256c1277de926f025cf8e54ab2aef0986b1e916e67
3
+ size 326963
SnowballTarget/SnowballTarget-299968.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd166fc81b42c7b6718c74aa07f5e502a968b36fd08a1b216e7683d8b750ff9a
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f7da96fea0ddb3234d3b3f6252d134c451d4d2b989ade5111215efe63f18dab
3
+ size 1938524
SnowballTarget/SnowballTarget-349992.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:919321b0064638441316b6a9153ab3d8b4371750b750826aa261dd554ba89d83
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82e8942d043ba3366372aed418d0a4e7fb5c38723d5ac612c70b5f33436d34f5
3
+ size 326963
SnowballTarget/SnowballTarget-349992.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8ea77bbc6d0950a28f9e4fe804b7bcfd3c137be4997bddd958c057bf0456cfd
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e0a6a8f22bc0a2d78aeb29cbb58d5f6e1745fe569199a5243f7167a68550c03
3
+ size 1938524
SnowballTarget/SnowballTarget-399992.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52c4fdf00951ac071db7791a42ef2e8f8c95cfa35202fdcf09ade05a71756781
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac2f8795efba0fd5e8915af6deb22f34a4500c321b37add4fe9b4deb04adebf5
3
+ size 326963
SnowballTarget/SnowballTarget-399992.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5f7fa2fedeed869c41c56fd0c4a9ddbb31dc6597a6a20fb39efd78f739a0b53
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79f86ef45f763c30754045d71a2cdd3d9ddc88acf5536e671ca943b6c3a493f0
3
+ size 1938524
SnowballTarget/SnowballTarget-449952.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db68880e823a5466170aa86807ceff2614b61c9e2328a4e707cb753d1310d7fb
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68b84f2960eb384159534fb0308c4c36f0c3826480ebcb53c5453b573722d2ef
3
+ size 326963
SnowballTarget/SnowballTarget-449952.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d6a73f580aaea7e27fed34dfbf982ef933d32d0a502bb1178a5e6d890558cb0c
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f71a38e4dbdc33ec62b6a03dd0aa20059ab511238fc7bd1b9bb704243b6ca6ab
3
+ size 1938524
SnowballTarget/SnowballTarget-499976.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd1e1fffdceb466face92cbe366d64a85094217c78e97ff1dd77c6c6d96cfd2c
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e8ac4c2cdef9a859e97da6a396fc7f43f5f4721549ffc9817d46446e261543f
3
+ size 326963
SnowballTarget/SnowballTarget-499976.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a9a717657ad40f64d1563acf12e29b42fbe96165e66c67d8eca5823c099bb71a
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afe648e6d1f5a43d3c8557778e1e88bbf5c1b852cba330293b1577f37d572901
3
+ size 1938524
SnowballTarget/SnowballTarget-500104.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd1e1fffdceb466face92cbe366d64a85094217c78e97ff1dd77c6c6d96cfd2c
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e8ac4c2cdef9a859e97da6a396fc7f43f5f4721549ffc9817d46446e261543f
3
+ size 326963
SnowballTarget/SnowballTarget-500104.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44a3cb9ab7ff29d9392e5df1a416eea411d1668d54159095bbd91ad0ea45f56a
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a13406d3b0801c22752ae24da1e21e324945a4b5bbdbd81213d740aabc2f83
3
+ size 1938524
SnowballTarget/SnowballTarget-99960.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:192510f4c007aea22cd111156c8dfaf98e72a9b6a591699b0c39ad5f48acdfe5
3
- size 646532
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1a1dd8c84f9e4996c2d3c9e1afd764c74a9ffc975c71afce20dd45da57764f9
3
+ size 326963
SnowballTarget/SnowballTarget-99960.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c37573de96f50298603c6573eb070bec5afd49a6ab67d14831e6180136ce55d8
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:106b5e721f18a2387a0224a15d4369dbe373b1d20b0e0da5e457d47ef59eede2
3
+ size 1938524
SnowballTarget/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44a3cb9ab7ff29d9392e5df1a416eea411d1668d54159095bbd91ad0ea45f56a
3
- size 3843968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97a13406d3b0801c22752ae24da1e21e324945a4b5bbdbd81213d740aabc2f83
3
+ size 1938524
SnowballTarget/{events.out.tfevents.1680429058.aa5e923e6146.7781.0 → events.out.tfevents.1680431052.aa5e923e6146.16497.0} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bc4877c85f3dccd17a261d8a6ddf2125a41489da8f988dd0623ddea02feb131f
3
- size 48249
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c20083069b402466b27be094ccd8a244c074527467ee517ef2d2426f50c0697
3
+ size 52586
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 256, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 500000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": true, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"SnowballTarget": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 128, "buffer_size": 2048, "learning_rate": 0.001, "beta": 0.007, "epsilon": 0.2, "lambd": 0.96, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 50000, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.99, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 10, "even_checkpoints": false, "max_steps": 500000, "time_horizon": 64, "summary_freq": 10000, "threaded": true, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./training-envs-executables/linux/SnowballTarget/SnowballTarget", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "SnowballTarget1", "initialize_from": null, "load_model": false, "resume": false, "force": true, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -5,10 +5,10 @@ behaviors:
5
  hyperparameters:
6
  batch_size: 128
7
  buffer_size: 2048
8
- learning_rate: 0.0003
9
- beta: 0.005
10
  epsilon: 0.2
11
- lambd: 0.95
12
  num_epoch: 3
13
  shared_critic: false
14
  learning_rate_schedule: linear
@@ -17,8 +17,8 @@ behaviors:
17
  checkpoint_interval: 50000
18
  network_settings:
19
  normalize: false
20
- hidden_units: 256
21
- num_layers: 2
22
  vis_encode_type: simple
23
  memory: null
24
  goal_conditioning_type: hyper
 
5
  hyperparameters:
6
  batch_size: 128
7
  buffer_size: 2048
8
+ learning_rate: 0.001
9
+ beta: 0.007
10
  epsilon: 0.2
11
+ lambd: 0.96
12
  num_epoch: 3
13
  shared_critic: false
14
  learning_rate_schedule: linear
 
17
  checkpoint_interval: 50000
18
  network_settings:
19
  normalize: false
20
+ hidden_units: 128
21
+ num_layers: 3
22
  vis_encode_type: simple
23
  memory: null
24
  goal_conditioning_type: hyper
run_logs/Player-0.log CHANGED
@@ -31,7 +31,7 @@ ALSA lib pcm.c:2642:(snd_pcm_open_noupdate) Unknown PCM default
31
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
32
  FMOD initialized on nosound output
33
  Begin MonoManager ReloadAssembly
34
- - Completed reload, in 0.081 seconds
35
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
36
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
37
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -42,7 +42,7 @@ ERROR: Shader Standard shader is not supported on this GPU (none of subshaders/f
42
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
43
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
44
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
45
- UnloadTime: 0.735158 ms
46
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
47
  requesting resize 84 x 84
48
  Setting up 1 worker threads for Enlighten.
@@ -50,7 +50,7 @@ Memory Statistics:
50
  [ALLOC_TEMP_TLS] TLS Allocator
51
  StackAllocators :
52
  [ALLOC_TEMP_MAIN]
53
- Peak usage frame count: [8.0 KB-16.0 KB]: 13425 frames, [16.0 KB-32.0 KB]: 227 frames, [2.0 MB-4.0 MB]: 1 frames
54
  Initial Block Size 4.0 MB
55
  Current Block Size 4.0 MB
56
  Peak Allocated Bytes 2.0 MB
@@ -166,22 +166,22 @@ Memory Statistics:
166
  Peak Allocated Bytes 0 B
167
  Overflow Count 0
168
  [ALLOC_DEFAULT] Dual Thread Allocator
169
- Peak main deferred allocation count 44
170
  [ALLOC_BUCKET]
171
  Large Block size 4.0 MB
172
  Used Block count 1
173
  Peak Allocated bytes 1.0 MB
174
  [ALLOC_DEFAULT_MAIN]
175
- Peak usage frame count: [4.0 MB-8.0 MB]: 11159 frames, [8.0 MB-16.0 MB]: 2494 frames
176
  Requested Block Size 16.0 MB
177
  Peak Block count 1
178
- Peak Allocated memory 9.7 MB
179
  Peak Large allocation bytes 0 B
180
  [ALLOC_DEFAULT_THREAD]
181
- Peak usage frame count: [16.0 MB-32.0 MB]: 13653 frames
182
  Requested Block Size 16.0 MB
183
  Peak Block count 1
184
- Peak Allocated memory 17.8 MB
185
  Peak Large allocation bytes 16.0 MB
186
  [ALLOC_TEMP_JOB_1_FRAME]
187
  Initial Block Size 2.0 MB
@@ -210,13 +210,13 @@ Memory Statistics:
210
  Used Block count 1
211
  Peak Allocated bytes 1.0 MB
212
  [ALLOC_GFX_MAIN]
213
- Peak usage frame count: [32.0 KB-64.0 KB]: 12485 frames, [64.0 KB-128.0 KB]: 1168 frames
214
  Requested Block Size 16.0 MB
215
  Peak Block count 1
216
- Peak Allocated memory 66.8 KB
217
  Peak Large allocation bytes 0 B
218
  [ALLOC_GFX_THREAD]
219
- Peak usage frame count: [32.0 KB-64.0 KB]: 13653 frames
220
  Requested Block Size 16.0 MB
221
  Peak Block count 1
222
  Peak Allocated memory 39.6 KB
@@ -228,13 +228,13 @@ Memory Statistics:
228
  Used Block count 1
229
  Peak Allocated bytes 1.0 MB
230
  [ALLOC_CACHEOBJECTS_MAIN]
231
- Peak usage frame count: [0.5 MB-1.0 MB]: 13653 frames
232
  Requested Block Size 4.0 MB
233
  Peak Block count 1
234
  Peak Allocated memory 0.6 MB
235
  Peak Large allocation bytes 0 B
236
  [ALLOC_CACHEOBJECTS_THREAD]
237
- Peak usage frame count: [0.5 MB-1.0 MB]: 13652 frames, [2.0 MB-4.0 MB]: 1 frames
238
  Requested Block Size 4.0 MB
239
  Peak Block count 1
240
  Peak Allocated memory 2.2 MB
@@ -246,13 +246,13 @@ Memory Statistics:
246
  Used Block count 1
247
  Peak Allocated bytes 1.0 MB
248
  [ALLOC_TYPETREE_MAIN]
249
- Peak usage frame count: [0-1.0 KB]: 13653 frames
250
  Requested Block Size 2.0 MB
251
  Peak Block count 1
252
  Peak Allocated memory 1.0 KB
253
  Peak Large allocation bytes 0 B
254
  [ALLOC_TYPETREE_THREAD]
255
- Peak usage frame count: [1.0 KB-2.0 KB]: 13653 frames
256
  Requested Block Size 2.0 MB
257
  Peak Block count 1
258
  Peak Allocated memory 1.7 KB
 
31
  FMOD failed to initialize the output device.: "Error initializing output device. " (60)
32
  FMOD initialized on nosound output
33
  Begin MonoManager ReloadAssembly
34
+ - Completed reload, in 0.080 seconds
35
  ERROR: Shader Sprites/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
36
  ERROR: Shader Sprites/Mask shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
37
  ERROR: Shader Legacy Shaders/VertexLit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
42
  WARNING: Shader Unsupported: 'Standard' - All subshaders removed
43
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
44
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
45
+ UnloadTime: 0.618109 ms
46
  ERROR: Shader UI/Default shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
47
  requesting resize 84 x 84
48
  Setting up 1 worker threads for Enlighten.
 
50
  [ALLOC_TEMP_TLS] TLS Allocator
51
  StackAllocators :
52
  [ALLOC_TEMP_MAIN]
53
+ Peak usage frame count: [8.0 KB-16.0 KB]: 13416 frames, [16.0 KB-32.0 KB]: 227 frames, [2.0 MB-4.0 MB]: 1 frames
54
  Initial Block Size 4.0 MB
55
  Current Block Size 4.0 MB
56
  Peak Allocated Bytes 2.0 MB
 
166
  Peak Allocated Bytes 0 B
167
  Overflow Count 0
168
  [ALLOC_DEFAULT] Dual Thread Allocator
169
+ Peak main deferred allocation count 45
170
  [ALLOC_BUCKET]
171
  Large Block size 4.0 MB
172
  Used Block count 1
173
  Peak Allocated bytes 1.0 MB
174
  [ALLOC_DEFAULT_MAIN]
175
+ Peak usage frame count: [4.0 MB-8.0 MB]: 11110 frames, [8.0 MB-16.0 MB]: 2534 frames
176
  Requested Block Size 16.0 MB
177
  Peak Block count 1
178
+ Peak Allocated memory 9.6 MB
179
  Peak Large allocation bytes 0 B
180
  [ALLOC_DEFAULT_THREAD]
181
+ Peak usage frame count: [16.0 MB-32.0 MB]: 13644 frames
182
  Requested Block Size 16.0 MB
183
  Peak Block count 1
184
+ Peak Allocated memory 17.7 MB
185
  Peak Large allocation bytes 16.0 MB
186
  [ALLOC_TEMP_JOB_1_FRAME]
187
  Initial Block Size 2.0 MB
 
210
  Used Block count 1
211
  Peak Allocated bytes 1.0 MB
212
  [ALLOC_GFX_MAIN]
213
+ Peak usage frame count: [32.0 KB-64.0 KB]: 13038 frames, [64.0 KB-128.0 KB]: 606 frames
214
  Requested Block Size 16.0 MB
215
  Peak Block count 1
216
+ Peak Allocated memory 67.7 KB
217
  Peak Large allocation bytes 0 B
218
  [ALLOC_GFX_THREAD]
219
+ Peak usage frame count: [32.0 KB-64.0 KB]: 13644 frames
220
  Requested Block Size 16.0 MB
221
  Peak Block count 1
222
  Peak Allocated memory 39.6 KB
 
228
  Used Block count 1
229
  Peak Allocated bytes 1.0 MB
230
  [ALLOC_CACHEOBJECTS_MAIN]
231
+ Peak usage frame count: [0.5 MB-1.0 MB]: 13644 frames
232
  Requested Block Size 4.0 MB
233
  Peak Block count 1
234
  Peak Allocated memory 0.6 MB
235
  Peak Large allocation bytes 0 B
236
  [ALLOC_CACHEOBJECTS_THREAD]
237
+ Peak usage frame count: [0.5 MB-1.0 MB]: 13643 frames, [2.0 MB-4.0 MB]: 1 frames
238
  Requested Block Size 4.0 MB
239
  Peak Block count 1
240
  Peak Allocated memory 2.2 MB
 
246
  Used Block count 1
247
  Peak Allocated bytes 1.0 MB
248
  [ALLOC_TYPETREE_MAIN]
249
+ Peak usage frame count: [0-1.0 KB]: 13644 frames
250
  Requested Block Size 2.0 MB
251
  Peak Block count 1
252
  Peak Allocated memory 1.0 KB
253
  Peak Large allocation bytes 0 B
254
  [ALLOC_TYPETREE_THREAD]
255
+ Peak usage frame count: [1.0 KB-2.0 KB]: 13644 frames
256
  Requested Block Size 2.0 MB
257
  Peak Block count 1
258
  Peak Allocated memory 1.7 KB
run_logs/timers.json CHANGED
@@ -2,15 +2,15 @@
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
- "value": 0.7794438004493713,
6
- "min": 0.769217848777771,
7
- "max": 2.8614861965179443,
8
  "count": 50
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
- "value": 7605.033203125,
12
- "min": 7520.01171875,
13
- "max": 29367.43359375,
14
  "count": 50
15
  },
16
  "SnowballTarget.Step.mean": {
@@ -26,15 +26,15 @@
26
  "count": 50
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
- "value": 13.735799789428711,
30
- "min": 0.1711287796497345,
31
- "max": 13.90906047821045,
32
  "count": 50
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
- "value": 2815.8388671875,
36
- "min": 33.19898223876953,
37
- "max": 2842.3603515625,
38
  "count": 50
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
@@ -50,39 +50,39 @@
50
  "count": 50
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
- "value": 0.07091386185683066,
54
- "min": 0.06160780617602879,
55
- "max": 0.0754181212882566,
56
  "count": 50
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
- "value": 0.35456930928415326,
60
- "min": 0.24643122470411516,
61
- "max": 0.377090606441283,
62
  "count": 50
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
- "value": 0.17666345916542353,
66
- "min": 0.09802576110757633,
67
- "max": 0.2818825277043324,
68
  "count": 50
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
- "value": 0.8833172958271176,
72
- "min": 0.3921030444303053,
73
- "max": 1.244458389925022,
74
  "count": 50
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
- "value": 3.0528989824000028e-06,
78
- "min": 3.0528989824000028e-06,
79
- "max": 0.00029675280108239997,
80
  "count": 50
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
83
- "value": 1.5264494912000015e-05,
84
- "min": 1.5264494912000015e-05,
85
- "max": 0.001454064015312,
86
  "count": 50
87
  },
88
  "SnowballTarget.Policy.Epsilon.mean": {
@@ -98,39 +98,39 @@
98
  "count": 50
99
  },
100
  "SnowballTarget.Policy.Beta.mean": {
101
- "value": 6.0778240000000044e-05,
102
- "min": 6.0778240000000044e-05,
103
- "max": 0.00494598824,
104
  "count": 50
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
107
- "value": 0.0003038912000000002,
108
- "min": 0.0003038912000000002,
109
- "max": 0.0242359312,
110
  "count": 50
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
- "value": 27.01818181818182,
114
- "min": 2.9318181818181817,
115
- "max": 27.40740740740741,
116
  "count": 50
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
- "value": 1486.0,
120
- "min": 129.0,
121
- "max": 1497.0,
122
  "count": 50
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
- "value": 27.01818181818182,
126
- "min": 2.9318181818181817,
127
- "max": 27.40740740740741,
128
  "count": 50
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
- "value": 1486.0,
132
- "min": 129.0,
133
- "max": 1497.0,
134
  "count": 50
135
  },
136
  "SnowballTarget.IsTraining.mean": {
@@ -148,7 +148,7 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1680429058",
152
  "python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics --force",
154
  "mlagents_version": "0.31.0.dev0",
@@ -156,59 +156,59 @@
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.11.0+cu102",
158
  "numpy_version": "1.21.2",
159
- "end_time_seconds": "1680430214"
160
  },
161
- "total": 1155.816538462,
162
  "count": 1,
163
- "self": 0.7799762839999858,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.10557375099983801,
167
  "count": 1,
168
- "self": 0.10557375099983801
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 1154.930988427,
172
  "count": 1,
173
- "self": 1.35956119197499,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 4.268539559999908,
177
  "count": 1,
178
- "self": 4.268539559999908
179
  },
180
  "TrainerController.advance": {
181
- "total": 1149.0903089550254,
182
- "count": 45507,
183
- "self": 0.6911831659599557,
184
  "children": {
185
  "env_step": {
186
- "total": 1148.3991257890655,
187
- "count": 45507,
188
- "self": 838.9038802160485,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 308.7796133849972,
192
- "count": 45507,
193
- "self": 4.300721927978202,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 304.478891457019,
197
- "count": 45507,
198
- "self": 304.478891457019
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 0.7156321880197538,
204
- "count": 45507,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 1151.229387694056,
209
- "count": 45507,
210
  "is_parallel": true,
211
- "self": 527.9214686660557,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +217,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.0019385410000722914,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.0006229280002116866,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0013156129998606048,
227
  "count": 10,
228
  "is_parallel": true,
229
- "self": 0.0013156129998606048
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.11472348799998144,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.0008184110001820954,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.00043315900006746233,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.00043315900006746233
244
  },
245
  "communicator.exchange": {
246
- "total": 0.1096483259998422,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.1096483259998422
250
  },
251
  "steps_from_proto": {
252
- "total": 0.003823591999889686,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.0005145469997387409,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.0033090450001509453,
259
  "count": 10,
260
  "is_parallel": true,
261
- "self": 0.0033090450001509453
262
  }
263
  }
264
  }
@@ -267,34 +267,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 623.3079190280002,
271
- "count": 45506,
272
  "is_parallel": true,
273
- "self": 24.74701056787626,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 14.338586199028668,
277
- "count": 45506,
278
  "is_parallel": true,
279
- "self": 14.338586199028668
280
  },
281
  "communicator.exchange": {
282
- "total": 502.19646343604745,
283
- "count": 45506,
284
  "is_parallel": true,
285
- "self": 502.19646343604745
286
  },
287
  "steps_from_proto": {
288
- "total": 82.02585882504786,
289
- "count": 45506,
290
  "is_parallel": true,
291
- "self": 16.147291615917993,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 65.87856720912987,
295
- "count": 455060,
296
  "is_parallel": true,
297
- "self": 65.87856720912987
298
  }
299
  }
300
  }
@@ -309,9 +309,9 @@
309
  }
310
  },
311
  "trainer_threads": {
312
- "total": 0.00011846999996123486,
313
  "count": 1,
314
- "self": 0.00011846999996123486,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
@@ -320,36 +320,36 @@
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
- "total": 1140.0188054531313,
324
- "count": 1042290,
325
  "is_parallel": true,
326
- "self": 25.206867358745285,
327
  "children": {
328
  "process_trajectory": {
329
- "total": 631.6825475373869,
330
- "count": 1042290,
331
  "is_parallel": true,
332
- "self": 628.1209678053867,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
- "total": 3.5615797320001548,
336
  "count": 10,
337
  "is_parallel": true,
338
- "self": 3.5615797320001548
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
- "total": 483.12939055699917,
344
  "count": 227,
345
  "is_parallel": true,
346
- "self": 178.141558771998,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
- "total": 304.98783178500116,
350
  "count": 11574,
351
  "is_parallel": true,
352
- "self": 304.98783178500116
353
  }
354
  }
355
  }
@@ -360,14 +360,14 @@
360
  }
361
  },
362
  "TrainerController._save_models": {
363
- "total": 0.21246024999982183,
364
  "count": 1,
365
- "self": 0.002099129999805882,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
- "total": 0.21036112000001594,
369
  "count": 1,
370
- "self": 0.21036112000001594
371
  }
372
  }
373
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "SnowballTarget.Policy.Entropy.mean": {
5
+ "value": 0.9076099991798401,
6
+ "min": 0.9076099991798401,
7
+ "max": 2.8623406887054443,
8
  "count": 50
9
  },
10
  "SnowballTarget.Policy.Entropy.sum": {
11
+ "value": 8785.6650390625,
12
+ "min": 8785.6650390625,
13
+ "max": 29344.716796875,
14
  "count": 50
15
  },
16
  "SnowballTarget.Step.mean": {
 
26
  "count": 50
27
  },
28
  "SnowballTarget.Policy.ExtrinsicValueEstimate.mean": {
29
+ "value": 14.108038902282715,
30
+ "min": 0.5381110310554504,
31
+ "max": 14.108038902282715,
32
  "count": 50
33
  },
34
  "SnowballTarget.Policy.ExtrinsicValueEstimate.sum": {
35
+ "value": 2892.14794921875,
36
+ "min": 104.39353942871094,
37
+ "max": 2892.14794921875,
38
  "count": 50
39
  },
40
  "SnowballTarget.Environment.EpisodeLength.mean": {
 
50
  "count": 50
51
  },
52
  "SnowballTarget.Losses.PolicyLoss.mean": {
53
+ "value": 0.07024131641887565,
54
+ "min": 0.05859752080954529,
55
+ "max": 0.0768773556619691,
56
  "count": 50
57
  },
58
  "SnowballTarget.Losses.PolicyLoss.sum": {
59
+ "value": 0.3512065820943783,
60
+ "min": 0.23439008323818117,
61
+ "max": 0.36810750479078563,
62
  "count": 50
63
  },
64
  "SnowballTarget.Losses.ValueLoss.mean": {
65
+ "value": 0.185192534327507,
66
+ "min": 0.14188875462038114,
67
+ "max": 0.3432107374218164,
68
  "count": 50
69
  },
70
  "SnowballTarget.Losses.ValueLoss.sum": {
71
+ "value": 0.9259626716375351,
72
+ "min": 0.5675550184815246,
73
+ "max": 1.6073153565327325,
74
  "count": 50
75
  },
76
  "SnowballTarget.Policy.LearningRate.mean": {
77
+ "value": 1.0176098982400003e-05,
78
+ "min": 1.0176098982400003e-05,
79
+ "max": 0.0009891760010824,
80
  "count": 50
81
  },
82
  "SnowballTarget.Policy.LearningRate.sum": {
83
+ "value": 5.088049491200002e-05,
84
+ "min": 5.088049491200002e-05,
85
+ "max": 0.0048468800153119995,
86
  "count": 50
87
  },
88
  "SnowballTarget.Policy.Epsilon.mean": {
 
98
  "count": 50
99
  },
100
  "SnowballTarget.Policy.Beta.mean": {
101
+ "value": 8.113024000000005e-05,
102
+ "min": 8.113024000000005e-05,
103
+ "max": 0.00692434024,
104
  "count": 50
105
  },
106
  "SnowballTarget.Policy.Beta.sum": {
107
+ "value": 0.00040565120000000027,
108
+ "min": 0.00040565120000000027,
109
+ "max": 0.0339296912,
110
  "count": 50
111
  },
112
  "SnowballTarget.Environment.CumulativeReward.mean": {
113
+ "value": 27.618181818181817,
114
+ "min": 3.8181818181818183,
115
+ "max": 27.98181818181818,
116
  "count": 50
117
  },
118
  "SnowballTarget.Environment.CumulativeReward.sum": {
119
+ "value": 1519.0,
120
+ "min": 168.0,
121
+ "max": 1539.0,
122
  "count": 50
123
  },
124
  "SnowballTarget.Policy.ExtrinsicReward.mean": {
125
+ "value": 27.618181818181817,
126
+ "min": 3.8181818181818183,
127
+ "max": 27.98181818181818,
128
  "count": 50
129
  },
130
  "SnowballTarget.Policy.ExtrinsicReward.sum": {
131
+ "value": 1519.0,
132
+ "min": 168.0,
133
+ "max": 1539.0,
134
  "count": 50
135
  },
136
  "SnowballTarget.IsTraining.mean": {
 
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1680431052",
152
  "python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
153
  "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/SnowballTarget.yaml --env=./training-envs-executables/linux/SnowballTarget/SnowballTarget --run-id=SnowballTarget1 --no-graphics --force",
154
  "mlagents_version": "0.31.0.dev0",
 
156
  "communication_protocol_version": "1.5.0",
157
  "pytorch_version": "1.11.0+cu102",
158
  "numpy_version": "1.21.2",
159
+ "end_time_seconds": "1680432199"
160
  },
161
+ "total": 1147.5846234949995,
162
  "count": 1,
163
+ "self": 0.3831024679989241,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.10527087800028312,
167
  "count": 1,
168
+ "self": 0.10527087800028312
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 1147.0962501490003,
172
  "count": 1,
173
+ "self": 1.3202531629540317,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 3.4788207030001104,
177
  "count": 1,
178
+ "self": 3.4788207030001104
179
  },
180
  "TrainerController.advance": {
181
+ "total": 1142.1677606410462,
182
+ "count": 45477,
183
+ "self": 0.6545610391285663,
184
  "children": {
185
  "env_step": {
186
+ "total": 1141.5131996019177,
187
+ "count": 45477,
188
+ "self": 831.4854457857655,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 309.3698402861228,
192
+ "count": 45477,
193
+ "self": 4.258624800032521,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 305.11121548609026,
197
+ "count": 45477,
198
+ "self": 305.11121548609026
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 0.6579135300294183,
204
+ "count": 45477,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 1143.2282684159159,
209
+ "count": 45477,
210
  "is_parallel": true,
211
+ "self": 532.258437057948,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.001822249999804626,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.0005693119987881801,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.001252938001016446,
227
  "count": 10,
228
  "is_parallel": true,
229
+ "self": 0.001252938001016446
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.047262445999876945,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.0003612219989008736,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 0.00031873800026005483,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 0.00031873800026005483
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.045516626000335236,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.045516626000335236
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.0010658600003807805,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 0.00023944299937284086,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.0008264170010079397,
259
  "count": 10,
260
  "is_parallel": true,
261
+ "self": 0.0008264170010079397
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 610.9698313579679,
271
+ "count": 45476,
272
  "is_parallel": true,
273
+ "self": 24.534858940867252,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 13.40537451399996,
277
+ "count": 45476,
278
  "is_parallel": true,
279
+ "self": 13.40537451399996
280
  },
281
  "communicator.exchange": {
282
+ "total": 492.1718367020362,
283
+ "count": 45476,
284
  "is_parallel": true,
285
+ "self": 492.1718367020362
286
  },
287
  "steps_from_proto": {
288
+ "total": 80.85776120106448,
289
+ "count": 45476,
290
  "is_parallel": true,
291
+ "self": 15.987068659212127,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 64.87069254185235,
295
+ "count": 454760,
296
  "is_parallel": true,
297
+ "self": 64.87069254185235
298
  }
299
  }
300
  }
 
309
  }
310
  },
311
  "trainer_threads": {
312
+ "total": 0.00011574400014069397,
313
  "count": 1,
314
+ "self": 0.00011574400014069397,
315
  "children": {
316
  "thread_root": {
317
  "total": 0.0,
 
320
  "self": 0.0,
321
  "children": {
322
  "trainer_advance": {
323
+ "total": 1133.4342856535195,
324
+ "count": 1034286,
325
  "is_parallel": true,
326
+ "self": 25.187041478791798,
327
  "children": {
328
  "process_trajectory": {
329
+ "total": 627.3703568027217,
330
+ "count": 1034286,
331
  "is_parallel": true,
332
+ "self": 624.8085960577223,
333
  "children": {
334
  "RLTrainer._checkpoint": {
335
+ "total": 2.5617607449994466,
336
  "count": 10,
337
  "is_parallel": true,
338
+ "self": 2.5617607449994466
339
  }
340
  }
341
  },
342
  "_update_policy": {
343
+ "total": 480.87688737200597,
344
  "count": 227,
345
  "is_parallel": true,
346
+ "self": 172.03355409504002,
347
  "children": {
348
  "TorchPPOOptimizer.update": {
349
+ "total": 308.84333327696595,
350
  "count": 11574,
351
  "is_parallel": true,
352
+ "self": 308.84333327696595
353
  }
354
  }
355
  }
 
360
  }
361
  },
362
  "TrainerController._save_models": {
363
+ "total": 0.12929989799977193,
364
  "count": 1,
365
+ "self": 0.000816151999970316,
366
  "children": {
367
  "RLTrainer._checkpoint": {
368
+ "total": 0.1284837459998016,
369
  "count": 1,
370
+ "self": 0.1284837459998016
371
  }
372
  }
373
  }
run_logs/training_status.json CHANGED
@@ -4,8 +4,8 @@
4
  {
5
  "steps": 99960,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
7
- "reward": 21.727272727272727,
8
- "creation_time": 1680429296.4610813,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
11
  ]
@@ -13,8 +13,8 @@
13
  {
14
  "steps": 149984,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
16
- "reward": 25.0,
17
- "creation_time": 1680429412.1024163,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
20
  ]
@@ -22,8 +22,8 @@
22
  {
23
  "steps": 199984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
25
- "reward": 25.90909090909091,
26
- "creation_time": 1680429525.4227335,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
29
  ]
@@ -31,8 +31,8 @@
31
  {
32
  "steps": 249944,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-249944.onnx",
34
- "reward": 26.454545454545453,
35
- "creation_time": 1680429639.5173824,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-249944.pt"
38
  ]
@@ -40,8 +40,8 @@
40
  {
41
  "steps": 299968,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-299968.onnx",
43
- "reward": 27.272727272727273,
44
- "creation_time": 1680429754.9410768,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-299968.pt"
47
  ]
@@ -49,8 +49,8 @@
49
  {
50
  "steps": 349992,
51
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-349992.onnx",
52
- "reward": 27.363636363636363,
53
- "creation_time": 1680429869.1451,
54
  "auxillary_file_paths": [
55
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-349992.pt"
56
  ]
@@ -58,8 +58,8 @@
58
  {
59
  "steps": 399992,
60
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-399992.onnx",
61
- "reward": 26.818181818181817,
62
- "creation_time": 1680429982.9439065,
63
  "auxillary_file_paths": [
64
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-399992.pt"
65
  ]
@@ -68,7 +68,7 @@
68
  "steps": 449952,
69
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.onnx",
70
  "reward": 27.181818181818183,
71
- "creation_time": 1680430097.4575415,
72
  "auxillary_file_paths": [
73
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.pt"
74
  ]
@@ -76,8 +76,8 @@
76
  {
77
  "steps": 499976,
78
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-499976.onnx",
79
- "reward": 27.363636363636363,
80
- "creation_time": 1680430213.0998704,
81
  "auxillary_file_paths": [
82
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-499976.pt"
83
  ]
@@ -85,8 +85,8 @@
85
  {
86
  "steps": 500104,
87
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.onnx",
88
- "reward": 27.363636363636363,
89
- "creation_time": 1680430213.3303428,
90
  "auxillary_file_paths": [
91
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
92
  ]
@@ -95,8 +95,8 @@
95
  "final_checkpoint": {
96
  "steps": 500104,
97
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
98
- "reward": 27.363636363636363,
99
- "creation_time": 1680430213.3303428,
100
  "auxillary_file_paths": [
101
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
102
  ]
 
4
  {
5
  "steps": 99960,
6
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.onnx",
7
+ "reward": 22.0,
8
+ "creation_time": 1680431286.9806924,
9
  "auxillary_file_paths": [
10
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-99960.pt"
11
  ]
 
13
  {
14
  "steps": 149984,
15
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.onnx",
16
+ "reward": 25.454545454545453,
17
+ "creation_time": 1680431402.9457452,
18
  "auxillary_file_paths": [
19
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-149984.pt"
20
  ]
 
22
  {
23
  "steps": 199984,
24
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.onnx",
25
+ "reward": 26.454545454545453,
26
+ "creation_time": 1680431518.3806846,
27
  "auxillary_file_paths": [
28
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-199984.pt"
29
  ]
 
31
  {
32
  "steps": 249944,
33
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-249944.onnx",
34
+ "reward": 25.90909090909091,
35
+ "creation_time": 1680431634.4951496,
36
  "auxillary_file_paths": [
37
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-249944.pt"
38
  ]
 
40
  {
41
  "steps": 299968,
42
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-299968.onnx",
43
+ "reward": 25.181818181818183,
44
+ "creation_time": 1680431749.2137773,
45
  "auxillary_file_paths": [
46
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-299968.pt"
47
  ]
 
49
  {
50
  "steps": 349992,
51
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-349992.onnx",
52
+ "reward": 27.636363636363637,
53
+ "creation_time": 1680431862.7742105,
54
  "auxillary_file_paths": [
55
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-349992.pt"
56
  ]
 
58
  {
59
  "steps": 399992,
60
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-399992.onnx",
61
+ "reward": 27.272727272727273,
62
+ "creation_time": 1680431974.3715184,
63
  "auxillary_file_paths": [
64
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-399992.pt"
65
  ]
 
68
  "steps": 449952,
69
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.onnx",
70
  "reward": 27.181818181818183,
71
+ "creation_time": 1680432087.5055535,
72
  "auxillary_file_paths": [
73
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-449952.pt"
74
  ]
 
76
  {
77
  "steps": 499976,
78
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-499976.onnx",
79
+ "reward": 26.545454545454547,
80
+ "creation_time": 1680432199.3106952,
81
  "auxillary_file_paths": [
82
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-499976.pt"
83
  ]
 
85
  {
86
  "steps": 500104,
87
  "file_path": "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.onnx",
88
+ "reward": 26.545454545454547,
89
+ "creation_time": 1680432199.4643004,
90
  "auxillary_file_paths": [
91
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
92
  ]
 
95
  "final_checkpoint": {
96
  "steps": 500104,
97
  "file_path": "results/SnowballTarget1/SnowballTarget.onnx",
98
+ "reward": 26.545454545454547,
99
+ "creation_time": 1680432199.4643004,
100
  "auxillary_file_paths": [
101
  "results/SnowballTarget1/SnowballTarget/SnowballTarget-500104.pt"
102
  ]