wangxso commited on
Commit
94b536b
·
verified ·
1 Parent(s): 1747112
Huggy.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba85d7d142c7d3d4e85a823e59cabc3cbbc2676addc6e6a06396cd9badb37493
3
- size 2271327
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcba2d82b5c801cc6f27e4cf9ee9228f639f82ca66b1482877b3b0b610feb756
3
+ size 2274020
Huggy/Huggy-1199995.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4109ceacd0456b3e7241d5680dadc8e30283203fb7013a9ad16358d56c811de5
3
+ size 2274020
Huggy/Huggy-1199995.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39c021287bf7e07fa883ee202aff86d26bf66400143ebcde6c540a68813a5c9b
3
+ size 13509946
Huggy/Huggy-1399940.onnx CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a20356e4f0283cc22f44b2128f511870fd00e767739e2875812790054a33808e
3
- size 2271327
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d5755614598335c22a8b890fdd89ac67e92615c7631ec640acf82b12c2ba892
3
+ size 2274020
Huggy/Huggy-1399940.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5b4a7cbe34530fda44d1dbc0d06f9f215d4321caa4d27ea8ae904ed5ed3e95ea
3
- size 13503717
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffd62ad6b024d7b4eda3a8bd0be708bfa57777e15531f92c3740d988745c6261
3
+ size 13509946
Huggy/Huggy-1599966.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a1b7e33b0f41c052db3728ae4fd83233c84304d9b048f14512b9584e8cb8680
3
+ size 2274020
Huggy/Huggy-1599966.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31f794d99a8ee039030acb1dafacad77f48f19155257994f30510fdbbc4efe08
3
+ size 13509946
Huggy/Huggy-1799988.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1bb15425e3427cf440698e0518d16a601bc05ef00460a45af87df010d3497d8
3
+ size 2274020
Huggy/Huggy-1799988.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3815286c7a1118fdc8913f0194939c4a658968b07221776c68c3a93ef4a576eb
3
+ size 13509946
Huggy/Huggy-199867.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9434860f1099b8c9ed54a05a4c6efc3b18febd0c02f60280b9656c81211cba18
3
+ size 2274020
Huggy/Huggy-199867.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a5dcb0f31c7bd8f2a2c92af4e5f7b70405f3a4444afc9e6aed39d6db29ca602
3
+ size 13509859
Huggy/Huggy-1999996.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1adf76ea8ef7dd7bba7a8d14e7d381a615dcee23b81a0fc25b881c595372ec4c
3
+ size 2274020
Huggy/Huggy-1999996.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a135b89800fa528a5a42da84d09b1fb44f898fe6b3117b414163fcf979495dd
3
+ size 13509946
Huggy/Huggy-2000076.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcba2d82b5c801cc6f27e4cf9ee9228f639f82ca66b1482877b3b0b610feb756
3
+ size 2274020
Huggy/Huggy-2000076.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3fa224b9ac63c3c73155573957aac07d8dcd4929b622c6a0e6dfcdfe87850d9
3
+ size 13509946
Huggy/Huggy-399973.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93b447c0d4af41441bce248448ee40122dd68a68dbf847825cd8627b0b5cc828
3
+ size 2274020
Huggy/Huggy-399973.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a09350cc9c02dfdd0df4db820b8e86279c8376a3922c0ce7de222514d9955be1
3
+ size 13509859
Huggy/Huggy-599911.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8849a7f42936142fe7670a42ff33ee910e704fbd87d95b5ee7ce049d92d6e27
3
+ size 2274020
Huggy/Huggy-599911.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd10174abb74135bcfff663392bc957b5e7f680ffdb148b2efb06bc93fcb6be6
3
+ size 13509859
Huggy/Huggy-799888.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:395435e2e801b42c53e198db3eaf25ed9795ffee7c2f7f6d706fa3d00bc76e03
3
+ size 2274020
Huggy/Huggy-799888.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b766ad3c0a948bef3058bb2cbb1a901a6e506d02cb2ce28129b4054ed679a7e
3
+ size 13509859
Huggy/Huggy-999982.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f65f16988d80eee08aca8c22cc1740dacf83ce3bc5a82f369d936996805b8f7
3
+ size 2274020
Huggy/Huggy-999982.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d6437fa9f16cd3f5f998dcb468a012831c41d1d0e4422886dfe51df26d6b1e8
3
+ size 13509859
Huggy/checkpoint.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac1578bb3182756dc80f70c4dcef23f741f1e088b3687e45111aa52423efe955
3
- size 13503717
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd7bf0a9bc10214df4b8e69e513ef7ede1e5592f2d87563795aeabdf2fbd2fba
3
+ size 13509685
Huggy/events.out.tfevents.1761203416.wangxso.11259.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19ef9df33df76b27cecec81f10b7a4d98e51087c3896eb3fa8ef50e42605ee05
3
+ size 416481
config.json CHANGED
@@ -1 +1 @@
1
- {"default_settings": null, "behaviors": {"Huggy": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 200000, "network_settings": {"normalize": true, "hidden_units": 512, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.995, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 15, "even_checkpoints": false, "max_steps": 2000000, "time_horizon": 1000, "summary_freq": 50000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./trained-envs-executables/linux/Huggy/Huggy", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Huggy", "initialize_from": null, "load_model": false, "resume": true, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
 
1
+ {"default_settings": null, "behaviors": {"Huggy": {"trainer_type": "ppo", "hyperparameters": {"batch_size": 2048, "buffer_size": 20480, "learning_rate": 0.0003, "beta": 0.005, "epsilon": 0.2, "lambd": 0.95, "num_epoch": 3, "shared_critic": false, "learning_rate_schedule": "linear", "beta_schedule": "linear", "epsilon_schedule": "linear"}, "checkpoint_interval": 200000, "network_settings": {"normalize": true, "hidden_units": 512, "num_layers": 3, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}, "reward_signals": {"extrinsic": {"gamma": 0.995, "strength": 1.0, "network_settings": {"normalize": false, "hidden_units": 128, "num_layers": 2, "vis_encode_type": "simple", "memory": null, "goal_conditioning_type": "hyper", "deterministic": false}}}, "init_path": null, "keep_checkpoints": 15, "even_checkpoints": false, "max_steps": 2000000, "time_horizon": 1000, "summary_freq": 50000, "threaded": false, "self_play": null, "behavioral_cloning": null}}, "env_settings": {"env_path": "./trained-envs-executables/linux/Huggy/Huggy", "env_args": null, "base_port": 5005, "num_envs": 1, "num_areas": 1, "timeout_wait": 60, "seed": -1, "max_lifetime_restarts": 10, "restarts_rate_limit_n": 1, "restarts_rate_limit_period_s": 60}, "engine_settings": {"width": 84, "height": 84, "quality_level": 5, "time_scale": 20, "target_frame_rate": -1, "capture_frame_rate": 60, "no_graphics": true, "no_graphics_monitor": false}, "environment_parameters": null, "checkpoint_settings": {"run_id": "Huggy2", "initialize_from": null, "load_model": false, "resume": false, "force": false, "train_model": false, "inference": false, "results_dir": "results"}, "torch_settings": {"device": null}, "debug": false}
configuration.yaml CHANGED
@@ -50,6 +50,7 @@ env_settings:
50
  base_port: 5005
51
  num_envs: 1
52
  num_areas: 1
 
53
  seed: -1
54
  max_lifetime_restarts: 10
55
  restarts_rate_limit_n: 1
@@ -62,12 +63,13 @@ engine_settings:
62
  target_frame_rate: -1
63
  capture_frame_rate: 60
64
  no_graphics: true
 
65
  environment_parameters: null
66
  checkpoint_settings:
67
- run_id: Huggy
68
  initialize_from: null
69
  load_model: false
70
- resume: true
71
  force: false
72
  train_model: false
73
  inference: false
 
50
  base_port: 5005
51
  num_envs: 1
52
  num_areas: 1
53
+ timeout_wait: 60
54
  seed: -1
55
  max_lifetime_restarts: 10
56
  restarts_rate_limit_n: 1
 
63
  target_frame_rate: -1
64
  capture_frame_rate: 60
65
  no_graphics: true
66
+ no_graphics_monitor: false
67
  environment_parameters: null
68
  checkpoint_settings:
69
+ run_id: Huggy2
70
  initialize_from: null
71
  load_model: false
72
+ resume: false
73
  force: false
74
  train_model: false
75
  inference: false
run_logs/Player-0.log CHANGED
@@ -1,37 +1,20 @@
1
- Mono path[0] = '/content/ml-agents/trained-envs-executables/linux/Huggy/Huggy_Data/Managed'
2
- Mono config path = '/content/ml-agents/trained-envs-executables/linux/Huggy/Huggy_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
 
 
 
5
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
6
- [Subsystems] Discovering subsystems at path /content/ml-agents/trained-envs-executables/linux/Huggy/Huggy_Data/UnitySubsystems
7
  Forcing GfxDevice: Null
8
  GfxDevice: creating device client; threaded=0; jobified=0
9
  NullGfxDevice:
10
  Version: NULL 1.0 [1.0]
11
  Renderer: Null Device
12
  Vendor: Unity Technologies
13
- ALSA lib confmisc.c:855:(parse_card) cannot find card '0'
14
- ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_card_inum returned error: No such file or directory
15
- ALSA lib confmisc.c:422:(snd_func_concat) error evaluating strings
16
- ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
17
- ALSA lib confmisc.c:1334:(snd_func_refer) error evaluating name
18
- ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
19
- ALSA lib conf.c:5701:(snd_config_expand) Evaluate error: No such file or directory
20
- ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
21
- FMOD failed to initialize the output device.: "Error initializing output device. " (60)
22
- Forced to initialize FMOD to to the device driver's system output rate 48000, this may impact performance and/or give inconsistent experiences compared to selected sample rate 48000
23
- ALSA lib confmisc.c:855:(parse_card) cannot find card '0'
24
- ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_card_inum returned error: No such file or directory
25
- ALSA lib confmisc.c:422:(snd_func_concat) error evaluating strings
26
- ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_concat returned error: No such file or directory
27
- ALSA lib confmisc.c:1334:(snd_func_refer) error evaluating name
28
- ALSA lib conf.c:5178:(_snd_config_evaluate) function snd_func_refer returned error: No such file or directory
29
- ALSA lib conf.c:5701:(snd_config_expand) Evaluate error: No such file or directory
30
- ALSA lib pcm.c:2664:(snd_pcm_open_noupdate) Unknown PCM default
31
- FMOD failed to initialize the output device.: "Error initializing output device. " (60)
32
- FMOD initialized on nosound output
33
  Begin MonoManager ReloadAssembly
34
- - Completed reload, in 0.099 seconds
35
  ERROR: Shader Hidden/Universal Render Pipeline/Blit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
36
  ERROR: Shader Hidden/Universal Render Pipeline/CopyDepth shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
37
  ERROR: Shader Hidden/Universal Render Pipeline/ScreenSpaceShadows shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
@@ -75,217 +58,6 @@ ERROR: Shader Universal Render Pipeline/Lit shader is not supported on this GPU
75
  WARNING: Shader Unsupported: 'Universal Render Pipeline/Lit' - All subshaders removed
76
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
77
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
78
- UnloadTime: 0.798049 ms
79
  requesting resize 84 x 84
80
- Setting up 1 worker threads for Enlighten.
81
- Memory Statistics:
82
- [ALLOC_TEMP_TLS] TLS Allocator
83
- StackAllocators :
84
- [ALLOC_TEMP_MAIN]
85
- Peak usage frame count: [4.0 KB-8.0 KB]: 26545 frames, [2.0 MB-4.0 MB]: 1 frames
86
- Initial Block Size 4.0 MB
87
- Current Block Size 4.0 MB
88
- Peak Allocated Bytes 3.6 MB
89
- Overflow Count 0
90
- [ALLOC_TEMP_Loading.AsyncRead]
91
- Initial Block Size 64.0 KB
92
- Current Block Size 64.0 KB
93
- Peak Allocated Bytes 192 B
94
- Overflow Count 0
95
- [ALLOC_TEMP_Loading.PreloadManager]
96
- Initial Block Size 256.0 KB
97
- Current Block Size 296.0 KB
98
- Peak Allocated Bytes 211.5 KB
99
- Overflow Count 4
100
- [ALLOC_TEMP_Background Job.Worker 8]
101
- Initial Block Size 32.0 KB
102
- Current Block Size 32.0 KB
103
- Peak Allocated Bytes 0 B
104
- Overflow Count 0
105
- [ALLOC_TEMP_Background Job.Worker 9]
106
- Initial Block Size 32.0 KB
107
- Current Block Size 32.0 KB
108
- Peak Allocated Bytes 0 B
109
- Overflow Count 0
110
- [ALLOC_TEMP_Job.Worker 0]
111
- Initial Block Size 256.0 KB
112
- Current Block Size 256.0 KB
113
- Peak Allocated Bytes 0 B
114
- Overflow Count 0
115
- [ALLOC_TEMP_Background Job.Worker 10]
116
- Initial Block Size 32.0 KB
117
- Current Block Size 32.0 KB
118
- Peak Allocated Bytes 0 B
119
- Overflow Count 0
120
- [ALLOC_TEMP_Background Job.Worker 14]
121
- Initial Block Size 32.0 KB
122
- Current Block Size 32.0 KB
123
- Peak Allocated Bytes 0 B
124
- Overflow Count 0
125
- [ALLOC_TEMP_Background Job.Worker 6]
126
- Initial Block Size 32.0 KB
127
- Current Block Size 32.0 KB
128
- Peak Allocated Bytes 0 B
129
- Overflow Count 0
130
- [ALLOC_TEMP_Background Job.Worker 12]
131
- Initial Block Size 32.0 KB
132
- Current Block Size 32.0 KB
133
- Peak Allocated Bytes 0 B
134
- Overflow Count 0
135
- [ALLOC_TEMP_EnlightenWorker]
136
- Initial Block Size 64.0 KB
137
- Current Block Size 64.0 KB
138
- Peak Allocated Bytes 0 B
139
- Overflow Count 0
140
- [ALLOC_TEMP_Background Job.Worker 15]
141
- Initial Block Size 32.0 KB
142
- Current Block Size 32.0 KB
143
- Peak Allocated Bytes 0 B
144
- Overflow Count 0
145
- [ALLOC_TEMP_Background Job.Worker 1]
146
- Initial Block Size 32.0 KB
147
- Current Block Size 32.0 KB
148
- Peak Allocated Bytes 0 B
149
- Overflow Count 0
150
- [ALLOC_TEMP_Background Job.Worker 2]
151
- Initial Block Size 32.0 KB
152
- Current Block Size 32.0 KB
153
- Peak Allocated Bytes 0 B
154
- Overflow Count 0
155
- [ALLOC_TEMP_Background Job.Worker 7]
156
- Initial Block Size 32.0 KB
157
- Current Block Size 32.0 KB
158
- Peak Allocated Bytes 0 B
159
- Overflow Count 0
160
- [ALLOC_TEMP_AssetGarbageCollectorHelper]
161
- Initial Block Size 64.0 KB
162
- Current Block Size 64.0 KB
163
- Peak Allocated Bytes 0 B
164
- Overflow Count 0
165
- [ALLOC_TEMP_Background Job.Worker 5]
166
- Initial Block Size 32.0 KB
167
- Current Block Size 32.0 KB
168
- Peak Allocated Bytes 0 B
169
- Overflow Count 0
170
- [ALLOC_TEMP_Background Job.Worker 13]
171
- Initial Block Size 32.0 KB
172
- Current Block Size 32.0 KB
173
- Peak Allocated Bytes 0 B
174
- Overflow Count 0
175
- [ALLOC_TEMP_Background Job.Worker 11]
176
- Initial Block Size 32.0 KB
177
- Current Block Size 32.0 KB
178
- Peak Allocated Bytes 0 B
179
- Overflow Count 0
180
- [ALLOC_TEMP_Background Job.Worker 3]
181
- Initial Block Size 32.0 KB
182
- Current Block Size 32.0 KB
183
- Peak Allocated Bytes 0 B
184
- Overflow Count 0
185
- [ALLOC_TEMP_Background Job.Worker 0]
186
- Initial Block Size 32.0 KB
187
- Current Block Size 32.0 KB
188
- Peak Allocated Bytes 0 B
189
- Overflow Count 0
190
- [ALLOC_TEMP_Background Job.Worker 4]
191
- Initial Block Size 32.0 KB
192
- Current Block Size 32.0 KB
193
- Peak Allocated Bytes 0 B
194
- Overflow Count 0
195
- [ALLOC_TEMP_BatchDeleteObjects]
196
- Initial Block Size 64.0 KB
197
- Current Block Size 64.0 KB
198
- Peak Allocated Bytes 0 B
199
- Overflow Count 0
200
- [ALLOC_DEFAULT] Dual Thread Allocator
201
- Peak main deferred allocation count 326
202
- [ALLOC_BUCKET]
203
- Large Block size 4.0 MB
204
- Used Block count 1
205
- Peak Allocated bytes 1.4 MB
206
- [ALLOC_DEFAULT_MAIN]
207
- Peak usage frame count: [16.0 MB-32.0 MB]: 26546 frames
208
- Requested Block Size 16.0 MB
209
- Peak Block count 1
210
- Peak Allocated memory 23.3 MB
211
- Peak Large allocation bytes 16.0 MB
212
- [ALLOC_DEFAULT_THREAD]
213
- Peak usage frame count: [2.0 MB-4.0 MB]: 26546 frames
214
- Requested Block Size 16.0 MB
215
- Peak Block count 1
216
- Peak Allocated memory 2.5 MB
217
- Peak Large allocation bytes 0 B
218
- [ALLOC_TEMP_JOB_1_FRAME]
219
- Initial Block Size 2.0 MB
220
- Used Block Count 1
221
- Overflow Count (too large) 0
222
- Overflow Count (full) 0
223
- [ALLOC_TEMP_JOB_2_FRAMES]
224
- Initial Block Size 2.0 MB
225
- Used Block Count 1
226
- Overflow Count (too large) 0
227
- Overflow Count (full) 0
228
- [ALLOC_TEMP_JOB_4_FRAMES (JobTemp)]
229
- Initial Block Size 2.0 MB
230
- Used Block Count 2
231
- Overflow Count (too large) 0
232
- Overflow Count (full) 0
233
- [ALLOC_TEMP_JOB_ASYNC (Background)]
234
- Initial Block Size 1.0 MB
235
- Used Block Count 3
236
- Overflow Count (too large) 0
237
- Overflow Count (full) 0
238
- [ALLOC_GFX] Dual Thread Allocator
239
- Peak main deferred allocation count 0
240
- [ALLOC_BUCKET]
241
- Large Block size 4.0 MB
242
- Used Block count 1
243
- Peak Allocated bytes 1.4 MB
244
- [ALLOC_GFX_MAIN]
245
- Peak usage frame count: [32.0 KB-64.0 KB]: 26545 frames, [64.0 KB-128.0 KB]: 1 frames
246
- Requested Block Size 16.0 MB
247
- Peak Block count 1
248
- Peak Allocated memory 65.6 KB
249
- Peak Large allocation bytes 0 B
250
- [ALLOC_GFX_THREAD]
251
- Peak usage frame count: [64.0 KB-128.0 KB]: 26546 frames
252
- Requested Block Size 16.0 MB
253
- Peak Block count 1
254
- Peak Allocated memory 81.8 KB
255
- Peak Large allocation bytes 0 B
256
- [ALLOC_CACHEOBJECTS] Dual Thread Allocator
257
- Peak main deferred allocation count 0
258
- [ALLOC_BUCKET]
259
- Large Block size 4.0 MB
260
- Used Block count 1
261
- Peak Allocated bytes 1.4 MB
262
- [ALLOC_CACHEOBJECTS_MAIN]
263
- Peak usage frame count: [1.0 MB-2.0 MB]: 26545 frames, [16.0 MB-32.0 MB]: 1 frames
264
- Requested Block Size 4.0 MB
265
- Peak Block count 2
266
- Peak Allocated memory 30.6 MB
267
- Peak Large allocation bytes 24.9 MB
268
- [ALLOC_CACHEOBJECTS_THREAD]
269
- Peak usage frame count: [0.5 MB-1.0 MB]: 26545 frames, [2.0 MB-4.0 MB]: 1 frames
270
- Requested Block Size 4.0 MB
271
- Peak Block count 1
272
- Peak Allocated memory 2.6 MB
273
- Peak Large allocation bytes 0 B
274
- [ALLOC_TYPETREE] Dual Thread Allocator
275
- Peak main deferred allocation count 0
276
- [ALLOC_BUCKET]
277
- Large Block size 4.0 MB
278
- Used Block count 1
279
- Peak Allocated bytes 1.4 MB
280
- [ALLOC_TYPETREE_MAIN]
281
- Peak usage frame count: [0-1.0 KB]: 26546 frames
282
- Requested Block Size 2.0 MB
283
- Peak Block count 1
284
- Peak Allocated memory 1.0 KB
285
- Peak Large allocation bytes 0 B
286
- [ALLOC_TYPETREE_THREAD]
287
- Peak usage frame count: [4.0 KB-8.0 KB]: 26546 frames
288
- Requested Block Size 2.0 MB
289
- Peak Block count 1
290
- Peak Allocated memory 7.3 KB
291
- Peak Large allocation bytes 0 B
 
1
+ Mono path[0] = '/home/wangxso/ml-agents/trained-envs-executables/linux/Huggy/Huggy_Data/Managed'
2
+ Mono config path = '/home/wangxso/ml-agents/trained-envs-executables/linux/Huggy/Huggy_Data/MonoBleedingEdge/etc'
3
  Preloaded 'lib_burst_generated.so'
4
  Preloaded 'libgrpc_csharp_ext.x64.so'
5
+ PlayerPrefs - Creating folder: /home/wangxso/.config/unity3d/Hugging Face
6
+ PlayerPrefs - Creating folder: /home/wangxso/.config/unity3d/Hugging Face/Huggy
7
+ Unable to load player prefs
8
  Initialize engine version: 2021.3.14f1 (eee1884e7226)
9
+ [Subsystems] Discovering subsystems at path /home/wangxso/ml-agents/trained-envs-executables/linux/Huggy/Huggy_Data/UnitySubsystems
10
  Forcing GfxDevice: Null
11
  GfxDevice: creating device client; threaded=0; jobified=0
12
  NullGfxDevice:
13
  Version: NULL 1.0 [1.0]
14
  Renderer: Null Device
15
  Vendor: Unity Technologies
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  Begin MonoManager ReloadAssembly
17
+ - Completed reload, in 0.070 seconds
18
  ERROR: Shader Hidden/Universal Render Pipeline/Blit shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
19
  ERROR: Shader Hidden/Universal Render Pipeline/CopyDepth shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
20
  ERROR: Shader Hidden/Universal Render Pipeline/ScreenSpaceShadows shader is not supported on this GPU (none of subshaders/fallbacks are suitable)
 
58
  WARNING: Shader Unsupported: 'Universal Render Pipeline/Lit' - All subshaders removed
59
  WARNING: Shader Did you use #pragma only_renderers and omit this platform?
60
  WARNING: Shader If subshaders removal was intentional, you may have forgotten turning Fallback off?
61
+ UnloadTime: 0.461257 ms
62
  requesting resize 84 x 84
63
+ Setting up 14 worker threads for Enlighten.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
run_logs/timers.json CHANGED
@@ -2,135 +2,135 @@
2
  "name": "root",
3
  "gauges": {
4
  "Huggy.Policy.Entropy.mean": {
5
- "value": 1.4052338600158691,
6
- "min": 1.4052338600158691,
7
- "max": 1.4272243976593018,
8
  "count": 40
9
  },
10
  "Huggy.Policy.Entropy.sum": {
11
- "value": 70142.25,
12
- "min": 54320.82421875,
13
- "max": 72881.484375,
14
  "count": 40
15
  },
16
  "Huggy.Environment.EpisodeLength.mean": {
17
- "value": 115.66281755196304,
18
- "min": 83.26182432432432,
19
- "max": 405.7073170731707,
20
  "count": 40
21
  },
22
  "Huggy.Environment.EpisodeLength.sum": {
23
- "value": 50082.0,
24
- "min": 33268.0,
25
- "max": 50082.0,
26
  "count": 40
27
  },
28
  "Huggy.Step.mean": {
29
- "value": 1999800.0,
30
- "min": 49970.0,
31
- "max": 1999800.0,
32
  "count": 40
33
  },
34
  "Huggy.Step.sum": {
35
- "value": 1999800.0,
36
- "min": 49970.0,
37
- "max": 1999800.0,
38
  "count": 40
39
  },
40
  "Huggy.Policy.ExtrinsicValueEstimate.mean": {
41
- "value": 2.3408315181732178,
42
- "min": 0.2081862837076187,
43
- "max": 2.451524019241333,
44
  "count": 40
45
  },
46
  "Huggy.Policy.ExtrinsicValueEstimate.sum": {
47
- "value": 1013.580078125,
48
- "min": 16.863088607788086,
49
- "max": 1433.427490234375,
50
  "count": 40
51
  },
52
  "Huggy.Environment.CumulativeReward.mean": {
53
- "value": 3.63955755542112,
54
- "min": 1.857966908702144,
55
- "max": 4.021299592803456,
56
  "count": 40
57
  },
58
  "Huggy.Environment.CumulativeReward.sum": {
59
- "value": 1575.928421497345,
60
- "min": 150.49531960487366,
61
- "max": 2322.0811572670937,
62
  "count": 40
63
  },
64
  "Huggy.Policy.ExtrinsicReward.mean": {
65
- "value": 3.63955755542112,
66
- "min": 1.857966908702144,
67
- "max": 4.021299592803456,
68
  "count": 40
69
  },
70
  "Huggy.Policy.ExtrinsicReward.sum": {
71
- "value": 1575.928421497345,
72
- "min": 150.49531960487366,
73
- "max": 2322.0811572670937,
74
  "count": 40
75
  },
76
  "Huggy.Losses.PolicyLoss.mean": {
77
- "value": 0.019831607847784954,
78
- "min": 0.013048832450294867,
79
- "max": 0.019831607847784954,
80
  "count": 40
81
  },
82
  "Huggy.Losses.PolicyLoss.sum": {
83
- "value": 0.05949482354335486,
84
- "min": 0.0171435411592635,
85
- "max": 0.05949482354335486,
86
  "count": 40
87
  },
88
  "Huggy.Losses.ValueLoss.mean": {
89
- "value": 0.04458648868732982,
90
- "min": 0.020119402495523294,
91
- "max": 0.05910428791410394,
92
  "count": 40
93
  },
94
  "Huggy.Losses.ValueLoss.sum": {
95
- "value": 0.13375946606198946,
96
- "min": 0.03190994672477245,
97
- "max": 0.1773128637423118,
98
  "count": 40
99
  },
100
  "Huggy.Policy.LearningRate.mean": {
101
- "value": 3.753548748850003e-06,
102
- "min": 3.753548748850003e-06,
103
- "max": 0.00029437785187405,
104
  "count": 40
105
  },
106
  "Huggy.Policy.LearningRate.sum": {
107
- "value": 1.126064624655001e-05,
108
- "min": 1.126064624655001e-05,
109
- "max": 0.00086460406179865,
110
  "count": 40
111
  },
112
  "Huggy.Policy.Epsilon.mean": {
113
- "value": 0.10125115,
114
- "min": 0.10125115,
115
- "max": 0.19812594999999994,
116
  "count": 40
117
  },
118
  "Huggy.Policy.Epsilon.sum": {
119
- "value": 0.30375345,
120
- "min": 0.19812594999999994,
121
- "max": 0.5882013500000001,
122
  "count": 40
123
  },
124
  "Huggy.Policy.Beta.mean": {
125
- "value": 7.243238500000006e-05,
126
- "min": 7.243238500000006e-05,
127
- "max": 0.004906484905,
128
  "count": 40
129
  },
130
  "Huggy.Policy.Beta.sum": {
131
- "value": 0.0002172971550000002,
132
- "min": 0.0002172971550000002,
133
- "max": 0.014411247365000001,
134
  "count": 40
135
  },
136
  "Huggy.IsTraining.mean": {
@@ -148,67 +148,67 @@
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
- "start_time_seconds": "1691567071",
152
- "python_version": "3.10.12 (main, Jun 11 2023, 05:26:28) [GCC 11.4.0]",
153
- "command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy --no-graphics --torch --resume",
154
- "mlagents_version": "0.31.0.dev0",
155
- "mlagents_envs_version": "0.31.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
- "pytorch_version": "1.11.0+cu102",
158
- "numpy_version": "1.21.2",
159
- "end_time_seconds": "1691569863"
160
  },
161
- "total": 2791.4339863550003,
162
  "count": 1,
163
- "self": 0.4898349050008619,
164
  "children": {
165
  "run_training.setup": {
166
- "total": 0.04169404199978999,
167
  "count": 1,
168
- "self": 0.04169404199978999
169
  },
170
  "TrainerController.start_learning": {
171
- "total": 2790.902457408,
172
  "count": 1,
173
- "self": 5.468968834914904,
174
  "children": {
175
  "TrainerController._reset_env": {
176
- "total": 5.264819663000026,
177
  "count": 1,
178
- "self": 5.264819663000026
179
  },
180
  "TrainerController.advance": {
181
- "total": 2780.0338667980845,
182
- "count": 229681,
183
- "self": 5.425858913085449,
184
  "children": {
185
  "env_step": {
186
- "total": 2154.274642819981,
187
- "count": 229681,
188
- "self": 1821.9675552119868,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
- "total": 328.89261201904037,
192
- "count": 229681,
193
- "self": 18.84305205207079,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
- "total": 310.0495599669696,
197
- "count": 221213,
198
- "self": 310.0495599669696
199
  }
200
  }
201
  },
202
  "workers": {
203
- "total": 3.4144755889540193,
204
- "count": 229681,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
- "total": 2782.239273359943,
209
- "count": 229681,
210
  "is_parallel": true,
211
- "self": 1291.8389174000313,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
@@ -217,48 +217,48 @@
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
- "total": 0.000775178999901982,
221
  "count": 1,
222
  "is_parallel": true,
223
- "self": 0.0002307089998794254,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
- "total": 0.0005444700000225566,
227
  "count": 2,
228
  "is_parallel": true,
229
- "self": 0.0005444700000225566
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
- "total": 0.05665175000012823,
235
  "count": 1,
236
  "is_parallel": true,
237
- "self": 0.00032068300015453133,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
- "total": 0.00025457099991399446,
241
  "count": 1,
242
  "is_parallel": true,
243
- "self": 0.00025457099991399446
244
  },
245
  "communicator.exchange": {
246
- "total": 0.054980911000029664,
247
  "count": 1,
248
  "is_parallel": true,
249
- "self": 0.054980911000029664
250
  },
251
  "steps_from_proto": {
252
- "total": 0.0010955850000300416,
253
  "count": 1,
254
  "is_parallel": true,
255
- "self": 0.00024733099985496665,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
- "total": 0.0008482540001750749,
259
  "count": 2,
260
  "is_parallel": true,
261
- "self": 0.0008482540001750749
262
  }
263
  }
264
  }
@@ -267,34 +267,34 @@
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
- "total": 1490.4003559599118,
271
- "count": 229680,
272
  "is_parallel": true,
273
- "self": 44.33401945211472,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
- "total": 89.4841478499659,
277
- "count": 229680,
278
  "is_parallel": true,
279
- "self": 89.4841478499659
280
  },
281
  "communicator.exchange": {
282
- "total": 1245.8389012948824,
283
- "count": 229680,
284
  "is_parallel": true,
285
- "self": 1245.8389012948824
286
  },
287
  "steps_from_proto": {
288
- "total": 110.74328736294888,
289
- "count": 229680,
290
  "is_parallel": true,
291
- "self": 39.128457268825514,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
- "total": 71.61483009412336,
295
- "count": 459360,
296
  "is_parallel": true,
297
- "self": 71.61483009412336
298
  }
299
  }
300
  }
@@ -307,31 +307,31 @@
307
  }
308
  },
309
  "trainer_advance": {
310
- "total": 620.3333650650179,
311
- "count": 229681,
312
- "self": 7.644895313100278,
313
  "children": {
314
  "process_trajectory": {
315
- "total": 151.47286381891718,
316
- "count": 229681,
317
- "self": 150.01683745191804,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
- "total": 1.45602636699914,
321
  "count": 10,
322
- "self": 1.45602636699914
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
- "total": 461.21560593300046,
328
- "count": 96,
329
- "self": 398.0527268219889,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
- "total": 63.162879111011534,
333
- "count": 2880,
334
- "self": 63.162879111011534
335
  }
336
  }
337
  }
@@ -340,19 +340,19 @@
340
  }
341
  },
342
  "trainer_threads": {
343
- "total": 8.380002327612601e-07,
344
  "count": 1,
345
- "self": 8.380002327612601e-07
346
  },
347
  "TrainerController._save_models": {
348
- "total": 0.1348012739999831,
349
  "count": 1,
350
- "self": 0.0030110279999462364,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
- "total": 0.13179024600003686,
354
  "count": 1,
355
- "self": 0.13179024600003686
356
  }
357
  }
358
  }
 
2
  "name": "root",
3
  "gauges": {
4
  "Huggy.Policy.Entropy.mean": {
5
+ "value": 1.4066259860992432,
6
+ "min": 1.4066259860992432,
7
+ "max": 1.426803469657898,
8
  "count": 40
9
  },
10
  "Huggy.Policy.Entropy.sum": {
11
+ "value": 69379.015625,
12
+ "min": 68254.140625,
13
+ "max": 77601.6640625,
14
  "count": 40
15
  },
16
  "Huggy.Environment.EpisodeLength.mean": {
17
+ "value": 87.04577464788733,
18
+ "min": 81.2495894909688,
19
+ "max": 417.44166666666666,
20
  "count": 40
21
  },
22
  "Huggy.Environment.EpisodeLength.sum": {
23
+ "value": 49442.0,
24
+ "min": 48814.0,
25
+ "max": 50169.0,
26
  "count": 40
27
  },
28
  "Huggy.Step.mean": {
29
+ "value": 1999996.0,
30
+ "min": 49533.0,
31
+ "max": 1999996.0,
32
  "count": 40
33
  },
34
  "Huggy.Step.sum": {
35
+ "value": 1999996.0,
36
+ "min": 49533.0,
37
+ "max": 1999996.0,
38
  "count": 40
39
  },
40
  "Huggy.Policy.ExtrinsicValueEstimate.mean": {
41
+ "value": 2.4411799907684326,
42
+ "min": 0.09867749363183975,
43
+ "max": 2.47532057762146,
44
  "count": 40
45
  },
46
  "Huggy.Policy.ExtrinsicValueEstimate.sum": {
47
+ "value": 1386.5902099609375,
48
+ "min": 11.742621421813965,
49
+ "max": 1504.4188232421875,
50
  "count": 40
51
  },
52
  "Huggy.Environment.CumulativeReward.mean": {
53
+ "value": 3.759947028273428,
54
+ "min": 1.7275031867648374,
55
+ "max": 3.9529070295899165,
56
  "count": 40
57
  },
58
  "Huggy.Environment.CumulativeReward.sum": {
59
+ "value": 2135.649912059307,
60
+ "min": 205.57287922501564,
61
+ "max": 2330.157479941845,
62
  "count": 40
63
  },
64
  "Huggy.Policy.ExtrinsicReward.mean": {
65
+ "value": 3.759947028273428,
66
+ "min": 1.7275031867648374,
67
+ "max": 3.9529070295899165,
68
  "count": 40
69
  },
70
  "Huggy.Policy.ExtrinsicReward.sum": {
71
+ "value": 2135.649912059307,
72
+ "min": 205.57287922501564,
73
+ "max": 2330.157479941845,
74
  "count": 40
75
  },
76
  "Huggy.Losses.PolicyLoss.mean": {
77
+ "value": 0.015968902152316233,
78
+ "min": 0.012641215541710457,
79
+ "max": 0.019383551332187682,
80
  "count": 40
81
  },
82
  "Huggy.Losses.PolicyLoss.sum": {
83
+ "value": 0.047906706456948694,
84
+ "min": 0.025282431083420914,
85
+ "max": 0.05526922386682903,
86
  "count": 40
87
  },
88
  "Huggy.Losses.ValueLoss.mean": {
89
+ "value": 0.05988577306270599,
90
+ "min": 0.02010363293811679,
91
+ "max": 0.06455774539046817,
92
  "count": 40
93
  },
94
  "Huggy.Losses.ValueLoss.sum": {
95
+ "value": 0.17965731918811798,
96
+ "min": 0.04020726587623358,
97
+ "max": 0.19367323617140453,
98
  "count": 40
99
  },
100
  "Huggy.Policy.LearningRate.mean": {
101
+ "value": 3.4192488602833368e-06,
102
+ "min": 3.4192488602833368e-06,
103
+ "max": 0.00029527852657382496,
104
  "count": 40
105
  },
106
  "Huggy.Policy.LearningRate.sum": {
107
+ "value": 1.025774658085001e-05,
108
+ "min": 1.025774658085001e-05,
109
+ "max": 0.0008438953687015496,
110
  "count": 40
111
  },
112
  "Huggy.Policy.Epsilon.mean": {
113
+ "value": 0.10113971666666666,
114
+ "min": 0.10113971666666666,
115
+ "max": 0.19842617499999995,
116
  "count": 40
117
  },
118
  "Huggy.Policy.Epsilon.sum": {
119
+ "value": 0.30341914999999997,
120
+ "min": 0.20741255,
121
+ "max": 0.58129845,
122
  "count": 40
123
  },
124
  "Huggy.Policy.Beta.mean": {
125
+ "value": 6.687186166666675e-05,
126
+ "min": 6.687186166666675e-05,
127
+ "max": 0.0049214661325,
128
  "count": 40
129
  },
130
  "Huggy.Policy.Beta.sum": {
131
+ "value": 0.00020061558500000023,
132
+ "min": 0.00020061558500000023,
133
+ "max": 0.014066792655,
134
  "count": 40
135
  },
136
  "Huggy.IsTraining.mean": {
 
148
  },
149
  "metadata": {
150
  "timer_format_version": "0.1.0",
151
+ "start_time_seconds": "1761203416",
152
+ "python_version": "3.10.12 (main, Jul 5 2023, 18:54:27) [GCC 11.2.0]",
153
+ "command_line_arguments": "/home/wangxso/anaconda3/envs/huggy/bin/mlagents-learn ./config/ppo/Huggy.yaml --env=./trained-envs-executables/linux/Huggy/Huggy --run-id=Huggy2 --no-graphics",
154
+ "mlagents_version": "1.2.0.dev0",
155
+ "mlagents_envs_version": "1.2.0.dev0",
156
  "communication_protocol_version": "1.5.0",
157
+ "pytorch_version": "2.8.0+cu128",
158
+ "numpy_version": "1.23.5",
159
+ "end_time_seconds": "1761204882"
160
  },
161
+ "total": 1466.3542388949998,
162
  "count": 1,
163
+ "self": 10.002818592999802,
164
  "children": {
165
  "run_training.setup": {
166
+ "total": 0.010576472000025205,
167
  "count": 1,
168
+ "self": 0.010576472000025205
169
  },
170
  "TrainerController.start_learning": {
171
+ "total": 1456.3408438299998,
172
  "count": 1,
173
+ "self": 2.3319793060509255,
174
  "children": {
175
  "TrainerController._reset_env": {
176
+ "total": 1.8015518860000839,
177
  "count": 1,
178
+ "self": 1.8015518860000839
179
  },
180
  "TrainerController.advance": {
181
+ "total": 1452.1483540049492,
182
+ "count": 232089,
183
+ "self": 2.1772256279562043,
184
  "children": {
185
  "env_step": {
186
+ "total": 1211.674216274936,
187
+ "count": 232089,
188
+ "self": 793.1521938349478,
189
  "children": {
190
  "SubprocessEnvManager._take_step": {
191
+ "total": 416.899643371005,
192
+ "count": 232089,
193
+ "self": 8.653769012024554,
194
  "children": {
195
  "TorchPolicy.evaluate": {
196
+ "total": 408.2458743589805,
197
+ "count": 222847,
198
+ "self": 408.2458743589805
199
  }
200
  }
201
  },
202
  "workers": {
203
+ "total": 1.622379068983264,
204
+ "count": 232089,
205
  "self": 0.0,
206
  "children": {
207
  "worker_root": {
208
+ "total": 1451.105851403961,
209
+ "count": 232089,
210
  "is_parallel": true,
211
+ "self": 787.6013764139241,
212
  "children": {
213
  "run_training.setup": {
214
  "total": 0.0,
 
217
  "self": 0.0,
218
  "children": {
219
  "steps_from_proto": {
220
+ "total": 0.0006009019999737575,
221
  "count": 1,
222
  "is_parallel": true,
223
+ "self": 0.00015025999982754001,
224
  "children": {
225
  "_process_rank_one_or_two_observation": {
226
+ "total": 0.00045064200014621747,
227
  "count": 2,
228
  "is_parallel": true,
229
+ "self": 0.00045064200014621747
230
  }
231
  }
232
  },
233
  "UnityEnvironment.step": {
234
+ "total": 0.011328645999924447,
235
  "count": 1,
236
  "is_parallel": true,
237
+ "self": 0.00011084399989158555,
238
  "children": {
239
  "UnityEnvironment._generate_step_input": {
240
+ "total": 9.075299999494746e-05,
241
  "count": 1,
242
  "is_parallel": true,
243
+ "self": 9.075299999494746e-05
244
  },
245
  "communicator.exchange": {
246
+ "total": 0.010878527000045324,
247
  "count": 1,
248
  "is_parallel": true,
249
+ "self": 0.010878527000045324
250
  },
251
  "steps_from_proto": {
252
+ "total": 0.0002485219999925903,
253
  "count": 1,
254
  "is_parallel": true,
255
+ "self": 8.557299997846712e-05,
256
  "children": {
257
  "_process_rank_one_or_two_observation": {
258
+ "total": 0.00016294900001412316,
259
  "count": 2,
260
  "is_parallel": true,
261
+ "self": 0.00016294900001412316
262
  }
263
  }
264
  }
 
267
  }
268
  },
269
  "UnityEnvironment.step": {
270
+ "total": 663.5044749900369,
271
+ "count": 232088,
272
  "is_parallel": true,
273
+ "self": 12.778950026080906,
274
  "children": {
275
  "UnityEnvironment._generate_step_input": {
276
+ "total": 20.749552779011083,
277
+ "count": 232088,
278
  "is_parallel": true,
279
+ "self": 20.749552779011083
280
  },
281
  "communicator.exchange": {
282
+ "total": 602.481897822021,
283
+ "count": 232088,
284
  "is_parallel": true,
285
+ "self": 602.481897822021
286
  },
287
  "steps_from_proto": {
288
+ "total": 27.494074362923925,
289
+ "count": 232088,
290
  "is_parallel": true,
291
+ "self": 8.758737116025713,
292
  "children": {
293
  "_process_rank_one_or_two_observation": {
294
+ "total": 18.73533724689821,
295
+ "count": 464176,
296
  "is_parallel": true,
297
+ "self": 18.73533724689821
298
  }
299
  }
300
  }
 
307
  }
308
  },
309
  "trainer_advance": {
310
+ "total": 238.29691210205704,
311
+ "count": 232089,
312
+ "self": 3.6165667471407232,
313
  "children": {
314
  "process_trajectory": {
315
+ "total": 91.49027175291599,
316
+ "count": 232089,
317
+ "self": 90.76382075791548,
318
  "children": {
319
  "RLTrainer._checkpoint": {
320
+ "total": 0.7264509950005049,
321
  "count": 10,
322
+ "self": 0.7264509950005049
323
  }
324
  }
325
  },
326
  "_update_policy": {
327
+ "total": 143.19007360200033,
328
+ "count": 97,
329
+ "self": 91.25993330401047,
330
  "children": {
331
  "TorchPPOOptimizer.update": {
332
+ "total": 51.93014029798985,
333
+ "count": 2910,
334
+ "self": 51.93014029798985
335
  }
336
  }
337
  }
 
340
  }
341
  },
342
  "trainer_threads": {
343
+ "total": 4.4199987314641476e-07,
344
  "count": 1,
345
+ "self": 4.4199987314641476e-07
346
  },
347
  "TrainerController._save_models": {
348
+ "total": 0.058958190999874205,
349
  "count": 1,
350
+ "self": 0.0009660860000622051,
351
  "children": {
352
  "RLTrainer._checkpoint": {
353
+ "total": 0.057992104999812,
354
  "count": 1,
355
+ "self": 0.057992104999812
356
  }
357
  }
358
  }
run_logs/training_status.json CHANGED
@@ -2,127 +2,118 @@
2
  "Huggy": {
3
  "checkpoints": [
4
  {
5
- "steps": 16924,
6
- "file_path": "results/Huggy/Huggy/Huggy-16924.onnx",
7
- "reward": 1.756937677330441,
8
- "creation_time": 1691567052.996866,
9
  "auxillary_file_paths": [
10
- "results/Huggy/Huggy/Huggy-16924.pt"
11
  ]
12
  },
13
  {
14
- "steps": 199739,
15
- "file_path": "results/Huggy/Huggy/Huggy-199739.onnx",
16
- "reward": 3.4687055515337595,
17
- "creation_time": 1691567319.7134223,
18
  "auxillary_file_paths": [
19
- "results/Huggy/Huggy/Huggy-199739.pt"
20
  ]
21
  },
22
  {
23
- "steps": 399720,
24
- "file_path": "results/Huggy/Huggy/Huggy-399720.onnx",
25
- "reward": 3.605525080217134,
26
- "creation_time": 1691567590.5574648,
27
  "auxillary_file_paths": [
28
- "results/Huggy/Huggy/Huggy-399720.pt"
29
  ]
30
  },
31
  {
32
- "steps": 599998,
33
- "file_path": "results/Huggy/Huggy/Huggy-599998.onnx",
34
- "reward": 3.9826613015988293,
35
- "creation_time": 1691567858.6299968,
36
  "auxillary_file_paths": [
37
- "results/Huggy/Huggy/Huggy-599998.pt"
38
  ]
39
  },
40
  {
41
- "steps": 799972,
42
- "file_path": "results/Huggy/Huggy/Huggy-799972.onnx",
43
- "reward": 3.7262471199035643,
44
- "creation_time": 1691568144.0322392,
45
  "auxillary_file_paths": [
46
- "results/Huggy/Huggy/Huggy-799972.pt"
47
  ]
48
  },
49
  {
50
- "steps": 999903,
51
- "file_path": "results/Huggy/Huggy/Huggy-999903.onnx",
52
- "reward": 3.9235518959852365,
53
- "creation_time": 1691568420.4665353,
54
  "auxillary_file_paths": [
55
- "results/Huggy/Huggy/Huggy-999903.pt"
56
- ]
57
- },
58
- {
59
- "steps": 1199993,
60
- "file_path": "results/Huggy/Huggy/Huggy-1199993.onnx",
61
- "reward": 3.5797291318981017,
62
- "creation_time": 1691568708.5528545,
63
- "auxillary_file_paths": [
64
- "results/Huggy/Huggy/Huggy-1199993.pt"
65
  ]
66
  },
67
  {
68
  "steps": 1399940,
69
- "file_path": "results/Huggy/Huggy/Huggy-1399940.onnx",
70
- "reward": 3.678971702402288,
71
- "creation_time": 1691568990.9485593,
72
  "auxillary_file_paths": [
73
- "results/Huggy/Huggy/Huggy-1399940.pt"
74
  ]
75
  },
76
  {
77
- "steps": 1599701,
78
- "file_path": "results/Huggy/Huggy/Huggy-1599701.onnx",
79
- "reward": 3.514644799681453,
80
- "creation_time": 1691569272.379971,
81
  "auxillary_file_paths": [
82
- "results/Huggy/Huggy/Huggy-1599701.pt"
83
  ]
84
  },
85
  {
86
- "steps": 1799435,
87
- "file_path": "results/Huggy/Huggy/Huggy-1799435.onnx",
88
- "reward": 3.458062550218979,
89
- "creation_time": 1691569557.544697,
90
  "auxillary_file_paths": [
91
- "results/Huggy/Huggy/Huggy-1799435.pt"
92
  ]
93
  },
94
  {
95
- "steps": 1999800,
96
- "file_path": "results/Huggy/Huggy/Huggy-1999800.onnx",
97
- "reward": 4.050866290142662,
98
- "creation_time": 1691569862.6583805,
99
  "auxillary_file_paths": [
100
- "results/Huggy/Huggy/Huggy-1999800.pt"
101
  ]
102
  },
103
  {
104
- "steps": 2000550,
105
- "file_path": "results/Huggy/Huggy/Huggy-2000550.onnx",
106
- "reward": 3.8553753327100706,
107
- "creation_time": 1691569862.8188043,
108
  "auxillary_file_paths": [
109
- "results/Huggy/Huggy/Huggy-2000550.pt"
110
  ]
111
  }
112
  ],
113
  "final_checkpoint": {
114
- "steps": 2000550,
115
- "file_path": "results/Huggy/Huggy.onnx",
116
- "reward": 3.8553753327100706,
117
- "creation_time": 1691569862.8188043,
118
  "auxillary_file_paths": [
119
- "results/Huggy/Huggy/Huggy-2000550.pt"
120
  ]
121
  }
122
  },
123
  "metadata": {
124
  "stats_format_version": "0.3.0",
125
- "mlagents_version": "0.31.0.dev0",
126
- "torch_version": "1.11.0+cu102"
127
  }
128
  }
 
2
  "Huggy": {
3
  "checkpoints": [
4
  {
5
+ "steps": 199867,
6
+ "file_path": "results/Huggy2/Huggy/Huggy-199867.onnx",
7
+ "reward": 3.2746736019376725,
8
+ "creation_time": 1761203557.2784374,
9
  "auxillary_file_paths": [
10
+ "results/Huggy2/Huggy/Huggy-199867.pt"
11
  ]
12
  },
13
  {
14
+ "steps": 399973,
15
+ "file_path": "results/Huggy2/Huggy/Huggy-399973.onnx",
16
+ "reward": 3.471751812965639,
17
+ "creation_time": 1761203698.933577,
18
  "auxillary_file_paths": [
19
+ "results/Huggy2/Huggy/Huggy-399973.pt"
20
  ]
21
  },
22
  {
23
+ "steps": 599911,
24
+ "file_path": "results/Huggy2/Huggy/Huggy-599911.onnx",
25
+ "reward": 3.7677173614501953,
26
+ "creation_time": 1761203840.117672,
27
  "auxillary_file_paths": [
28
+ "results/Huggy2/Huggy/Huggy-599911.pt"
29
  ]
30
  },
31
  {
32
+ "steps": 799888,
33
+ "file_path": "results/Huggy2/Huggy/Huggy-799888.onnx",
34
+ "reward": 4.034348468334354,
35
+ "creation_time": 1761203982.7505534,
36
  "auxillary_file_paths": [
37
+ "results/Huggy2/Huggy/Huggy-799888.pt"
38
  ]
39
  },
40
  {
41
+ "steps": 999982,
42
+ "file_path": "results/Huggy2/Huggy/Huggy-999982.onnx",
43
+ "reward": 3.9394379349442215,
44
+ "creation_time": 1761204126.683781,
45
  "auxillary_file_paths": [
46
+ "results/Huggy2/Huggy/Huggy-999982.pt"
47
  ]
48
  },
49
  {
50
+ "steps": 1199995,
51
+ "file_path": "results/Huggy2/Huggy/Huggy-1199995.onnx",
52
+ "reward": 3.8306074531182,
53
+ "creation_time": 1761204273.2620282,
54
  "auxillary_file_paths": [
55
+ "results/Huggy2/Huggy/Huggy-1199995.pt"
 
 
 
 
 
 
 
 
 
56
  ]
57
  },
58
  {
59
  "steps": 1399940,
60
+ "file_path": "results/Huggy2/Huggy/Huggy-1399940.onnx",
61
+ "reward": 3.957194642951855,
62
+ "creation_time": 1761204423.3528085,
63
  "auxillary_file_paths": [
64
+ "results/Huggy2/Huggy/Huggy-1399940.pt"
65
  ]
66
  },
67
  {
68
+ "steps": 1599966,
69
+ "file_path": "results/Huggy2/Huggy/Huggy-1599966.onnx",
70
+ "reward": 3.8169545004280603,
71
+ "creation_time": 1761204570.9987407,
72
  "auxillary_file_paths": [
73
+ "results/Huggy2/Huggy/Huggy-1599966.pt"
74
  ]
75
  },
76
  {
77
+ "steps": 1799988,
78
+ "file_path": "results/Huggy2/Huggy/Huggy-1799988.onnx",
79
+ "reward": 3.7128463983535767,
80
+ "creation_time": 1761204722.7814953,
81
  "auxillary_file_paths": [
82
+ "results/Huggy2/Huggy/Huggy-1799988.pt"
83
  ]
84
  },
85
  {
86
+ "steps": 1999996,
87
+ "file_path": "results/Huggy2/Huggy/Huggy-1999996.onnx",
88
+ "reward": 3.3463553537925086,
89
+ "creation_time": 1761204872.350479,
90
  "auxillary_file_paths": [
91
+ "results/Huggy2/Huggy/Huggy-1999996.pt"
92
  ]
93
  },
94
  {
95
+ "steps": 2000076,
96
+ "file_path": "results/Huggy2/Huggy/Huggy-2000076.onnx",
97
+ "reward": 3.4634731245040893,
98
+ "creation_time": 1761204872.4119837,
99
  "auxillary_file_paths": [
100
+ "results/Huggy2/Huggy/Huggy-2000076.pt"
101
  ]
102
  }
103
  ],
104
  "final_checkpoint": {
105
+ "steps": 2000076,
106
+ "file_path": "results/Huggy2/Huggy.onnx",
107
+ "reward": 3.4634731245040893,
108
+ "creation_time": 1761204872.4119837,
109
  "auxillary_file_paths": [
110
+ "results/Huggy2/Huggy/Huggy-2000076.pt"
111
  ]
112
  }
113
  },
114
  "metadata": {
115
  "stats_format_version": "0.3.0",
116
+ "mlagents_version": "1.2.0.dev0",
117
+ "torch_version": "2.8.0+cu128"
118
  }
119
  }