Liberow commited on
Commit
00d89ed
·
verified ·
1 Parent(s): ed96d25

Upload folder using huggingface_hub

Browse files
Files changed (37) hide show
  1. .gitattributes +2 -0
  2. smolvla_omy/checkpoints/005000/pretrained_model/config.json +83 -0
  3. smolvla_omy/checkpoints/005000/pretrained_model/model.safetensors +3 -0
  4. smolvla_omy/checkpoints/005000/pretrained_model/train_config.json +195 -0
  5. smolvla_omy/checkpoints/005000/training_state/optimizer_param_groups.json +526 -0
  6. smolvla_omy/checkpoints/005000/training_state/optimizer_state.safetensors +3 -0
  7. smolvla_omy/checkpoints/005000/training_state/rng_state.safetensors +3 -0
  8. smolvla_omy/checkpoints/005000/training_state/scheduler_state.json +15 -0
  9. smolvla_omy/checkpoints/005000/training_state/training_step.json +3 -0
  10. smolvla_omy/wandb/debug-internal.log +8 -0
  11. smolvla_omy/wandb/debug.log +21 -0
  12. smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/output.log +631 -0
  13. smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/requirements.txt +173 -0
  14. smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/wandb-metadata.json +64 -0
  15. smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-core.log +6 -0
  16. smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-internal.log +8 -0
  17. smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug.log +21 -0
  18. smolvla_omy/wandb/run-20251106_162149-8gs9of7r/run-8gs9of7r.wandb +3 -0
  19. smolvla_omy_2000_steps/checkpoints/002000/pretrained_model/config.json +83 -0
  20. smolvla_omy_2000_steps/checkpoints/002000/pretrained_model/model.safetensors +3 -0
  21. smolvla_omy_2000_steps/checkpoints/002000/pretrained_model/train_config.json +195 -0
  22. smolvla_omy_2000_steps/checkpoints/002000/training_state/optimizer_param_groups.json +526 -0
  23. smolvla_omy_2000_steps/checkpoints/002000/training_state/optimizer_state.safetensors +3 -0
  24. smolvla_omy_2000_steps/checkpoints/002000/training_state/rng_state.safetensors +3 -0
  25. smolvla_omy_2000_steps/checkpoints/002000/training_state/scheduler_state.json +15 -0
  26. smolvla_omy_2000_steps/checkpoints/002000/training_state/training_step.json +3 -0
  27. smolvla_omy_2000_steps/wandb/debug-internal.log +11 -0
  28. smolvla_omy_2000_steps/wandb/debug.log +23 -0
  29. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/config.yaml +232 -0
  30. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/output.log +137 -0
  31. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/requirements.txt +172 -0
  32. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/wandb-metadata.json +64 -0
  33. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/wandb-summary.json +1 -0
  34. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/logs/debug-core.log +14 -0
  35. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/logs/debug-internal.log +11 -0
  36. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/logs/debug.log +23 -0
  37. smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/run-p7vw7943.wandb +3 -0
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ smolvla_omy/wandb/run-20251106_162149-8gs9of7r/run-8gs9of7r.wandb filter=lfs diff=lfs merge=lfs -text
37
+ smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/run-p7vw7943.wandb filter=lfs diff=lfs merge=lfs -text
smolvla_omy/checkpoints/005000/pretrained_model/config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "normalization_mapping": {
5
+ "VISUAL": "IDENTITY",
6
+ "STATE": "MEAN_STD",
7
+ "ACTION": "MEAN_STD"
8
+ },
9
+ "input_features": {
10
+ "observation.image": {
11
+ "type": "VISUAL",
12
+ "shape": [
13
+ 3,
14
+ 256,
15
+ 256
16
+ ]
17
+ },
18
+ "observation.wrist_image": {
19
+ "type": "VISUAL",
20
+ "shape": [
21
+ 3,
22
+ 256,
23
+ 256
24
+ ]
25
+ },
26
+ "observation.state": {
27
+ "type": "STATE",
28
+ "shape": [
29
+ 6
30
+ ]
31
+ }
32
+ },
33
+ "output_features": {
34
+ "action": {
35
+ "type": "ACTION",
36
+ "shape": [
37
+ 7
38
+ ]
39
+ }
40
+ },
41
+ "device": "cuda",
42
+ "use_amp": false,
43
+ "chunk_size": 5,
44
+ "n_action_steps": 5,
45
+ "max_state_dim": 32,
46
+ "max_action_dim": 32,
47
+ "resize_imgs_with_padding": [
48
+ 512,
49
+ 512
50
+ ],
51
+ "empty_cameras": 0,
52
+ "adapt_to_pi_aloha": false,
53
+ "use_delta_joint_actions_aloha": false,
54
+ "tokenizer_max_length": 48,
55
+ "num_steps": 10,
56
+ "use_cache": true,
57
+ "freeze_vision_encoder": true,
58
+ "train_expert_only": true,
59
+ "train_state_proj": true,
60
+ "optimizer_lr": 0.0001,
61
+ "optimizer_betas": [
62
+ 0.9,
63
+ 0.95
64
+ ],
65
+ "optimizer_eps": 1e-08,
66
+ "optimizer_weight_decay": 1e-10,
67
+ "optimizer_grad_clip_norm": 10,
68
+ "scheduler_warmup_steps": 1000,
69
+ "scheduler_decay_steps": 30000,
70
+ "scheduler_decay_lr": 2.5e-06,
71
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
72
+ "load_vlm_weights": false,
73
+ "add_image_special_tokens": false,
74
+ "attention_mode": "cross_attn",
75
+ "prefix_length": -1,
76
+ "pad_language_to": "longest",
77
+ "num_expert_layers": -1,
78
+ "num_vlm_layers": 16,
79
+ "self_attn_every_n_layers": 2,
80
+ "expert_width_multiplier": 0.75,
81
+ "min_period": 0.004,
82
+ "max_period": 4.0
83
+ }
smolvla_omy/checkpoints/005000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:794c736c8a4faf3263611cc5cf8327f2fe3772773c68b23757596bcfb2d99580
3
+ size 1197790032
smolvla_omy/checkpoints/005000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "omy_pnp_language",
4
+ "root": "./demo_data_language",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "revision": null,
64
+ "use_imagenet_stats": true,
65
+ "video_backend": "torchcodec"
66
+ },
67
+ "env": null,
68
+ "policy": {
69
+ "type": "smolvla",
70
+ "n_obs_steps": 1,
71
+ "normalization_mapping": {
72
+ "VISUAL": "IDENTITY",
73
+ "STATE": "MEAN_STD",
74
+ "ACTION": "MEAN_STD"
75
+ },
76
+ "input_features": {
77
+ "observation.image": {
78
+ "type": "VISUAL",
79
+ "shape": [
80
+ 3,
81
+ 256,
82
+ 256
83
+ ]
84
+ },
85
+ "observation.wrist_image": {
86
+ "type": "VISUAL",
87
+ "shape": [
88
+ 3,
89
+ 256,
90
+ 256
91
+ ]
92
+ },
93
+ "observation.state": {
94
+ "type": "STATE",
95
+ "shape": [
96
+ 6
97
+ ]
98
+ }
99
+ },
100
+ "output_features": {
101
+ "action": {
102
+ "type": "ACTION",
103
+ "shape": [
104
+ 7
105
+ ]
106
+ }
107
+ },
108
+ "device": "cuda",
109
+ "use_amp": false,
110
+ "chunk_size": 5,
111
+ "n_action_steps": 5,
112
+ "max_state_dim": 32,
113
+ "max_action_dim": 32,
114
+ "resize_imgs_with_padding": [
115
+ 512,
116
+ 512
117
+ ],
118
+ "empty_cameras": 0,
119
+ "adapt_to_pi_aloha": false,
120
+ "use_delta_joint_actions_aloha": false,
121
+ "tokenizer_max_length": 48,
122
+ "num_steps": 10,
123
+ "use_cache": true,
124
+ "freeze_vision_encoder": true,
125
+ "train_expert_only": true,
126
+ "train_state_proj": true,
127
+ "optimizer_lr": 0.0001,
128
+ "optimizer_betas": [
129
+ 0.9,
130
+ 0.95
131
+ ],
132
+ "optimizer_eps": 1e-08,
133
+ "optimizer_weight_decay": 1e-10,
134
+ "optimizer_grad_clip_norm": 10,
135
+ "scheduler_warmup_steps": 1000,
136
+ "scheduler_decay_steps": 30000,
137
+ "scheduler_decay_lr": 2.5e-06,
138
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
139
+ "load_vlm_weights": false,
140
+ "add_image_special_tokens": false,
141
+ "attention_mode": "cross_attn",
142
+ "prefix_length": -1,
143
+ "pad_language_to": "longest",
144
+ "num_expert_layers": -1,
145
+ "num_vlm_layers": 16,
146
+ "self_attn_every_n_layers": 2,
147
+ "expert_width_multiplier": 0.75,
148
+ "min_period": 0.004,
149
+ "max_period": 4.0
150
+ },
151
+ "output_dir": "ckpt/smolvla_omy",
152
+ "job_name": "smolvla_omy",
153
+ "resume": false,
154
+ "seed": 42,
155
+ "num_workers": 8,
156
+ "batch_size": 16,
157
+ "steps": 20000,
158
+ "eval_freq": -1,
159
+ "log_freq": 50,
160
+ "save_checkpoint": true,
161
+ "save_freq": 5000,
162
+ "use_policy_training_preset": true,
163
+ "optimizer": {
164
+ "type": "adamw",
165
+ "lr": 0.0001,
166
+ "weight_decay": 1e-10,
167
+ "grad_clip_norm": 10,
168
+ "betas": [
169
+ 0.9,
170
+ 0.95
171
+ ],
172
+ "eps": 1e-08
173
+ },
174
+ "scheduler": {
175
+ "type": "cosine_decay_with_warmup",
176
+ "num_warmup_steps": 1000,
177
+ "num_decay_steps": 30000,
178
+ "peak_lr": 0.0001,
179
+ "decay_lr": 2.5e-06
180
+ },
181
+ "eval": {
182
+ "n_episodes": 50,
183
+ "batch_size": 50,
184
+ "use_async_envs": false
185
+ },
186
+ "wandb": {
187
+ "enable": true,
188
+ "disable_artifact": true,
189
+ "project": "smolvla_omy",
190
+ "entity": "ai-liber",
191
+ "notes": null,
192
+ "run_id": null,
193
+ "mode": null
194
+ }
195
+ }
smolvla_omy/checkpoints/005000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,526 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 9.34687384344914e-05,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "foreach": null,
12
+ "maximize": false,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "initial_lr": 0.0001,
17
+ "params": [
18
+ 0,
19
+ 1,
20
+ 2,
21
+ 3,
22
+ 4,
23
+ 5,
24
+ 6,
25
+ 7,
26
+ 8,
27
+ 9,
28
+ 10,
29
+ 11,
30
+ 12,
31
+ 13,
32
+ 14,
33
+ 15,
34
+ 16,
35
+ 17,
36
+ 18,
37
+ 19,
38
+ 20,
39
+ 21,
40
+ 22,
41
+ 23,
42
+ 24,
43
+ 25,
44
+ 26,
45
+ 27,
46
+ 28,
47
+ 29,
48
+ 30,
49
+ 31,
50
+ 32,
51
+ 33,
52
+ 34,
53
+ 35,
54
+ 36,
55
+ 37,
56
+ 38,
57
+ 39,
58
+ 40,
59
+ 41,
60
+ 42,
61
+ 43,
62
+ 44,
63
+ 45,
64
+ 46,
65
+ 47,
66
+ 48,
67
+ 49,
68
+ 50,
69
+ 51,
70
+ 52,
71
+ 53,
72
+ 54,
73
+ 55,
74
+ 56,
75
+ 57,
76
+ 58,
77
+ 59,
78
+ 60,
79
+ 61,
80
+ 62,
81
+ 63,
82
+ 64,
83
+ 65,
84
+ 66,
85
+ 67,
86
+ 68,
87
+ 69,
88
+ 70,
89
+ 71,
90
+ 72,
91
+ 73,
92
+ 74,
93
+ 75,
94
+ 76,
95
+ 77,
96
+ 78,
97
+ 79,
98
+ 80,
99
+ 81,
100
+ 82,
101
+ 83,
102
+ 84,
103
+ 85,
104
+ 86,
105
+ 87,
106
+ 88,
107
+ 89,
108
+ 90,
109
+ 91,
110
+ 92,
111
+ 93,
112
+ 94,
113
+ 95,
114
+ 96,
115
+ 97,
116
+ 98,
117
+ 99,
118
+ 100,
119
+ 101,
120
+ 102,
121
+ 103,
122
+ 104,
123
+ 105,
124
+ 106,
125
+ 107,
126
+ 108,
127
+ 109,
128
+ 110,
129
+ 111,
130
+ 112,
131
+ 113,
132
+ 114,
133
+ 115,
134
+ 116,
135
+ 117,
136
+ 118,
137
+ 119,
138
+ 120,
139
+ 121,
140
+ 122,
141
+ 123,
142
+ 124,
143
+ 125,
144
+ 126,
145
+ 127,
146
+ 128,
147
+ 129,
148
+ 130,
149
+ 131,
150
+ 132,
151
+ 133,
152
+ 134,
153
+ 135,
154
+ 136,
155
+ 137,
156
+ 138,
157
+ 139,
158
+ 140,
159
+ 141,
160
+ 142,
161
+ 143,
162
+ 144,
163
+ 145,
164
+ 146,
165
+ 147,
166
+ 148,
167
+ 149,
168
+ 150,
169
+ 151,
170
+ 152,
171
+ 153,
172
+ 154,
173
+ 155,
174
+ 156,
175
+ 157,
176
+ 158,
177
+ 159,
178
+ 160,
179
+ 161,
180
+ 162,
181
+ 163,
182
+ 164,
183
+ 165,
184
+ 166,
185
+ 167,
186
+ 168,
187
+ 169,
188
+ 170,
189
+ 171,
190
+ 172,
191
+ 173,
192
+ 174,
193
+ 175,
194
+ 176,
195
+ 177,
196
+ 178,
197
+ 179,
198
+ 180,
199
+ 181,
200
+ 182,
201
+ 183,
202
+ 184,
203
+ 185,
204
+ 186,
205
+ 187,
206
+ 188,
207
+ 189,
208
+ 190,
209
+ 191,
210
+ 192,
211
+ 193,
212
+ 194,
213
+ 195,
214
+ 196,
215
+ 197,
216
+ 198,
217
+ 199,
218
+ 200,
219
+ 201,
220
+ 202,
221
+ 203,
222
+ 204,
223
+ 205,
224
+ 206,
225
+ 207,
226
+ 208,
227
+ 209,
228
+ 210,
229
+ 211,
230
+ 212,
231
+ 213,
232
+ 214,
233
+ 215,
234
+ 216,
235
+ 217,
236
+ 218,
237
+ 219,
238
+ 220,
239
+ 221,
240
+ 222,
241
+ 223,
242
+ 224,
243
+ 225,
244
+ 226,
245
+ 227,
246
+ 228,
247
+ 229,
248
+ 230,
249
+ 231,
250
+ 232,
251
+ 233,
252
+ 234,
253
+ 235,
254
+ 236,
255
+ 237,
256
+ 238,
257
+ 239,
258
+ 240,
259
+ 241,
260
+ 242,
261
+ 243,
262
+ 244,
263
+ 245,
264
+ 246,
265
+ 247,
266
+ 248,
267
+ 249,
268
+ 250,
269
+ 251,
270
+ 252,
271
+ 253,
272
+ 254,
273
+ 255,
274
+ 256,
275
+ 257,
276
+ 258,
277
+ 259,
278
+ 260,
279
+ 261,
280
+ 262,
281
+ 263,
282
+ 264,
283
+ 265,
284
+ 266,
285
+ 267,
286
+ 268,
287
+ 269,
288
+ 270,
289
+ 271,
290
+ 272,
291
+ 273,
292
+ 274,
293
+ 275,
294
+ 276,
295
+ 277,
296
+ 278,
297
+ 279,
298
+ 280,
299
+ 281,
300
+ 282,
301
+ 283,
302
+ 284,
303
+ 285,
304
+ 286,
305
+ 287,
306
+ 288,
307
+ 289,
308
+ 290,
309
+ 291,
310
+ 292,
311
+ 293,
312
+ 294,
313
+ 295,
314
+ 296,
315
+ 297,
316
+ 298,
317
+ 299,
318
+ 300,
319
+ 301,
320
+ 302,
321
+ 303,
322
+ 304,
323
+ 305,
324
+ 306,
325
+ 307,
326
+ 308,
327
+ 309,
328
+ 310,
329
+ 311,
330
+ 312,
331
+ 313,
332
+ 314,
333
+ 315,
334
+ 316,
335
+ 317,
336
+ 318,
337
+ 319,
338
+ 320,
339
+ 321,
340
+ 322,
341
+ 323,
342
+ 324,
343
+ 325,
344
+ 326,
345
+ 327,
346
+ 328,
347
+ 329,
348
+ 330,
349
+ 331,
350
+ 332,
351
+ 333,
352
+ 334,
353
+ 335,
354
+ 336,
355
+ 337,
356
+ 338,
357
+ 339,
358
+ 340,
359
+ 341,
360
+ 342,
361
+ 343,
362
+ 344,
363
+ 345,
364
+ 346,
365
+ 347,
366
+ 348,
367
+ 349,
368
+ 350,
369
+ 351,
370
+ 352,
371
+ 353,
372
+ 354,
373
+ 355,
374
+ 356,
375
+ 357,
376
+ 358,
377
+ 359,
378
+ 360,
379
+ 361,
380
+ 362,
381
+ 363,
382
+ 364,
383
+ 365,
384
+ 366,
385
+ 367,
386
+ 368,
387
+ 369,
388
+ 370,
389
+ 371,
390
+ 372,
391
+ 373,
392
+ 374,
393
+ 375,
394
+ 376,
395
+ 377,
396
+ 378,
397
+ 379,
398
+ 380,
399
+ 381,
400
+ 382,
401
+ 383,
402
+ 384,
403
+ 385,
404
+ 386,
405
+ 387,
406
+ 388,
407
+ 389,
408
+ 390,
409
+ 391,
410
+ 392,
411
+ 393,
412
+ 394,
413
+ 395,
414
+ 396,
415
+ 397,
416
+ 398,
417
+ 399,
418
+ 400,
419
+ 401,
420
+ 402,
421
+ 403,
422
+ 404,
423
+ 405,
424
+ 406,
425
+ 407,
426
+ 408,
427
+ 409,
428
+ 410,
429
+ 411,
430
+ 412,
431
+ 413,
432
+ 414,
433
+ 415,
434
+ 416,
435
+ 417,
436
+ 418,
437
+ 419,
438
+ 420,
439
+ 421,
440
+ 422,
441
+ 423,
442
+ 424,
443
+ 425,
444
+ 426,
445
+ 427,
446
+ 428,
447
+ 429,
448
+ 430,
449
+ 431,
450
+ 432,
451
+ 433,
452
+ 434,
453
+ 435,
454
+ 436,
455
+ 437,
456
+ 438,
457
+ 439,
458
+ 440,
459
+ 441,
460
+ 442,
461
+ 443,
462
+ 444,
463
+ 445,
464
+ 446,
465
+ 447,
466
+ 448,
467
+ 449,
468
+ 450,
469
+ 451,
470
+ 452,
471
+ 453,
472
+ 454,
473
+ 455,
474
+ 456,
475
+ 457,
476
+ 458,
477
+ 459,
478
+ 460,
479
+ 461,
480
+ 462,
481
+ 463,
482
+ 464,
483
+ 465,
484
+ 466,
485
+ 467,
486
+ 468,
487
+ 469,
488
+ 470,
489
+ 471,
490
+ 472,
491
+ 473,
492
+ 474,
493
+ 475,
494
+ 476,
495
+ 477,
496
+ 478,
497
+ 479,
498
+ 480,
499
+ 481,
500
+ 482,
501
+ 483,
502
+ 484,
503
+ 485,
504
+ 486,
505
+ 487,
506
+ 488,
507
+ 489,
508
+ 490,
509
+ 491,
510
+ 492,
511
+ 493,
512
+ 494,
513
+ 495,
514
+ 496,
515
+ 497,
516
+ 498,
517
+ 499,
518
+ 500,
519
+ 501,
520
+ 502,
521
+ 503,
522
+ 504,
523
+ 505
524
+ ]
525
+ }
526
+ ]
smolvla_omy/checkpoints/005000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28dbf586dce9a55eb5a3987438be5f06488031966430c8e6279de94992d23109
3
+ size 412659164
smolvla_omy/checkpoints/005000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afe5806563981ea1faee236e400cf9c5eca89fef62fed6aa1a580b085a4cb11e
3
+ size 15708
smolvla_omy/checkpoints/005000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 5000,
6
+ "verbose": false,
7
+ "_step_count": 5001,
8
+ "_get_lr_called_within_step": false,
9
+ "_last_lr": [
10
+ 9.34687384344914e-05
11
+ ],
12
+ "lr_lambdas": [
13
+ null
14
+ ]
15
+ }
smolvla_omy/checkpoints/005000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 5000
3
+ }
smolvla_omy/wandb/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-06T16:21:49.810658746+08:00","level":"INFO","msg":"stream: starting","core version":"0.22.3"}
2
+ {"time":"2025-11-06T16:21:50.381496059+08:00","level":"INFO","msg":"stream: created new stream","id":"8gs9of7r"}
3
+ {"time":"2025-11-06T16:21:50.381670461+08:00","level":"INFO","msg":"handler: started","stream_id":"8gs9of7r"}
4
+ {"time":"2025-11-06T16:21:50.38195187+08:00","level":"INFO","msg":"stream: started","id":"8gs9of7r"}
5
+ {"time":"2025-11-06T16:21:50.381970995+08:00","level":"INFO","msg":"writer: started","stream_id":"8gs9of7r"}
6
+ {"time":"2025-11-06T16:21:50.381975054+08:00","level":"INFO","msg":"sender: started","stream_id":"8gs9of7r"}
7
+ {"time":"2025-11-06T18:28:51.825582909+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
8
+ {"time":"2025-11-06T19:15:51.940995974+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
smolvla_omy/wandb/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 16:21:49,585 INFO MainThread:528220 [wandb_setup.py:_flush():81] Current SDK version is 0.22.3
2
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Configure stats pid to 528220
3
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Loading settings from /root/.config/wandb/settings
4
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Loading settings from /workspace/liber/embodied_ai/lerobot-mujoco-tutorial/wandb/settings
5
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:setup_run_log_directory():706] Logging user logs to ckpt/smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug.log
7
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:setup_run_log_directory():707] Logging internal logs to ckpt/smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-internal.log
8
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:init():833] calling init triggers
9
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:init():838] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'omy_pnp_language', 'root': './demo_data_language', 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec'}, 'env': None, 'policy': {'type': 'smolvla', 'n_obs_steps': 1, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'chunk_size': 5, 'n_action_steps': 5, 'max_state_dim': 32, 'max_action_dim': 32, 'resize_imgs_with_padding': [512, 512], 'empty_cameras': 0, 'adapt_to_pi_aloha': False, 'use_delta_joint_actions_aloha': False, 'tokenizer_max_length': 48, 'num_steps': 10, 'use_cache': True, 'freeze_vision_encoder': True, 'train_expert_only': True, 'train_state_proj': True, 'optimizer_lr': 0.0001, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 1e-10, 'optimizer_grad_clip_norm': 10, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'vlm_model_name': 'HuggingFaceTB/SmolVLM2-500M-Video-Instruct', 'load_vlm_weights': False, 'add_image_special_tokens': False, 'attention_mode': 'cross_attn', 'prefix_length': -1, 'pad_language_to': 'longest', 'num_expert_layers': -1, 'num_vlm_layers': 16, 'self_attn_every_n_layers': 2, 'expert_width_multiplier': 0.75, 'min_period': 0.004, 'max_period': 4.0}, 'output_dir': 'ckpt/smolvla_omy', 'job_name': 'smolvla_omy', 'resume': False, 'seed': 42, 'num_workers': 8, 'batch_size': 16, 'steps': 20000, 'eval_freq': -1, 'log_freq': 50, 'save_checkpoint': True, 'save_freq': 5000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 0.0001, 'weight_decay': 1e-10, 'grad_clip_norm': 10, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 0.0001, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': True, 'project': 'smolvla_omy', 'entity': 'ai-liber', 'notes': None, 'run_id': None, 'mode': None}, '_wandb': {}}
11
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:init():881] starting backend
12
+ 2025-11-06 16:21:49,796 INFO MainThread:528220 [wandb_init.py:init():884] sending inform_init request
13
+ 2025-11-06 16:21:49,803 INFO MainThread:528220 [wandb_init.py:init():892] backend started and connected
14
+ 2025-11-06 16:21:49,806 INFO MainThread:528220 [wandb_init.py:init():962] updated telemetry
15
+ 2025-11-06 16:21:49,819 INFO MainThread:528220 [wandb_init.py:init():986] communicating run to backend with 90.0 second timeout
16
+ 2025-11-06 16:21:51,055 INFO MainThread:528220 [wandb_init.py:init():1033] starting run threads in backend
17
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-11-06 16:21:51,416 INFO MainThread:528220 [wandb_init.py:init():1073] run started, returning control to user process
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/output.log ADDED
@@ -0,0 +1,631 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Logs will be synced with wandb.
2
+ INFO 2025-11-06 16:21:51 ndb_utils.py:96 Track this run --> https://wandb.ai/ai-liber/smolvla_omy/runs/8gs9of7r
3
+ INFO 2025-11-06 16:21:51 in_model.py:127 Creating dataset
4
+ Resolving data files: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 20/20 [00:00<00:00, 18149.30it/s]
5
+ INFO 2025-11-06 16:21:52 in_model.py:138 Creating policy
6
+ Reducing the number of VLM layers to 16 ...
7
+ INFO 2025-11-06 16:23:01 in_model.py:148 Creating optimizer and scheduler
8
+ INFO 2025-11-06 16:23:01 in_model.py:160 Output dir: ckpt/smolvla_omy
9
+ INFO 2025-11-06 16:23:01 in_model.py:163 cfg.steps=20000 (20K)
10
+ INFO 2025-11-06 16:23:01 in_model.py:164 dataset.num_frames=2621 (3K)
11
+ INFO 2025-11-06 16:23:01 in_model.py:165 dataset.num_episodes=20
12
+ INFO 2025-11-06 16:23:01 in_model.py:166 num_learnable_params=99880992 (100M)
13
+ INFO 2025-11-06 16:23:01 in_model.py:167 num_total_params=450046216 (450M)
14
+ INFO 2025-11-06 16:23:01 in_model.py:206 Start offline training on a fixed dataset
15
+ INFO 2025-11-06 16:23:52 in_model.py:236 step:50 smpl:800 ep:6 epch:0.31 loss:0.306 grdn:2.807 lr:2.6e-06 updt_s:0.998 data_s:0.022
16
+ WARNING 2025-11-06 16:23:52 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
17
+ WARNING 2025-11-06 16:23:52 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
18
+ INFO 2025-11-06 16:24:41 in_model.py:236 step:100 smpl:2K ep:12 epch:0.61 loss:0.223 grdn:1.959 lr:7.6e-06 updt_s:0.982 data_s:0.000
19
+ WARNING 2025-11-06 16:24:41 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
20
+ WARNING 2025-11-06 16:24:41 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
21
+ INFO 2025-11-06 16:25:32 in_model.py:236 step:150 smpl:2K ep:18 epch:0.92 loss:0.166 grdn:1.444 lr:1.3e-05 updt_s:1.013 data_s:0.000
22
+ WARNING 2025-11-06 16:25:32 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
23
+ WARNING 2025-11-06 16:25:32 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
24
+ INFO 2025-11-06 16:26:26 in_model.py:236 step:200 smpl:3K ep:24 epch:1.22 loss:0.112 grdn:1.042 lr:1.8e-05 updt_s:1.049 data_s:0.021
25
+ WARNING 2025-11-06 16:26:26 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
26
+ WARNING 2025-11-06 16:26:26 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
27
+ INFO 2025-11-06 16:27:18 in_model.py:236 step:250 smpl:4K ep:31 epch:1.53 loss:0.091 grdn:0.947 lr:2.3e-05 updt_s:1.048 data_s:0.000
28
+ WARNING 2025-11-06 16:27:18 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
29
+ WARNING 2025-11-06 16:27:18 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
30
+ INFO 2025-11-06 16:28:11 in_model.py:236 step:300 smpl:5K ep:37 epch:1.83 loss:0.079 grdn:0.972 lr:2.8e-05 updt_s:1.060 data_s:0.000
31
+ WARNING 2025-11-06 16:28:11 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
32
+ WARNING 2025-11-06 16:28:11 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
33
+ INFO 2025-11-06 16:29:05 in_model.py:236 step:350 smpl:6K ep:43 epch:2.14 loss:0.068 grdn:0.944 lr:3.3e-05 updt_s:1.056 data_s:0.021
34
+ WARNING 2025-11-06 16:29:05 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
35
+ WARNING 2025-11-06 16:29:05 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
36
+ INFO 2025-11-06 16:29:59 in_model.py:236 step:400 smpl:6K ep:49 epch:2.44 loss:0.062 grdn:0.921 lr:3.8e-05 updt_s:1.063 data_s:0.000
37
+ WARNING 2025-11-06 16:29:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
38
+ WARNING 2025-11-06 16:29:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
39
+ INFO 2025-11-06 16:30:52 in_model.py:236 step:450 smpl:7K ep:55 epch:2.75 loss:0.062 grdn:0.979 lr:4.3e-05 updt_s:1.062 data_s:0.000
40
+ WARNING 2025-11-06 16:30:52 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
41
+ WARNING 2025-11-06 16:30:52 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
42
+ INFO 2025-11-06 16:31:46 in_model.py:236 step:500 smpl:8K ep:61 epch:3.05 loss:0.055 grdn:0.987 lr:4.8e-05 updt_s:1.055 data_s:0.021
43
+ WARNING 2025-11-06 16:31:46 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
44
+ WARNING 2025-11-06 16:31:46 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
45
+ INFO 2025-11-06 16:32:39 in_model.py:236 step:550 smpl:9K ep:67 epch:3.36 loss:0.046 grdn:0.843 lr:5.3e-05 updt_s:1.053 data_s:0.000
46
+ WARNING 2025-11-06 16:32:39 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
47
+ WARNING 2025-11-06 16:32:39 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
48
+ INFO 2025-11-06 16:33:32 in_model.py:236 step:600 smpl:10K ep:73 epch:3.66 loss:0.050 grdn:0.955 lr:5.8e-05 updt_s:1.054 data_s:0.000
49
+ WARNING 2025-11-06 16:33:32 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
50
+ WARNING 2025-11-06 16:33:32 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
51
+ INFO 2025-11-06 16:34:25 in_model.py:236 step:650 smpl:10K ep:79 epch:3.97 loss:0.050 grdn:1.030 lr:6.3e-05 updt_s:1.058 data_s:0.000
52
+ WARNING 2025-11-06 16:34:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
53
+ WARNING 2025-11-06 16:34:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
54
+ INFO 2025-11-06 16:35:19 in_model.py:236 step:700 smpl:11K ep:85 epch:4.27 loss:0.044 grdn:0.878 lr:6.8e-05 updt_s:1.055 data_s:0.021
55
+ WARNING 2025-11-06 16:35:19 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
56
+ WARNING 2025-11-06 16:35:19 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
57
+ INFO 2025-11-06 16:36:12 in_model.py:236 step:750 smpl:12K ep:92 epch:4.58 loss:0.048 grdn:0.911 lr:7.3e-05 updt_s:1.064 data_s:0.000
58
+ WARNING 2025-11-06 16:36:12 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
59
+ WARNING 2025-11-06 16:36:12 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
60
+ INFO 2025-11-06 16:37:05 in_model.py:236 step:800 smpl:13K ep:98 epch:4.88 loss:0.050 grdn:0.946 lr:7.8e-05 updt_s:1.063 data_s:0.000
61
+ WARNING 2025-11-06 16:37:05 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
62
+ WARNING 2025-11-06 16:37:05 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
63
+ INFO 2025-11-06 16:37:59 in_model.py:236 step:850 smpl:14K ep:104 epch:5.19 loss:0.045 grdn:0.904 lr:8.3e-05 updt_s:1.056 data_s:0.024
64
+ WARNING 2025-11-06 16:37:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
65
+ WARNING 2025-11-06 16:37:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
66
+ INFO 2025-11-06 16:38:53 in_model.py:236 step:900 smpl:14K ep:110 epch:5.49 loss:0.056 grdn:1.043 lr:8.8e-05 updt_s:1.062 data_s:0.000
67
+ WARNING 2025-11-06 16:38:53 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
68
+ WARNING 2025-11-06 16:38:53 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
69
+ INFO 2025-11-06 16:39:46 in_model.py:236 step:950 smpl:15K ep:116 epch:5.80 loss:0.056 grdn:1.051 lr:9.3e-05 updt_s:1.058 data_s:0.000
70
+ WARNING 2025-11-06 16:39:46 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
71
+ WARNING 2025-11-06 16:39:46 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
72
+ INFO 2025-11-06 16:40:40 in_model.py:236 step:1K smpl:16K ep:122 epch:6.10 loss:0.051 grdn:0.937 lr:9.8e-05 updt_s:1.056 data_s:0.020
73
+ WARNING 2025-11-06 16:40:40 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
74
+ WARNING 2025-11-06 16:40:40 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
75
+ INFO 2025-11-06 16:41:33 in_model.py:236 step:1K smpl:17K ep:128 epch:6.41 loss:0.047 grdn:0.883 lr:1.0e-04 updt_s:1.060 data_s:0.000
76
+ WARNING 2025-11-06 16:41:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
77
+ WARNING 2025-11-06 16:41:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
78
+ INFO 2025-11-06 16:42:26 in_model.py:236 step:1K smpl:18K ep:134 epch:6.71 loss:0.056 grdn:0.951 lr:1.0e-04 updt_s:1.059 data_s:0.000
79
+ WARNING 2025-11-06 16:42:26 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
80
+ WARNING 2025-11-06 16:42:26 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
81
+ INFO 2025-11-06 16:43:20 in_model.py:236 step:1K smpl:18K ep:140 epch:7.02 loss:0.055 grdn:0.955 lr:1.0e-04 updt_s:1.057 data_s:0.022
82
+ WARNING 2025-11-06 16:43:20 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
83
+ WARNING 2025-11-06 16:43:20 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
84
+ INFO 2025-11-06 16:44:13 in_model.py:236 step:1K smpl:19K ep:147 epch:7.33 loss:0.049 grdn:0.953 lr:1.0e-04 updt_s:1.056 data_s:0.000
85
+ WARNING 2025-11-06 16:44:13 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
86
+ WARNING 2025-11-06 16:44:13 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
87
+ INFO 2025-11-06 16:45:06 in_model.py:236 step:1K smpl:20K ep:153 epch:7.63 loss:0.045 grdn:0.868 lr:1.0e-04 updt_s:1.062 data_s:0.000
88
+ WARNING 2025-11-06 16:45:06 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
89
+ WARNING 2025-11-06 16:45:06 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
90
+ INFO 2025-11-06 16:45:59 in_model.py:236 step:1K smpl:21K ep:159 epch:7.94 loss:0.053 grdn:0.922 lr:1.0e-04 updt_s:1.061 data_s:0.000
91
+ WARNING 2025-11-06 16:45:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
92
+ WARNING 2025-11-06 16:45:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
93
+ INFO 2025-11-06 16:46:53 in_model.py:236 step:1K smpl:22K ep:165 epch:8.24 loss:0.054 grdn:0.953 lr:1.0e-04 updt_s:1.055 data_s:0.021
94
+ WARNING 2025-11-06 16:46:53 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
95
+ WARNING 2025-11-06 16:46:53 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
96
+ INFO 2025-11-06 16:47:46 in_model.py:236 step:1K smpl:22K ep:171 epch:8.55 loss:0.044 grdn:0.778 lr:9.9e-05 updt_s:1.057 data_s:0.000
97
+ WARNING 2025-11-06 16:47:46 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
98
+ WARNING 2025-11-06 16:47:46 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
99
+ INFO 2025-11-06 16:48:39 in_model.py:236 step:1K smpl:23K ep:177 epch:8.85 loss:0.044 grdn:0.855 lr:9.9e-05 updt_s:1.055 data_s:0.000
100
+ WARNING 2025-11-06 16:48:39 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
101
+ WARNING 2025-11-06 16:48:39 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
102
+ INFO 2025-11-06 16:49:33 in_model.py:236 step:2K smpl:24K ep:183 epch:9.16 loss:0.045 grdn:0.835 lr:9.9e-05 updt_s:1.050 data_s:0.021
103
+ WARNING 2025-11-06 16:49:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
104
+ WARNING 2025-11-06 16:49:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
105
+ INFO 2025-11-06 16:50:26 in_model.py:236 step:2K smpl:25K ep:189 epch:9.46 loss:0.043 grdn:0.803 lr:9.9e-05 updt_s:1.053 data_s:0.000
106
+ WARNING 2025-11-06 16:50:26 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
107
+ WARNING 2025-11-06 16:50:26 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
108
+ INFO 2025-11-06 16:51:19 in_model.py:236 step:2K smpl:26K ep:195 epch:9.77 loss:0.047 grdn:0.870 lr:9.9e-05 updt_s:1.060 data_s:0.000
109
+ WARNING 2025-11-06 16:51:19 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
110
+ WARNING 2025-11-06 16:51:19 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
111
+ INFO 2025-11-06 16:52:13 in_model.py:236 step:2K smpl:26K ep:201 epch:10.07 loss:0.042 grdn:0.784 lr:9.9e-05 updt_s:1.057 data_s:0.024
112
+ WARNING 2025-11-06 16:52:13 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
113
+ WARNING 2025-11-06 16:52:13 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
114
+ INFO 2025-11-06 16:53:06 in_model.py:236 step:2K smpl:27K ep:208 epch:10.38 loss:0.043 grdn:0.775 lr:9.9e-05 updt_s:1.053 data_s:0.000
115
+ WARNING 2025-11-06 16:53:06 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
116
+ WARNING 2025-11-06 16:53:06 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
117
+ INFO 2025-11-06 16:53:59 in_model.py:236 step:2K smpl:28K ep:214 epch:10.68 loss:0.042 grdn:0.799 lr:9.9e-05 updt_s:1.056 data_s:0.000
118
+ WARNING 2025-11-06 16:53:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
119
+ WARNING 2025-11-06 16:53:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
120
+ INFO 2025-11-06 16:54:52 in_model.py:236 step:2K smpl:29K ep:220 epch:10.99 loss:0.038 grdn:0.714 lr:9.9e-05 updt_s:1.052 data_s:0.000
121
+ WARNING 2025-11-06 16:54:52 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
122
+ WARNING 2025-11-06 16:54:52 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
123
+ INFO 2025-11-06 16:55:45 in_model.py:236 step:2K smpl:30K ep:226 epch:11.29 loss:0.040 grdn:0.753 lr:9.9e-05 updt_s:1.049 data_s:0.021
124
+ WARNING 2025-11-06 16:55:45 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
125
+ WARNING 2025-11-06 16:55:45 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
126
+ INFO 2025-11-06 16:56:38 in_model.py:236 step:2K smpl:30K ep:232 epch:11.60 loss:0.038 grdn:0.699 lr:9.9e-05 updt_s:1.055 data_s:0.000
127
+ WARNING 2025-11-06 16:56:38 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
128
+ WARNING 2025-11-06 16:56:38 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
129
+ INFO 2025-11-06 16:57:31 in_model.py:236 step:2K smpl:31K ep:238 epch:11.90 loss:0.038 grdn:0.660 lr:9.9e-05 updt_s:1.054 data_s:0.000
130
+ WARNING 2025-11-06 16:57:31 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
131
+ WARNING 2025-11-06 16:57:31 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
132
+ INFO 2025-11-06 16:58:25 in_model.py:236 step:2K smpl:32K ep:244 epch:12.21 loss:0.034 grdn:0.633 lr:9.9e-05 updt_s:1.048 data_s:0.021
133
+ WARNING 2025-11-06 16:58:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
134
+ WARNING 2025-11-06 16:58:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
135
+ INFO 2025-11-06 16:59:17 in_model.py:236 step:2K smpl:33K ep:250 epch:12.51 loss:0.039 grdn:0.733 lr:9.9e-05 updt_s:1.055 data_s:0.000
136
+ WARNING 2025-11-06 16:59:18 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
137
+ WARNING 2025-11-06 16:59:18 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
138
+ INFO 2025-11-06 17:00:10 in_model.py:236 step:2K smpl:34K ep:256 epch:12.82 loss:0.037 grdn:0.686 lr:9.9e-05 updt_s:1.056 data_s:0.000
139
+ WARNING 2025-11-06 17:00:10 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
140
+ WARNING 2025-11-06 17:00:10 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
141
+ INFO 2025-11-06 17:01:04 in_model.py:236 step:2K smpl:34K ep:262 epch:13.12 loss:0.038 grdn:0.758 lr:9.9e-05 updt_s:1.048 data_s:0.021
142
+ WARNING 2025-11-06 17:01:04 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
143
+ WARNING 2025-11-06 17:01:04 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
144
+ INFO 2025-11-06 17:01:57 in_model.py:236 step:2K smpl:35K ep:269 epch:13.43 loss:0.034 grdn:0.704 lr:9.9e-05 updt_s:1.052 data_s:0.000
145
+ WARNING 2025-11-06 17:01:57 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
146
+ WARNING 2025-11-06 17:01:57 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
147
+ INFO 2025-11-06 17:02:50 in_model.py:236 step:2K smpl:36K ep:275 epch:13.74 loss:0.030 grdn:0.661 lr:9.9e-05 updt_s:1.052 data_s:0.000
148
+ WARNING 2025-11-06 17:02:50 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
149
+ WARNING 2025-11-06 17:02:50 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
150
+ INFO 2025-11-06 17:03:43 in_model.py:236 step:2K smpl:37K ep:281 epch:14.04 loss:0.036 grdn:0.715 lr:9.9e-05 updt_s:1.048 data_s:0.022
151
+ WARNING 2025-11-06 17:03:43 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
152
+ WARNING 2025-11-06 17:03:43 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
153
+ INFO 2025-11-06 17:04:36 in_model.py:236 step:2K smpl:38K ep:287 epch:14.35 loss:0.039 grdn:0.753 lr:9.9e-05 updt_s:1.056 data_s:0.000
154
+ WARNING 2025-11-06 17:04:36 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
155
+ WARNING 2025-11-06 17:04:36 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
156
+ INFO 2025-11-06 17:05:30 in_model.py:236 step:2K smpl:38K ep:293 epch:14.65 loss:0.035 grdn:0.735 lr:9.8e-05 updt_s:1.063 data_s:0.000
157
+ WARNING 2025-11-06 17:05:30 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
158
+ WARNING 2025-11-06 17:05:30 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
159
+ INFO 2025-11-06 17:06:23 in_model.py:236 step:2K smpl:39K ep:299 epch:14.96 loss:0.031 grdn:0.641 lr:9.8e-05 updt_s:1.064 data_s:0.000
160
+ WARNING 2025-11-06 17:06:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
161
+ WARNING 2025-11-06 17:06:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
162
+ INFO 2025-11-06 17:07:17 in_model.py:236 step:2K smpl:40K ep:305 epch:15.26 loss:0.038 grdn:0.796 lr:9.8e-05 updt_s:1.059 data_s:0.020
163
+ WARNING 2025-11-06 17:07:17 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
164
+ WARNING 2025-11-06 17:07:17 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
165
+ INFO 2025-11-06 17:08:10 in_model.py:236 step:3K smpl:41K ep:311 epch:15.57 loss:0.036 grdn:0.678 lr:9.8e-05 updt_s:1.058 data_s:0.000
166
+ WARNING 2025-11-06 17:08:10 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
167
+ WARNING 2025-11-06 17:08:10 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
168
+ INFO 2025-11-06 17:09:03 in_model.py:236 step:3K smpl:42K ep:317 epch:15.87 loss:0.036 grdn:0.720 lr:9.8e-05 updt_s:1.063 data_s:0.000
169
+ WARNING 2025-11-06 17:09:03 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
170
+ WARNING 2025-11-06 17:09:03 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
171
+ INFO 2025-11-06 17:09:57 in_model.py:236 step:3K smpl:42K ep:324 epch:16.18 loss:0.028 grdn:0.588 lr:9.8e-05 updt_s:1.059 data_s:0.021
172
+ WARNING 2025-11-06 17:09:57 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
173
+ WARNING 2025-11-06 17:09:57 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
174
+ INFO 2025-11-06 17:10:51 in_model.py:236 step:3K smpl:43K ep:330 epch:16.48 loss:0.029 grdn:0.607 lr:9.8e-05 updt_s:1.060 data_s:0.000
175
+ WARNING 2025-11-06 17:10:51 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
176
+ WARNING 2025-11-06 17:10:51 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
177
+ INFO 2025-11-06 17:11:44 in_model.py:236 step:3K smpl:44K ep:336 epch:16.79 loss:0.029 grdn:0.611 lr:9.8e-05 updt_s:1.055 data_s:0.000
178
+ WARNING 2025-11-06 17:11:44 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
179
+ WARNING 2025-11-06 17:11:44 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
180
+ INFO 2025-11-06 17:12:37 in_model.py:236 step:3K smpl:45K ep:342 epch:17.09 loss:0.031 grdn:0.626 lr:9.8e-05 updt_s:1.049 data_s:0.020
181
+ WARNING 2025-11-06 17:12:37 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
182
+ WARNING 2025-11-06 17:12:37 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
183
+ INFO 2025-11-06 17:13:30 in_model.py:236 step:3K smpl:46K ep:348 epch:17.40 loss:0.030 grdn:0.638 lr:9.8e-05 updt_s:1.054 data_s:0.000
184
+ WARNING 2025-11-06 17:13:30 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
185
+ WARNING 2025-11-06 17:13:30 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
186
+ INFO 2025-11-06 17:14:23 in_model.py:236 step:3K smpl:46K ep:354 epch:17.70 loss:0.035 grdn:0.653 lr:9.8e-05 updt_s:1.055 data_s:0.000
187
+ WARNING 2025-11-06 17:14:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
188
+ WARNING 2025-11-06 17:14:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
189
+ INFO 2025-11-06 17:15:16 in_model.py:236 step:3K smpl:47K ep:360 epch:18.01 loss:0.036 grdn:0.653 lr:9.8e-05 updt_s:1.055 data_s:0.000
190
+ WARNING 2025-11-06 17:15:16 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
191
+ WARNING 2025-11-06 17:15:16 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
192
+ INFO 2025-11-06 17:16:09 in_model.py:236 step:3K smpl:48K ep:366 epch:18.31 loss:0.025 grdn:0.577 lr:9.8e-05 updt_s:1.049 data_s:0.021
193
+ WARNING 2025-11-06 17:16:09 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
194
+ WARNING 2025-11-06 17:16:09 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
195
+ INFO 2025-11-06 17:17:02 in_model.py:236 step:3K smpl:49K ep:372 epch:18.62 loss:0.036 grdn:0.641 lr:9.8e-05 updt_s:1.056 data_s:0.000
196
+ WARNING 2025-11-06 17:17:02 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
197
+ WARNING 2025-11-06 17:17:02 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
198
+ INFO 2025-11-06 17:17:55 in_model.py:236 step:3K smpl:50K ep:378 epch:18.92 loss:0.028 grdn:0.591 lr:9.7e-05 updt_s:1.056 data_s:0.000
199
+ WARNING 2025-11-06 17:17:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
200
+ WARNING 2025-11-06 17:17:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
201
+ INFO 2025-11-06 17:18:49 in_model.py:236 step:3K smpl:50K ep:385 epch:19.23 loss:0.027 grdn:0.589 lr:9.7e-05 updt_s:1.049 data_s:0.020
202
+ WARNING 2025-11-06 17:18:49 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
203
+ WARNING 2025-11-06 17:18:49 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
204
+ INFO 2025-11-06 17:19:42 in_model.py:236 step:3K smpl:51K ep:391 epch:19.53 loss:0.027 grdn:0.577 lr:9.7e-05 updt_s:1.053 data_s:0.000
205
+ WARNING 2025-11-06 17:19:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
206
+ WARNING 2025-11-06 17:19:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
207
+ INFO 2025-11-06 17:20:35 in_model.py:236 step:3K smpl:52K ep:397 epch:19.84 loss:0.028 grdn:0.567 lr:9.7e-05 updt_s:1.058 data_s:0.000
208
+ WARNING 2025-11-06 17:20:35 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
209
+ WARNING 2025-11-06 17:20:35 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
210
+ INFO 2025-11-06 17:21:28 in_model.py:236 step:3K smpl:53K ep:403 epch:20.14 loss:0.029 grdn:0.603 lr:9.7e-05 updt_s:1.049 data_s:0.020
211
+ WARNING 2025-11-06 17:21:28 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
212
+ WARNING 2025-11-06 17:21:28 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
213
+ INFO 2025-11-06 17:22:21 in_model.py:236 step:3K smpl:54K ep:409 epch:20.45 loss:0.030 grdn:0.611 lr:9.7e-05 updt_s:1.056 data_s:0.000
214
+ WARNING 2025-11-06 17:22:21 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
215
+ WARNING 2025-11-06 17:22:21 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
216
+ INFO 2025-11-06 17:23:14 in_model.py:236 step:3K smpl:54K ep:415 epch:20.76 loss:0.025 grdn:0.572 lr:9.7e-05 updt_s:1.057 data_s:0.000
217
+ WARNING 2025-11-06 17:23:14 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
218
+ WARNING 2025-11-06 17:23:14 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
219
+ INFO 2025-11-06 17:24:08 in_model.py:236 step:3K smpl:55K ep:421 epch:21.06 loss:0.030 grdn:0.592 lr:9.7e-05 updt_s:1.050 data_s:0.021
220
+ WARNING 2025-11-06 17:24:08 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
221
+ WARNING 2025-11-06 17:24:08 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
222
+ INFO 2025-11-06 17:25:01 in_model.py:236 step:4K smpl:56K ep:427 epch:21.37 loss:0.029 grdn:0.615 lr:9.7e-05 updt_s:1.054 data_s:0.000
223
+ WARNING 2025-11-06 17:25:01 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
224
+ WARNING 2025-11-06 17:25:01 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
225
+ INFO 2025-11-06 17:25:54 in_model.py:236 step:4K smpl:57K ep:433 epch:21.67 loss:0.025 grdn:0.524 lr:9.7e-05 updt_s:1.056 data_s:0.000
226
+ WARNING 2025-11-06 17:25:54 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
227
+ WARNING 2025-11-06 17:25:54 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
228
+ INFO 2025-11-06 17:26:47 in_model.py:236 step:4K smpl:58K ep:440 epch:21.98 loss:0.032 grdn:0.574 lr:9.7e-05 updt_s:1.053 data_s:0.000
229
+ WARNING 2025-11-06 17:26:47 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
230
+ WARNING 2025-11-06 17:26:47 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
231
+ INFO 2025-11-06 17:27:40 in_model.py:236 step:4K smpl:58K ep:446 epch:22.28 loss:0.025 grdn:0.527 lr:9.7e-05 updt_s:1.045 data_s:0.022
232
+ WARNING 2025-11-06 17:27:40 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
233
+ WARNING 2025-11-06 17:27:40 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
234
+ INFO 2025-11-06 17:28:33 in_model.py:236 step:4K smpl:59K ep:452 epch:22.59 loss:0.030 grdn:0.578 lr:9.6e-05 updt_s:1.046 data_s:0.000
235
+ WARNING 2025-11-06 17:28:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
236
+ WARNING 2025-11-06 17:28:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
237
+ INFO 2025-11-06 17:29:25 in_model.py:236 step:4K smpl:60K ep:458 epch:22.89 loss:0.025 grdn:0.504 lr:9.6e-05 updt_s:1.054 data_s:0.000
238
+ WARNING 2025-11-06 17:29:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
239
+ WARNING 2025-11-06 17:29:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
240
+ INFO 2025-11-06 17:30:19 in_model.py:236 step:4K smpl:61K ep:464 epch:23.20 loss:0.030 grdn:0.586 lr:9.6e-05 updt_s:1.048 data_s:0.021
241
+ WARNING 2025-11-06 17:30:19 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
242
+ WARNING 2025-11-06 17:30:19 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
243
+ INFO 2025-11-06 17:31:12 in_model.py:236 step:4K smpl:62K ep:470 epch:23.50 loss:0.032 grdn:0.626 lr:9.6e-05 updt_s:1.054 data_s:0.000
244
+ WARNING 2025-11-06 17:31:12 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
245
+ WARNING 2025-11-06 17:31:12 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
246
+ INFO 2025-11-06 17:32:05 in_model.py:236 step:4K smpl:62K ep:476 epch:23.81 loss:0.032 grdn:0.594 lr:9.6e-05 updt_s:1.052 data_s:0.000
247
+ WARNING 2025-11-06 17:32:05 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
248
+ WARNING 2025-11-06 17:32:05 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
249
+ INFO 2025-11-06 17:32:58 in_model.py:236 step:4K smpl:63K ep:482 epch:24.11 loss:0.025 grdn:0.536 lr:9.6e-05 updt_s:1.045 data_s:0.022
250
+ WARNING 2025-11-06 17:32:58 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
251
+ WARNING 2025-11-06 17:32:58 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
252
+ INFO 2025-11-06 17:33:51 in_model.py:236 step:4K smpl:64K ep:488 epch:24.42 loss:0.025 grdn:0.529 lr:9.6e-05 updt_s:1.049 data_s:0.000
253
+ WARNING 2025-11-06 17:33:51 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
254
+ WARNING 2025-11-06 17:33:51 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
255
+ INFO 2025-11-06 17:34:44 in_model.py:236 step:4K smpl:65K ep:494 epch:24.72 loss:0.024 grdn:0.528 lr:9.6e-05 updt_s:1.052 data_s:0.000
256
+ WARNING 2025-11-06 17:34:44 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
257
+ WARNING 2025-11-06 17:34:44 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
258
+ INFO 2025-11-06 17:35:36 in_model.py:236 step:4K smpl:66K ep:501 epch:25.03 loss:0.026 grdn:0.545 lr:9.6e-05 updt_s:1.053 data_s:0.000
259
+ WARNING 2025-11-06 17:35:36 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
260
+ WARNING 2025-11-06 17:35:36 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
261
+ INFO 2025-11-06 17:36:30 in_model.py:236 step:4K smpl:66K ep:507 epch:25.33 loss:0.027 grdn:0.549 lr:9.6e-05 updt_s:1.052 data_s:0.021
262
+ WARNING 2025-11-06 17:36:30 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
263
+ WARNING 2025-11-06 17:36:30 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
264
+ INFO 2025-11-06 17:37:23 in_model.py:236 step:4K smpl:67K ep:513 epch:25.64 loss:0.025 grdn:0.554 lr:9.5e-05 updt_s:1.056 data_s:0.000
265
+ WARNING 2025-11-06 17:37:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
266
+ WARNING 2025-11-06 17:37:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
267
+ INFO 2025-11-06 17:38:16 in_model.py:236 step:4K smpl:68K ep:519 epch:25.94 loss:0.023 grdn:0.522 lr:9.5e-05 updt_s:1.054 data_s:0.000
268
+ WARNING 2025-11-06 17:38:16 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
269
+ WARNING 2025-11-06 17:38:16 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
270
+ INFO 2025-11-06 17:39:10 in_model.py:236 step:4K smpl:69K ep:525 epch:26.25 loss:0.025 grdn:0.564 lr:9.5e-05 updt_s:1.050 data_s:0.020
271
+ WARNING 2025-11-06 17:39:10 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
272
+ WARNING 2025-11-06 17:39:10 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
273
+ INFO 2025-11-06 17:40:03 in_model.py:236 step:4K smpl:70K ep:531 epch:26.55 loss:0.027 grdn:0.561 lr:9.5e-05 updt_s:1.056 data_s:0.000
274
+ WARNING 2025-11-06 17:40:03 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
275
+ WARNING 2025-11-06 17:40:03 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
276
+ INFO 2025-11-06 17:40:55 in_model.py:236 step:4K smpl:70K ep:537 epch:26.86 loss:0.027 grdn:0.616 lr:9.5e-05 updt_s:1.055 data_s:0.000
277
+ WARNING 2025-11-06 17:40:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
278
+ WARNING 2025-11-06 17:40:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
279
+ INFO 2025-11-06 17:41:49 in_model.py:236 step:4K smpl:71K ep:543 epch:27.17 loss:0.030 grdn:0.585 lr:9.5e-05 updt_s:1.047 data_s:0.021
280
+ WARNING 2025-11-06 17:41:49 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
281
+ WARNING 2025-11-06 17:41:49 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
282
+ INFO 2025-11-06 17:42:42 in_model.py:236 step:4K smpl:72K ep:549 epch:27.47 loss:0.023 grdn:0.508 lr:9.5e-05 updt_s:1.053 data_s:0.000
283
+ WARNING 2025-11-06 17:42:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
284
+ WARNING 2025-11-06 17:42:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
285
+ INFO 2025-11-06 17:43:35 in_model.py:236 step:5K smpl:73K ep:556 epch:27.78 loss:0.023 grdn:0.472 lr:9.5e-05 updt_s:1.054 data_s:0.000
286
+ WARNING 2025-11-06 17:43:35 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
287
+ WARNING 2025-11-06 17:43:35 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
288
+ INFO 2025-11-06 17:44:28 in_model.py:236 step:5K smpl:74K ep:562 epch:28.08 loss:0.025 grdn:0.559 lr:9.5e-05 updt_s:1.048 data_s:0.021
289
+ WARNING 2025-11-06 17:44:28 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
290
+ WARNING 2025-11-06 17:44:28 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
291
+ INFO 2025-11-06 17:45:21 in_model.py:236 step:5K smpl:74K ep:568 epch:28.39 loss:0.023 grdn:0.552 lr:9.4e-05 updt_s:1.054 data_s:0.000
292
+ WARNING 2025-11-06 17:45:21 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
293
+ WARNING 2025-11-06 17:45:21 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
294
+ INFO 2025-11-06 17:46:14 in_model.py:236 step:5K smpl:75K ep:574 epch:28.69 loss:0.027 grdn:0.555 lr:9.4e-05 updt_s:1.054 data_s:0.000
295
+ WARNING 2025-11-06 17:46:14 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
296
+ WARNING 2025-11-06 17:46:14 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
297
+ INFO 2025-11-06 17:47:07 in_model.py:236 step:5K smpl:76K ep:580 epch:29.00 loss:0.025 grdn:0.503 lr:9.4e-05 updt_s:1.056 data_s:0.000
298
+ WARNING 2025-11-06 17:47:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
299
+ WARNING 2025-11-06 17:47:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
300
+ INFO 2025-11-06 17:48:00 in_model.py:236 step:5K smpl:77K ep:586 epch:29.30 loss:0.024 grdn:0.523 lr:9.4e-05 updt_s:1.045 data_s:0.020
301
+ WARNING 2025-11-06 17:48:00 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
302
+ WARNING 2025-11-06 17:48:00 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
303
+ INFO 2025-11-06 17:48:53 in_model.py:236 step:5K smpl:78K ep:592 epch:29.61 loss:0.023 grdn:0.479 lr:9.4e-05 updt_s:1.054 data_s:0.000
304
+ WARNING 2025-11-06 17:48:53 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
305
+ WARNING 2025-11-06 17:48:53 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
306
+ INFO 2025-11-06 17:49:46 in_model.py:236 step:5K smpl:78K ep:598 epch:29.91 loss:0.026 grdn:0.566 lr:9.4e-05 updt_s:1.056 data_s:0.000
307
+ WARNING 2025-11-06 17:49:46 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
308
+ WARNING 2025-11-06 17:49:46 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
309
+ INFO 2025-11-06 17:50:40 in_model.py:236 step:5K smpl:79K ep:604 epch:30.22 loss:0.021 grdn:0.490 lr:9.4e-05 updt_s:1.051 data_s:0.020
310
+ WARNING 2025-11-06 17:50:40 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
311
+ WARNING 2025-11-06 17:50:40 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
312
+ INFO 2025-11-06 17:51:33 in_model.py:236 step:5K smpl:80K ep:610 epch:30.52 loss:0.021 grdn:0.485 lr:9.4e-05 updt_s:1.061 data_s:0.000
313
+ WARNING 2025-11-06 17:51:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
314
+ WARNING 2025-11-06 17:51:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
315
+ INFO 2025-11-06 17:51:33 in_model.py:245 Checkpoint policy after step 5000
316
+ INFO 2025-11-06 17:52:29 in_model.py:236 step:5K smpl:81K ep:617 epch:30.83 loss:0.022 grdn:0.503 lr:9.3e-05 updt_s:1.052 data_s:0.000
317
+ WARNING 2025-11-06 17:52:29 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
318
+ WARNING 2025-11-06 17:52:29 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
319
+ INFO 2025-11-06 17:53:23 in_model.py:236 step:5K smpl:82K ep:623 epch:31.13 loss:0.025 grdn:0.549 lr:9.3e-05 updt_s:1.058 data_s:0.022
320
+ WARNING 2025-11-06 17:53:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
321
+ WARNING 2025-11-06 17:53:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
322
+ INFO 2025-11-06 17:54:17 in_model.py:236 step:5K smpl:82K ep:629 epch:31.44 loss:0.022 grdn:0.462 lr:9.3e-05 updt_s:1.061 data_s:0.000
323
+ WARNING 2025-11-06 17:54:17 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
324
+ WARNING 2025-11-06 17:54:17 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
325
+ INFO 2025-11-06 17:55:10 in_model.py:236 step:5K smpl:83K ep:635 epch:31.74 loss:0.020 grdn:0.458 lr:9.3e-05 updt_s:1.061 data_s:0.000
326
+ WARNING 2025-11-06 17:55:10 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
327
+ WARNING 2025-11-06 17:55:10 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
328
+ INFO 2025-11-06 17:56:04 in_model.py:236 step:5K smpl:84K ep:641 epch:32.05 loss:0.019 grdn:0.441 lr:9.3e-05 updt_s:1.059 data_s:0.023
329
+ WARNING 2025-11-06 17:56:04 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
330
+ WARNING 2025-11-06 17:56:04 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
331
+ INFO 2025-11-06 17:56:57 in_model.py:236 step:5K smpl:85K ep:647 epch:32.35 loss:0.023 grdn:0.502 lr:9.3e-05 updt_s:1.065 data_s:0.000
332
+ WARNING 2025-11-06 17:56:57 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
333
+ WARNING 2025-11-06 17:56:57 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
334
+ INFO 2025-11-06 17:57:51 in_model.py:236 step:5K smpl:86K ep:653 epch:32.66 loss:0.024 grdn:0.487 lr:9.3e-05 updt_s:1.069 data_s:0.000
335
+ WARNING 2025-11-06 17:57:51 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
336
+ WARNING 2025-11-06 17:57:51 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
337
+ INFO 2025-11-06 17:58:45 in_model.py:236 step:5K smpl:86K ep:659 epch:32.96 loss:0.020 grdn:0.441 lr:9.2e-05 updt_s:1.073 data_s:0.000
338
+ WARNING 2025-11-06 17:58:45 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
339
+ WARNING 2025-11-06 17:58:45 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
340
+ INFO 2025-11-06 17:59:39 in_model.py:236 step:5K smpl:87K ep:665 epch:33.27 loss:0.024 grdn:0.523 lr:9.2e-05 updt_s:1.064 data_s:0.021
341
+ WARNING 2025-11-06 17:59:39 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
342
+ WARNING 2025-11-06 17:59:39 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
343
+ INFO 2025-11-06 18:00:33 in_model.py:236 step:6K smpl:88K ep:671 epch:33.57 loss:0.022 grdn:0.471 lr:9.2e-05 updt_s:1.072 data_s:0.000
344
+ WARNING 2025-11-06 18:00:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
345
+ WARNING 2025-11-06 18:00:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
346
+ INFO 2025-11-06 18:01:27 in_model.py:236 step:6K smpl:89K ep:678 epch:33.88 loss:0.021 grdn:0.422 lr:9.2e-05 updt_s:1.071 data_s:0.000
347
+ WARNING 2025-11-06 18:01:27 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
348
+ WARNING 2025-11-06 18:01:27 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
349
+ INFO 2025-11-06 18:02:21 in_model.py:236 step:6K smpl:90K ep:684 epch:34.19 loss:0.025 grdn:0.555 lr:9.2e-05 updt_s:1.061 data_s:0.022
350
+ WARNING 2025-11-06 18:02:21 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
351
+ WARNING 2025-11-06 18:02:21 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
352
+ INFO 2025-11-06 18:03:15 in_model.py:236 step:6K smpl:90K ep:690 epch:34.49 loss:0.020 grdn:0.464 lr:9.2e-05 updt_s:1.068 data_s:0.000
353
+ WARNING 2025-11-06 18:03:15 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
354
+ WARNING 2025-11-06 18:03:15 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
355
+ INFO 2025-11-06 18:04:08 in_model.py:236 step:6K smpl:91K ep:696 epch:34.80 loss:0.023 grdn:0.500 lr:9.2e-05 updt_s:1.066 data_s:0.000
356
+ WARNING 2025-11-06 18:04:08 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
357
+ WARNING 2025-11-06 18:04:08 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
358
+ INFO 2025-11-06 18:05:02 in_model.py:236 step:6K smpl:92K ep:702 epch:35.10 loss:0.021 grdn:0.453 lr:9.1e-05 updt_s:1.060 data_s:0.022
359
+ WARNING 2025-11-06 18:05:02 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
360
+ WARNING 2025-11-06 18:05:02 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
361
+ INFO 2025-11-06 18:05:56 in_model.py:236 step:6K smpl:93K ep:708 epch:35.41 loss:0.022 grdn:0.518 lr:9.1e-05 updt_s:1.072 data_s:0.000
362
+ WARNING 2025-11-06 18:05:56 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
363
+ WARNING 2025-11-06 18:05:56 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
364
+ INFO 2025-11-06 18:06:50 in_model.py:236 step:6K smpl:94K ep:714 epch:35.71 loss:0.021 grdn:0.456 lr:9.1e-05 updt_s:1.081 data_s:0.000
365
+ WARNING 2025-11-06 18:06:50 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
366
+ WARNING 2025-11-06 18:06:50 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
367
+ INFO 2025-11-06 18:07:44 in_model.py:236 step:6K smpl:94K ep:720 epch:36.02 loss:0.020 grdn:0.463 lr:9.1e-05 updt_s:1.082 data_s:0.000
368
+ WARNING 2025-11-06 18:07:44 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
369
+ WARNING 2025-11-06 18:07:44 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
370
+ INFO 2025-11-06 18:08:39 in_model.py:236 step:6K smpl:95K ep:726 epch:36.32 loss:0.024 grdn:0.583 lr:9.1e-05 updt_s:1.066 data_s:0.021
371
+ WARNING 2025-11-06 18:08:39 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
372
+ WARNING 2025-11-06 18:08:39 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
373
+ INFO 2025-11-06 18:09:32 in_model.py:236 step:6K smpl:96K ep:733 epch:36.63 loss:0.019 grdn:0.503 lr:9.1e-05 updt_s:1.062 data_s:0.000
374
+ WARNING 2025-11-06 18:09:32 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
375
+ WARNING 2025-11-06 18:09:32 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
376
+ INFO 2025-11-06 18:10:26 in_model.py:236 step:6K smpl:97K ep:739 epch:36.93 loss:0.021 grdn:0.497 lr:9.1e-05 updt_s:1.064 data_s:0.000
377
+ WARNING 2025-11-06 18:10:26 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
378
+ WARNING 2025-11-06 18:10:26 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
379
+ INFO 2025-11-06 18:11:20 in_model.py:236 step:6K smpl:98K ep:745 epch:37.24 loss:0.020 grdn:0.467 lr:9.0e-05 updt_s:1.056 data_s:0.022
380
+ WARNING 2025-11-06 18:11:20 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
381
+ WARNING 2025-11-06 18:11:20 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
382
+ INFO 2025-11-06 18:12:13 in_model.py:236 step:6K smpl:98K ep:751 epch:37.54 loss:0.019 grdn:0.423 lr:9.0e-05 updt_s:1.058 data_s:0.000
383
+ WARNING 2025-11-06 18:12:13 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
384
+ WARNING 2025-11-06 18:12:13 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
385
+ INFO 2025-11-06 18:13:06 in_model.py:236 step:6K smpl:99K ep:757 epch:37.85 loss:0.020 grdn:0.476 lr:9.0e-05 updt_s:1.059 data_s:0.000
386
+ WARNING 2025-11-06 18:13:06 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
387
+ WARNING 2025-11-06 18:13:06 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
388
+ INFO 2025-11-06 18:13:59 in_model.py:236 step:6K smpl:100K ep:763 epch:38.15 loss:0.024 grdn:0.498 lr:9.0e-05 updt_s:1.044 data_s:0.022
389
+ WARNING 2025-11-06 18:13:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
390
+ WARNING 2025-11-06 18:13:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
391
+ INFO 2025-11-06 18:14:52 in_model.py:236 step:6K smpl:101K ep:769 epch:38.46 loss:0.025 grdn:0.478 lr:9.0e-05 updt_s:1.051 data_s:0.000
392
+ WARNING 2025-11-06 18:14:52 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
393
+ WARNING 2025-11-06 18:14:52 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
394
+ INFO 2025-11-06 18:15:44 in_model.py:236 step:6K smpl:102K ep:775 epch:38.76 loss:0.020 grdn:0.436 lr:9.0e-05 updt_s:1.049 data_s:0.000
395
+ WARNING 2025-11-06 18:15:44 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
396
+ WARNING 2025-11-06 18:15:44 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
397
+ INFO 2025-11-06 18:16:38 in_model.py:236 step:6K smpl:102K ep:781 epch:39.07 loss:0.020 grdn:0.487 lr:9.0e-05 updt_s:1.049 data_s:0.022
398
+ WARNING 2025-11-06 18:16:38 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
399
+ WARNING 2025-11-06 18:16:38 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
400
+ INFO 2025-11-06 18:17:30 in_model.py:236 step:6K smpl:103K ep:787 epch:39.37 loss:0.020 grdn:0.449 lr:8.9e-05 updt_s:1.041 data_s:0.000
401
+ WARNING 2025-11-06 18:17:30 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
402
+ WARNING 2025-11-06 18:17:30 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
403
+ INFO 2025-11-06 18:18:23 in_model.py:236 step:6K smpl:104K ep:794 epch:39.68 loss:0.018 grdn:0.453 lr:8.9e-05 updt_s:1.047 data_s:0.000
404
+ WARNING 2025-11-06 18:18:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
405
+ WARNING 2025-11-06 18:18:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
406
+ INFO 2025-11-06 18:19:15 in_model.py:236 step:7K smpl:105K ep:800 epch:39.98 loss:0.020 grdn:0.454 lr:8.9e-05 updt_s:1.045 data_s:0.000
407
+ WARNING 2025-11-06 18:19:15 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
408
+ WARNING 2025-11-06 18:19:15 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
409
+ INFO 2025-11-06 18:20:09 in_model.py:236 step:7K smpl:106K ep:806 epch:40.29 loss:0.023 grdn:0.511 lr:8.9e-05 updt_s:1.043 data_s:0.021
410
+ WARNING 2025-11-06 18:20:09 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
411
+ WARNING 2025-11-06 18:20:09 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
412
+ INFO 2025-11-06 18:21:01 in_model.py:236 step:7K smpl:106K ep:812 epch:40.60 loss:0.018 grdn:0.436 lr:8.9e-05 updt_s:1.046 data_s:0.000
413
+ WARNING 2025-11-06 18:21:01 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
414
+ WARNING 2025-11-06 18:21:01 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
415
+ INFO 2025-11-06 18:21:53 in_model.py:236 step:7K smpl:107K ep:818 epch:40.90 loss:0.021 grdn:0.485 lr:8.9e-05 updt_s:1.038 data_s:0.000
416
+ WARNING 2025-11-06 18:21:53 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
417
+ WARNING 2025-11-06 18:21:53 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
418
+ INFO 2025-11-06 18:22:46 in_model.py:236 step:7K smpl:108K ep:824 epch:41.21 loss:0.020 grdn:0.435 lr:8.8e-05 updt_s:1.035 data_s:0.021
419
+ WARNING 2025-11-06 18:22:46 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
420
+ WARNING 2025-11-06 18:22:46 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
421
+ INFO 2025-11-06 18:23:38 in_model.py:236 step:7K smpl:109K ep:830 epch:41.51 loss:0.023 grdn:0.482 lr:8.8e-05 updt_s:1.039 data_s:0.000
422
+ WARNING 2025-11-06 18:23:38 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
423
+ WARNING 2025-11-06 18:23:38 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
424
+ INFO 2025-11-06 18:24:30 in_model.py:236 step:7K smpl:110K ep:836 epch:41.82 loss:0.021 grdn:0.454 lr:8.8e-05 updt_s:1.040 data_s:0.000
425
+ WARNING 2025-11-06 18:24:30 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
426
+ WARNING 2025-11-06 18:24:30 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
427
+ INFO 2025-11-06 18:25:23 in_model.py:236 step:7K smpl:110K ep:842 epch:42.12 loss:0.021 grdn:0.461 lr:8.8e-05 updt_s:1.031 data_s:0.021
428
+ WARNING 2025-11-06 18:25:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
429
+ WARNING 2025-11-06 18:25:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
430
+ INFO 2025-11-06 18:26:15 in_model.py:236 step:7K smpl:111K ep:849 epch:42.43 loss:0.021 grdn:0.455 lr:8.8e-05 updt_s:1.038 data_s:0.000
431
+ WARNING 2025-11-06 18:26:15 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
432
+ WARNING 2025-11-06 18:26:15 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
433
+ INFO 2025-11-06 18:27:07 in_model.py:236 step:7K smpl:112K ep:855 epch:42.73 loss:0.020 grdn:0.468 lr:8.8e-05 updt_s:1.036 data_s:0.000
434
+ WARNING 2025-11-06 18:27:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
435
+ WARNING 2025-11-06 18:27:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
436
+ INFO 2025-11-06 18:27:59 in_model.py:236 step:7K smpl:113K ep:861 epch:43.04 loss:0.020 grdn:0.423 lr:8.7e-05 updt_s:1.038 data_s:0.000
437
+ WARNING 2025-11-06 18:27:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
438
+ WARNING 2025-11-06 18:27:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
439
+ INFO 2025-11-06 18:28:52 in_model.py:236 step:7K smpl:114K ep:867 epch:43.34 loss:0.022 grdn:0.467 lr:8.7e-05 updt_s:1.032 data_s:0.020
440
+ WARNING 2025-11-06 18:28:52 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
441
+ WARNING 2025-11-06 18:28:52 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
442
+ INFO 2025-11-06 18:29:44 in_model.py:236 step:7K smpl:114K ep:873 epch:43.65 loss:0.021 grdn:0.453 lr:8.7e-05 updt_s:1.038 data_s:0.000
443
+ WARNING 2025-11-06 18:29:44 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
444
+ WARNING 2025-11-06 18:29:44 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
445
+ INFO 2025-11-06 18:30:36 in_model.py:236 step:7K smpl:115K ep:879 epch:43.95 loss:0.016 grdn:0.421 lr:8.7e-05 updt_s:1.036 data_s:0.000
446
+ WARNING 2025-11-06 18:30:36 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
447
+ WARNING 2025-11-06 18:30:36 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
448
+ INFO 2025-11-06 18:31:28 in_model.py:236 step:7K smpl:116K ep:885 epch:44.26 loss:0.020 grdn:0.437 lr:8.7e-05 updt_s:1.020 data_s:0.021
449
+ WARNING 2025-11-06 18:31:28 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
450
+ WARNING 2025-11-06 18:31:28 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
451
+ INFO 2025-11-06 18:32:20 in_model.py:236 step:7K smpl:117K ep:891 epch:44.56 loss:0.020 grdn:0.457 lr:8.7e-05 updt_s:1.025 data_s:0.000
452
+ WARNING 2025-11-06 18:32:20 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
453
+ WARNING 2025-11-06 18:32:20 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
454
+ INFO 2025-11-06 18:33:11 in_model.py:236 step:7K smpl:118K ep:897 epch:44.87 loss:0.023 grdn:0.471 lr:8.6e-05 updt_s:1.028 data_s:0.000
455
+ WARNING 2025-11-06 18:33:11 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
456
+ WARNING 2025-11-06 18:33:11 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
457
+ INFO 2025-11-06 18:34:04 in_model.py:236 step:7K smpl:118K ep:903 epch:45.17 loss:0.020 grdn:0.456 lr:8.6e-05 updt_s:1.022 data_s:0.022
458
+ WARNING 2025-11-06 18:34:04 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
459
+ WARNING 2025-11-06 18:34:04 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
460
+ INFO 2025-11-06 18:34:55 in_model.py:236 step:7K smpl:119K ep:910 epch:45.48 loss:0.017 grdn:0.464 lr:8.6e-05 updt_s:1.029 data_s:0.000
461
+ WARNING 2025-11-06 18:34:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
462
+ WARNING 2025-11-06 18:34:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
463
+ INFO 2025-11-06 18:35:47 in_model.py:236 step:8K smpl:120K ep:916 epch:45.78 loss:0.019 grdn:0.429 lr:8.6e-05 updt_s:1.032 data_s:0.000
464
+ WARNING 2025-11-06 18:35:47 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
465
+ WARNING 2025-11-06 18:35:47 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
466
+ INFO 2025-11-06 18:36:39 in_model.py:236 step:8K smpl:121K ep:922 epch:46.09 loss:0.020 grdn:0.434 lr:8.6e-05 updt_s:1.024 data_s:0.022
467
+ WARNING 2025-11-06 18:36:39 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
468
+ WARNING 2025-11-06 18:36:39 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
469
+ INFO 2025-11-06 18:37:31 in_model.py:236 step:8K smpl:122K ep:928 epch:46.39 loss:0.021 grdn:0.469 lr:8.5e-05 updt_s:1.032 data_s:0.000
470
+ WARNING 2025-11-06 18:37:31 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
471
+ WARNING 2025-11-06 18:37:31 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
472
+ INFO 2025-11-06 18:38:23 in_model.py:236 step:8K smpl:122K ep:934 epch:46.70 loss:0.018 grdn:0.417 lr:8.5e-05 updt_s:1.031 data_s:0.000
473
+ WARNING 2025-11-06 18:38:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
474
+ WARNING 2025-11-06 18:38:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
475
+ INFO 2025-11-06 18:39:14 in_model.py:236 step:8K smpl:123K ep:940 epch:47.00 loss:0.017 grdn:0.450 lr:8.5e-05 updt_s:1.027 data_s:0.000
476
+ WARNING 2025-11-06 18:39:14 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
477
+ WARNING 2025-11-06 18:39:14 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
478
+ INFO 2025-11-06 18:40:07 in_model.py:236 step:8K smpl:124K ep:946 epch:47.31 loss:0.019 grdn:0.491 lr:8.5e-05 updt_s:1.019 data_s:0.022
479
+ WARNING 2025-11-06 18:40:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
480
+ WARNING 2025-11-06 18:40:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
481
+ INFO 2025-11-06 18:40:58 in_model.py:236 step:8K smpl:125K ep:952 epch:47.62 loss:0.018 grdn:0.401 lr:8.5e-05 updt_s:1.019 data_s:0.000
482
+ WARNING 2025-11-06 18:40:58 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
483
+ WARNING 2025-11-06 18:40:58 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
484
+ INFO 2025-11-06 18:41:49 in_model.py:236 step:8K smpl:126K ep:958 epch:47.92 loss:0.017 grdn:0.428 lr:8.5e-05 updt_s:1.022 data_s:0.000
485
+ WARNING 2025-11-06 18:41:49 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
486
+ WARNING 2025-11-06 18:41:49 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
487
+ INFO 2025-11-06 18:42:41 in_model.py:236 step:8K smpl:126K ep:965 epch:48.23 loss:0.018 grdn:0.441 lr:8.4e-05 updt_s:1.015 data_s:0.021
488
+ WARNING 2025-11-06 18:42:41 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
489
+ WARNING 2025-11-06 18:42:41 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
490
+ INFO 2025-11-06 18:43:32 in_model.py:236 step:8K smpl:127K ep:971 epch:48.53 loss:0.017 grdn:0.438 lr:8.4e-05 updt_s:1.025 data_s:0.000
491
+ WARNING 2025-11-06 18:43:32 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
492
+ WARNING 2025-11-06 18:43:32 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
493
+ INFO 2025-11-06 18:44:24 in_model.py:236 step:8K smpl:128K ep:977 epch:48.84 loss:0.017 grdn:0.396 lr:8.4e-05 updt_s:1.024 data_s:0.000
494
+ WARNING 2025-11-06 18:44:24 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
495
+ WARNING 2025-11-06 18:44:24 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
496
+ INFO 2025-11-06 18:45:16 in_model.py:236 step:8K smpl:129K ep:983 epch:49.14 loss:0.022 grdn:0.510 lr:8.4e-05 updt_s:1.018 data_s:0.022
497
+ WARNING 2025-11-06 18:45:16 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
498
+ WARNING 2025-11-06 18:45:16 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
499
+ INFO 2025-11-06 18:46:07 in_model.py:236 step:8K smpl:130K ep:989 epch:49.45 loss:0.015 grdn:0.357 lr:8.4e-05 updt_s:1.025 data_s:0.000
500
+ WARNING 2025-11-06 18:46:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
501
+ WARNING 2025-11-06 18:46:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
502
+ INFO 2025-11-06 18:46:58 in_model.py:236 step:8K smpl:130K ep:995 epch:49.75 loss:0.018 grdn:0.442 lr:8.3e-05 updt_s:1.023 data_s:0.000
503
+ WARNING 2025-11-06 18:46:58 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
504
+ WARNING 2025-11-06 18:46:58 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
505
+ INFO 2025-11-06 18:47:50 in_model.py:236 step:8K smpl:131K ep:1K epch:50.06 loss:0.020 grdn:0.416 lr:8.3e-05 updt_s:1.020 data_s:0.000
506
+ WARNING 2025-11-06 18:47:50 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
507
+ WARNING 2025-11-06 18:47:50 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
508
+ INFO 2025-11-06 18:48:42 in_model.py:236 step:8K smpl:132K ep:1K epch:50.36 loss:0.019 grdn:0.433 lr:8.3e-05 updt_s:1.019 data_s:0.021
509
+ WARNING 2025-11-06 18:48:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
510
+ WARNING 2025-11-06 18:48:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
511
+ INFO 2025-11-06 18:49:33 in_model.py:236 step:8K smpl:133K ep:1K epch:50.67 loss:0.018 grdn:0.458 lr:8.3e-05 updt_s:1.020 data_s:0.000
512
+ WARNING 2025-11-06 18:49:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
513
+ WARNING 2025-11-06 18:49:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
514
+ INFO 2025-11-06 18:50:24 in_model.py:236 step:8K smpl:134K ep:1K epch:50.97 loss:0.016 grdn:0.340 lr:8.3e-05 updt_s:1.022 data_s:0.000
515
+ WARNING 2025-11-06 18:50:24 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
516
+ WARNING 2025-11-06 18:50:24 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
517
+ INFO 2025-11-06 18:51:16 in_model.py:236 step:8K smpl:134K ep:1K epch:51.28 loss:0.018 grdn:0.419 lr:8.2e-05 updt_s:1.014 data_s:0.020
518
+ WARNING 2025-11-06 18:51:16 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
519
+ WARNING 2025-11-06 18:51:16 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
520
+ INFO 2025-11-06 18:52:07 in_model.py:236 step:8K smpl:135K ep:1K epch:51.58 loss:0.019 grdn:0.436 lr:8.2e-05 updt_s:1.019 data_s:0.000
521
+ WARNING 2025-11-06 18:52:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
522
+ WARNING 2025-11-06 18:52:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
523
+ INFO 2025-11-06 18:52:59 in_model.py:236 step:8K smpl:136K ep:1K epch:51.89 loss:0.016 grdn:0.383 lr:8.2e-05 updt_s:1.020 data_s:0.000
524
+ WARNING 2025-11-06 18:52:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
525
+ WARNING 2025-11-06 18:52:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
526
+ INFO 2025-11-06 18:53:50 in_model.py:236 step:9K smpl:137K ep:1K epch:52.19 loss:0.019 grdn:0.446 lr:8.2e-05 updt_s:1.012 data_s:0.022
527
+ WARNING 2025-11-06 18:53:50 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
528
+ WARNING 2025-11-06 18:53:50 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
529
+ INFO 2025-11-06 18:54:42 in_model.py:236 step:9K smpl:138K ep:1K epch:52.50 loss:0.018 grdn:0.429 lr:8.2e-05 updt_s:1.022 data_s:0.000
530
+ WARNING 2025-11-06 18:54:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
531
+ WARNING 2025-11-06 18:54:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
532
+ INFO 2025-11-06 18:55:33 in_model.py:236 step:9K smpl:138K ep:1K epch:52.80 loss:0.016 grdn:0.401 lr:8.1e-05 updt_s:1.022 data_s:0.000
533
+ WARNING 2025-11-06 18:55:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
534
+ WARNING 2025-11-06 18:55:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
535
+ INFO 2025-11-06 18:56:25 in_model.py:236 step:9K smpl:139K ep:1K epch:53.11 loss:0.018 grdn:0.394 lr:8.1e-05 updt_s:1.015 data_s:0.020
536
+ WARNING 2025-11-06 18:56:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
537
+ WARNING 2025-11-06 18:56:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
538
+ INFO 2025-11-06 18:57:16 in_model.py:236 step:9K smpl:140K ep:1K epch:53.41 loss:0.018 grdn:0.447 lr:8.1e-05 updt_s:1.018 data_s:0.000
539
+ WARNING 2025-11-06 18:57:16 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
540
+ WARNING 2025-11-06 18:57:16 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
541
+ INFO 2025-11-06 18:58:07 in_model.py:236 step:9K smpl:141K ep:1K epch:53.72 loss:0.018 grdn:0.415 lr:8.1e-05 updt_s:1.021 data_s:0.000
542
+ WARNING 2025-11-06 18:58:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
543
+ WARNING 2025-11-06 18:58:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
544
+ INFO 2025-11-06 18:58:58 in_model.py:236 step:9K smpl:142K ep:1K epch:54.03 loss:0.016 grdn:0.393 lr:8.1e-05 updt_s:1.021 data_s:0.000
545
+ WARNING 2025-11-06 18:58:58 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
546
+ WARNING 2025-11-06 18:58:58 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
547
+ INFO 2025-11-06 18:59:50 in_model.py:236 step:9K smpl:142K ep:1K epch:54.33 loss:0.021 grdn:0.461 lr:8.0e-05 updt_s:1.016 data_s:0.022
548
+ WARNING 2025-11-06 18:59:50 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
549
+ WARNING 2025-11-06 18:59:50 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
550
+ INFO 2025-11-06 19:00:42 in_model.py:236 step:9K smpl:143K ep:1K epch:54.64 loss:0.017 grdn:0.391 lr:8.0e-05 updt_s:1.026 data_s:0.000
551
+ WARNING 2025-11-06 19:00:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
552
+ WARNING 2025-11-06 19:00:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
553
+ INFO 2025-11-06 19:01:33 in_model.py:236 step:9K smpl:144K ep:1K epch:54.94 loss:0.015 grdn:0.400 lr:8.0e-05 updt_s:1.025 data_s:0.000
554
+ WARNING 2025-11-06 19:01:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
555
+ WARNING 2025-11-06 19:01:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
556
+ INFO 2025-11-06 19:02:25 in_model.py:236 step:9K smpl:145K ep:1K epch:55.25 loss:0.017 grdn:0.433 lr:8.0e-05 updt_s:1.017 data_s:0.021
557
+ WARNING 2025-11-06 19:02:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
558
+ WARNING 2025-11-06 19:02:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
559
+ INFO 2025-11-06 19:03:17 in_model.py:236 step:9K smpl:146K ep:1K epch:55.55 loss:0.016 grdn:0.401 lr:8.0e-05 updt_s:1.023 data_s:0.000
560
+ WARNING 2025-11-06 19:03:17 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
561
+ WARNING 2025-11-06 19:03:17 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
562
+ INFO 2025-11-06 19:04:08 in_model.py:236 step:9K smpl:146K ep:1K epch:55.86 loss:0.019 grdn:0.424 lr:7.9e-05 updt_s:1.023 data_s:0.000
563
+ WARNING 2025-11-06 19:04:08 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
564
+ WARNING 2025-11-06 19:04:08 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
565
+ INFO 2025-11-06 19:05:00 in_model.py:236 step:9K smpl:147K ep:1K epch:56.16 loss:0.016 grdn:0.393 lr:7.9e-05 updt_s:1.017 data_s:0.021
566
+ WARNING 2025-11-06 19:05:00 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
567
+ WARNING 2025-11-06 19:05:00 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
568
+ INFO 2025-11-06 19:05:51 in_model.py:236 step:9K smpl:148K ep:1K epch:56.47 loss:0.020 grdn:0.404 lr:7.9e-05 updt_s:1.019 data_s:0.000
569
+ WARNING 2025-11-06 19:05:51 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
570
+ WARNING 2025-11-06 19:05:51 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
571
+ INFO 2025-11-06 19:06:42 in_model.py:236 step:9K smpl:149K ep:1K epch:56.77 loss:0.015 grdn:0.359 lr:7.9e-05 updt_s:1.017 data_s:0.000
572
+ WARNING 2025-11-06 19:06:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
573
+ WARNING 2025-11-06 19:06:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
574
+ INFO 2025-11-06 19:07:34 in_model.py:236 step:9K smpl:150K ep:1K epch:57.08 loss:0.016 grdn:0.374 lr:7.9e-05 updt_s:1.009 data_s:0.020
575
+ WARNING 2025-11-06 19:07:34 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
576
+ WARNING 2025-11-06 19:07:34 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
577
+ INFO 2025-11-06 19:08:25 in_model.py:236 step:9K smpl:150K ep:1K epch:57.38 loss:0.015 grdn:0.369 lr:7.8e-05 updt_s:1.024 data_s:0.000
578
+ WARNING 2025-11-06 19:08:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
579
+ WARNING 2025-11-06 19:08:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
580
+ INFO 2025-11-06 19:09:17 in_model.py:236 step:9K smpl:151K ep:1K epch:57.69 loss:0.016 grdn:0.401 lr:7.8e-05 updt_s:1.038 data_s:0.000
581
+ WARNING 2025-11-06 19:09:17 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
582
+ WARNING 2025-11-06 19:09:17 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
583
+ INFO 2025-11-06 19:10:10 in_model.py:236 step:10K smpl:152K ep:1K epch:57.99 loss:0.018 grdn:0.440 lr:7.8e-05 updt_s:1.053 data_s:0.000
584
+ WARNING 2025-11-06 19:10:10 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
585
+ WARNING 2025-11-06 19:10:10 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
586
+ INFO 2025-11-06 19:11:04 in_model.py:236 step:10K smpl:153K ep:1K epch:58.30 loss:0.013 grdn:0.358 lr:7.8e-05 updt_s:1.051 data_s:0.021
587
+ WARNING 2025-11-06 19:11:04 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
588
+ WARNING 2025-11-06 19:11:04 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
589
+ INFO 2025-11-06 19:11:57 in_model.py:236 step:10K smpl:154K ep:1K epch:58.60 loss:0.015 grdn:0.359 lr:7.7e-05 updt_s:1.062 data_s:0.000
590
+ WARNING 2025-11-06 19:11:57 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
591
+ WARNING 2025-11-06 19:11:57 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
592
+ INFO 2025-11-06 19:12:50 in_model.py:236 step:10K smpl:154K ep:1K epch:58.91 loss:0.015 grdn:0.406 lr:7.7e-05 updt_s:1.061 data_s:0.000
593
+ WARNING 2025-11-06 19:12:50 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
594
+ WARNING 2025-11-06 19:12:50 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
595
+ INFO 2025-11-06 19:13:44 in_model.py:236 step:10K smpl:155K ep:1K epch:59.21 loss:0.014 grdn:0.357 lr:7.7e-05 updt_s:1.055 data_s:0.020
596
+ WARNING 2025-11-06 19:13:44 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
597
+ WARNING 2025-11-06 19:13:44 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
598
+ INFO 2025-11-06 19:14:37 in_model.py:236 step:10K smpl:156K ep:1K epch:59.52 loss:0.018 grdn:0.424 lr:7.7e-05 updt_s:1.061 data_s:0.000
599
+ WARNING 2025-11-06 19:14:37 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
600
+ WARNING 2025-11-06 19:14:37 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
601
+ INFO 2025-11-06 19:15:30 in_model.py:236 step:10K smpl:157K ep:1K epch:59.82 loss:0.017 grdn:0.395 lr:7.7e-05 updt_s:1.062 data_s:0.000
602
+ WARNING 2025-11-06 19:15:30 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
603
+ WARNING 2025-11-06 19:15:30 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
604
+ INFO 2025-11-06 19:16:24 in_model.py:236 step:10K smpl:158K ep:1K epch:60.13 loss:0.017 grdn:0.400 lr:7.6e-05 updt_s:1.056 data_s:0.020
605
+ WARNING 2025-11-06 19:16:24 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
606
+ WARNING 2025-11-06 19:16:24 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
607
+ INFO 2025-11-06 19:17:18 in_model.py:236 step:10K smpl:158K ep:1K epch:60.43 loss:0.015 grdn:0.362 lr:7.6e-05 updt_s:1.061 data_s:0.000
608
+ WARNING 2025-11-06 19:17:18 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
609
+ WARNING 2025-11-06 19:17:18 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
610
+ INFO 2025-11-06 19:18:11 in_model.py:236 step:10K smpl:159K ep:1K epch:60.74 loss:0.016 grdn:0.421 lr:7.6e-05 updt_s:1.064 data_s:0.000
611
+ WARNING 2025-11-06 19:18:11 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
612
+ WARNING 2025-11-06 19:18:11 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
613
+ INFO 2025-11-06 19:19:04 in_model.py:236 step:10K smpl:160K ep:1K epch:61.05 loss:0.015 grdn:0.381 lr:7.6e-05 updt_s:1.064 data_s:0.000
614
+ WARNING 2025-11-06 19:19:04 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
615
+ WARNING 2025-11-06 19:19:04 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
616
+ INFO 2025-11-06 19:19:04 in_model.py:245 Checkpoint policy after step 10000
617
+ INFO 2025-11-06 19:20:01 in_model.py:236 step:10K smpl:161K ep:1K epch:61.35 loss:0.015 grdn:0.362 lr:7.6e-05 updt_s:1.044 data_s:0.023
618
+ WARNING 2025-11-06 19:20:01 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
619
+ WARNING 2025-11-06 19:20:01 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
620
+ INFO 2025-11-06 19:20:54 in_model.py:236 step:10K smpl:162K ep:1K epch:61.66 loss:0.016 grdn:0.378 lr:7.5e-05 updt_s:1.061 data_s:0.000
621
+ WARNING 2025-11-06 19:20:54 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
622
+ WARNING 2025-11-06 19:20:54 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
623
+ INFO 2025-11-06 19:21:47 in_model.py:236 step:10K smpl:162K ep:1K epch:61.96 loss:0.015 grdn:0.408 lr:7.5e-05 updt_s:1.062 data_s:0.000
624
+ WARNING 2025-11-06 19:21:47 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
625
+ WARNING 2025-11-06 19:21:47 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
626
+ INFO 2025-11-06 19:22:41 in_model.py:236 step:10K smpl:163K ep:1K epch:62.27 loss:0.019 grdn:0.436 lr:7.5e-05 updt_s:1.055 data_s:0.020
627
+ WARNING 2025-11-06 19:22:41 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
628
+ WARNING 2025-11-06 19:22:41 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
629
+ INFO 2025-11-06 19:23:35 in_model.py:236 step:10K smpl:164K ep:1K epch:62.57 loss:0.013 grdn:0.339 lr:7.5e-05 updt_s:1.067 data_s:0.000
630
+ WARNING 2025-11-06 19:23:35 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
631
+ WARNING 2025-11-06 19:23:35 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/requirements.txt ADDED
@@ -0,0 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ pyparsing==3.2.5
2
+ nvidia-cuda-cupti-cu12==12.4.127
3
+ pip==25.2
4
+ setuptools==80.9.0
5
+ gitdb==4.0.12
6
+ networkx==3.4.2
7
+ propcache==0.4.1
8
+ aiohappyeyeballs==2.6.1
9
+ aiosignal==1.4.0
10
+ Werkzeug==3.1.3
11
+ nvidia-nvtx-cu12==12.4.127
12
+ typing-inspect==0.9.0
13
+ urllib3==2.5.0
14
+ imageio-ffmpeg==0.6.0
15
+ num2words==0.5.14
16
+ PySocks==1.7.1
17
+ datasets==3.4.1
18
+ PyScreeze==1.0.1
19
+ tokenizers==0.21.4
20
+ yarl==1.22.0
21
+ MouseInfo==0.1.3
22
+ mypy_extensions==1.1.0
23
+ pyzmq==27.1.0
24
+ beautifulsoup4==4.14.2
25
+ importlib_metadata==8.7.0
26
+ evdev==1.9.2
27
+ sentry-sdk==2.43.0
28
+ wcwidth==0.2.14
29
+ typing_extensions==4.15.0
30
+ certifi==2025.10.5
31
+ importlib_resources==6.5.2
32
+ hf-xet==1.2.0
33
+ frozenlist==1.8.0
34
+ annotated-types==0.7.0
35
+ PyGetWindow==0.0.9
36
+ draccus==0.10.0
37
+ rerun-sdk==0.26.2
38
+ cmake==4.1.2
39
+ PyAutoGUI==0.9.54
40
+ gdown==5.2.0
41
+ pyarrow==22.0.0
42
+ nvidia-cusparse-cu12==12.3.1.170
43
+ torch==2.6.0
44
+ cycler==0.12.1
45
+ av==16.0.1
46
+ pytz==2025.2
47
+ tzdata==2025.2
48
+ zipp==3.23.0
49
+ prompt_toolkit==3.0.52
50
+ itsdangerous==2.2.0
51
+ lerobot==0.1.0
52
+ smmap==5.0.2
53
+ pyserial==3.5
54
+ charset-normalizer==3.4.4
55
+ numpy==2.2.6
56
+ psutil==7.1.3
57
+ zarr==2.18.3
58
+ matplotlib==3.10.7
59
+ llvmlite==0.45.1
60
+ torchaudio==2.6.0
61
+ huggingface-hub==0.36.0
62
+ MarkupSafe==3.0.3
63
+ packaging==25.0
64
+ PyYAML==6.0.3
65
+ Flask==3.1.2
66
+ PyOpenGL-accelerate==3.1.10
67
+ hf_transfer==0.1.9
68
+ pydantic_core==2.41.4
69
+ pandas==2.3.3
70
+ mergedeep==1.3.4
71
+ h5py==3.15.1
72
+ transformers==4.50.3
73
+ torchcodec==0.8.1
74
+ termcolor==3.2.0
75
+ toml==0.10.2
76
+ imageio==2.37.0
77
+ nvidia-cuda-nvrtc-cu12==12.4.127
78
+ asciitree==0.3.3
79
+ antlr4-python3-runtime==4.9.3
80
+ jsonlines==4.0.0
81
+ scipy==1.15.3
82
+ pyperclip==1.11.0
83
+ pillow==12.0.0
84
+ click==8.3.0
85
+ torchvision==0.21.0
86
+ cffi==2.0.0
87
+ etils==1.13.0
88
+ Farama-Notifications==0.0.4
89
+ nvidia-cufft-cu12==11.2.1.3
90
+ kiwisolver==1.4.9
91
+ triton==3.2.0
92
+ fonttools==4.60.1
93
+ pycparser==2.23
94
+ absl-py==2.3.1
95
+ requests==2.32.5
96
+ six==1.17.0
97
+ nvidia-nccl-cu12==2.21.5
98
+ pytweening==1.2.0
99
+ docopt==0.6.2
100
+ contourpy==1.3.2
101
+ nvidia-cusolver-cu12==11.6.1.9
102
+ pynput==1.8.1
103
+ Jinja2==3.1.6
104
+ einops==0.8.1
105
+ numcodecs==0.13.1
106
+ async-timeout==5.0.1
107
+ blinker==1.9.0
108
+ pydantic==2.12.3
109
+ xxhash==3.6.0
110
+ cloudpickle==3.1.2
111
+ diffusers==0.35.2
112
+ PyOpenGL==3.1.10
113
+ glfw==2.10.0
114
+ soupsieve==2.8
115
+ GitPython==3.1.45
116
+ fsspec==2024.12.0
117
+ aiohttp==3.13.2
118
+ opencv-python-headless==4.12.0.88
119
+ gymnasium==0.29.1
120
+ python3-xlib==0.15
121
+ mpmath==1.3.0
122
+ nvidia-curand-cu12==10.3.5.147
123
+ wandb==0.22.3
124
+ nvidia-cublas-cu12==12.4.5.8
125
+ nvidia-cuda-runtime-cu12==12.4.127
126
+ python-xlib==0.33
127
+ multiprocess==0.70.16
128
+ pfzy==0.3.4
129
+ orderly-set==5.5.0
130
+ nvidia-nvjitlink-cu12==12.4.127
131
+ accelerate==1.11.0
132
+ regex==2025.11.3
133
+ omegaconf==2.3.0
134
+ safetensors==0.5.3
135
+ filelock==3.20.0
136
+ numba==0.62.1
137
+ nvidia-cusparselt-cu12==0.6.2
138
+ dill==0.3.8
139
+ pyyaml-include==1.4.1
140
+ platformdirs==4.5.0
141
+ multidict==6.7.0
142
+ protobuf==6.33.0
143
+ attrs==25.4.0
144
+ inquirerpy==0.3.4
145
+ PyMsgBox==2.0.1
146
+ pymunk==6.11.1
147
+ nvidia-cudnn-cu12==9.1.0.70
148
+ typing-inspection==0.4.2
149
+ fasteners==0.20
150
+ PyRect==0.2.0
151
+ idna==3.11
152
+ deepdiff==8.6.1
153
+ wheel==0.45.1
154
+ sympy==1.13.1
155
+ python-dateutil==2.9.0.post0
156
+ mujoco==3.1.6
157
+ tqdm==4.67.1
158
+ backports.tarfile==1.2.0
159
+ typeguard==4.3.0
160
+ zipp==3.19.2
161
+ inflect==7.3.1
162
+ more-itertools==10.3.0
163
+ tomli==2.0.1
164
+ jaraco.collections==5.1.0
165
+ jaraco.functools==4.0.1
166
+ typing_extensions==4.12.2
167
+ importlib_metadata==8.0.0
168
+ jaraco.text==3.12.1
169
+ platformdirs==4.2.2
170
+ jaraco.context==5.3.0
171
+ wheel==0.45.1
172
+ packaging==24.2
173
+ autocommand==2.2.2
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/wandb-metadata.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-51-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.10.19",
4
+ "startedAt": "2025-11-06T08:21:49.583849Z",
5
+ "args": [
6
+ "--config_path",
7
+ "smolvla_omy.yaml"
8
+ ],
9
+ "program": "/workspace/liber/embodied_ai/lerobot-mujoco-tutorial/train_model.py",
10
+ "codePath": "train_model.py",
11
+ "codePathLocal": "train_model.py",
12
+ "git": {
13
+ "remote": "git@github.com:liberow/lerobot-mujoco-tutorial.git",
14
+ "commit": "b11377a0ce7a3c32cefe260c6dce959b0db2bdda"
15
+ },
16
+ "root": "ckpt/smolvla_omy",
17
+ "host": "c81ab3b21da2",
18
+ "executable": "/opt/conda/envs/lerobot_mujoco_tutorial/bin/python",
19
+ "cpu_count": 40,
20
+ "cpu_count_logical": 80,
21
+ "gpu": "Tesla V100-SXM2-32GB",
22
+ "gpu_count": 4,
23
+ "disk": {
24
+ "/": {
25
+ "total": "1676411871232",
26
+ "used": "1481895833600"
27
+ }
28
+ },
29
+ "memory": {
30
+ "total": "134767624192"
31
+ },
32
+ "gpu_nvidia": [
33
+ {
34
+ "name": "Tesla V100-SXM2-32GB",
35
+ "memoryTotal": "34359738368",
36
+ "cudaCores": 5120,
37
+ "architecture": "Volta",
38
+ "uuid": "GPU-b29b46dc-42a3-9fcc-be17-50b8b8972960"
39
+ },
40
+ {
41
+ "name": "Tesla V100-SXM2-32GB",
42
+ "memoryTotal": "34359738368",
43
+ "cudaCores": 5120,
44
+ "architecture": "Volta",
45
+ "uuid": "GPU-368aa17c-5504-4d63-4df2-339d38a17fe3"
46
+ },
47
+ {
48
+ "name": "Tesla V100-SXM2-32GB",
49
+ "memoryTotal": "34359738368",
50
+ "cudaCores": 5120,
51
+ "architecture": "Volta",
52
+ "uuid": "GPU-acf7b3cc-f154-087a-3a06-d57e023c09be"
53
+ },
54
+ {
55
+ "name": "Tesla V100-SXM2-32GB",
56
+ "memoryTotal": "34359738368",
57
+ "cudaCores": 5120,
58
+ "architecture": "Volta",
59
+ "uuid": "GPU-86b115b3-0485-dd33-56f3-385a1140ae36"
60
+ }
61
+ ],
62
+ "cudaVersion": "12.4",
63
+ "writerId": "qj4c6zn4buu7s32sv5nl108gvz7rnfa0"
64
+ }
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-core.log ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"time":"2025-11-06T16:21:49.620623718+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpfd2c2eff/port-528220.txt","pid":528220,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-11-06T16:21:49.621744188+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":528220}
3
+ {"time":"2025-11-06T16:21:49.621734866+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-528220-528551-3638943209/socket","Net":"unix"}}
4
+ {"time":"2025-11-06T16:21:49.795665893+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-11-06T16:21:49.810358292+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"8gs9of7r","id":"1(@)"}
6
+ {"time":"2025-11-06T16:21:50.381980279+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"8gs9of7r","id":"1(@)"}
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-internal.log ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-06T16:21:49.810658746+08:00","level":"INFO","msg":"stream: starting","core version":"0.22.3"}
2
+ {"time":"2025-11-06T16:21:50.381496059+08:00","level":"INFO","msg":"stream: created new stream","id":"8gs9of7r"}
3
+ {"time":"2025-11-06T16:21:50.381670461+08:00","level":"INFO","msg":"handler: started","stream_id":"8gs9of7r"}
4
+ {"time":"2025-11-06T16:21:50.38195187+08:00","level":"INFO","msg":"stream: started","id":"8gs9of7r"}
5
+ {"time":"2025-11-06T16:21:50.381970995+08:00","level":"INFO","msg":"writer: started","stream_id":"8gs9of7r"}
6
+ {"time":"2025-11-06T16:21:50.381975054+08:00","level":"INFO","msg":"sender: started","stream_id":"8gs9of7r"}
7
+ {"time":"2025-11-06T18:28:51.825582909+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
8
+ {"time":"2025-11-06T19:15:51.940995974+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug.log ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 16:21:49,585 INFO MainThread:528220 [wandb_setup.py:_flush():81] Current SDK version is 0.22.3
2
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Configure stats pid to 528220
3
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Loading settings from /root/.config/wandb/settings
4
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Loading settings from /workspace/liber/embodied_ai/lerobot-mujoco-tutorial/wandb/settings
5
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:setup_run_log_directory():706] Logging user logs to ckpt/smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug.log
7
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:setup_run_log_directory():707] Logging internal logs to ckpt/smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-internal.log
8
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:init():833] calling init triggers
9
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:init():838] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'omy_pnp_language', 'root': './demo_data_language', 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec'}, 'env': None, 'policy': {'type': 'smolvla', 'n_obs_steps': 1, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'chunk_size': 5, 'n_action_steps': 5, 'max_state_dim': 32, 'max_action_dim': 32, 'resize_imgs_with_padding': [512, 512], 'empty_cameras': 0, 'adapt_to_pi_aloha': False, 'use_delta_joint_actions_aloha': False, 'tokenizer_max_length': 48, 'num_steps': 10, 'use_cache': True, 'freeze_vision_encoder': True, 'train_expert_only': True, 'train_state_proj': True, 'optimizer_lr': 0.0001, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 1e-10, 'optimizer_grad_clip_norm': 10, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'vlm_model_name': 'HuggingFaceTB/SmolVLM2-500M-Video-Instruct', 'load_vlm_weights': False, 'add_image_special_tokens': False, 'attention_mode': 'cross_attn', 'prefix_length': -1, 'pad_language_to': 'longest', 'num_expert_layers': -1, 'num_vlm_layers': 16, 'self_attn_every_n_layers': 2, 'expert_width_multiplier': 0.75, 'min_period': 0.004, 'max_period': 4.0}, 'output_dir': 'ckpt/smolvla_omy', 'job_name': 'smolvla_omy', 'resume': False, 'seed': 42, 'num_workers': 8, 'batch_size': 16, 'steps': 20000, 'eval_freq': -1, 'log_freq': 50, 'save_checkpoint': True, 'save_freq': 5000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 0.0001, 'weight_decay': 1e-10, 'grad_clip_norm': 10, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 0.0001, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': True, 'project': 'smolvla_omy', 'entity': 'ai-liber', 'notes': None, 'run_id': None, 'mode': None}, '_wandb': {}}
11
+ 2025-11-06 16:21:49,586 INFO MainThread:528220 [wandb_init.py:init():881] starting backend
12
+ 2025-11-06 16:21:49,796 INFO MainThread:528220 [wandb_init.py:init():884] sending inform_init request
13
+ 2025-11-06 16:21:49,803 INFO MainThread:528220 [wandb_init.py:init():892] backend started and connected
14
+ 2025-11-06 16:21:49,806 INFO MainThread:528220 [wandb_init.py:init():962] updated telemetry
15
+ 2025-11-06 16:21:49,819 INFO MainThread:528220 [wandb_init.py:init():986] communicating run to backend with 90.0 second timeout
16
+ 2025-11-06 16:21:51,055 INFO MainThread:528220 [wandb_init.py:init():1033] starting run threads in backend
17
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-11-06 16:21:51,416 INFO MainThread:528220 [wandb_init.py:init():1073] run started, returning control to user process
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/run-8gs9of7r.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:122f2a5f19ad90752dff6274a7d5884caf0566af458d12f45841bd744233e332
3
+ size 1277952
smolvla_omy_2000_steps/checkpoints/002000/pretrained_model/config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "normalization_mapping": {
5
+ "VISUAL": "IDENTITY",
6
+ "STATE": "MEAN_STD",
7
+ "ACTION": "MEAN_STD"
8
+ },
9
+ "input_features": {
10
+ "observation.image": {
11
+ "type": "VISUAL",
12
+ "shape": [
13
+ 3,
14
+ 256,
15
+ 256
16
+ ]
17
+ },
18
+ "observation.wrist_image": {
19
+ "type": "VISUAL",
20
+ "shape": [
21
+ 3,
22
+ 256,
23
+ 256
24
+ ]
25
+ },
26
+ "observation.state": {
27
+ "type": "STATE",
28
+ "shape": [
29
+ 6
30
+ ]
31
+ }
32
+ },
33
+ "output_features": {
34
+ "action": {
35
+ "type": "ACTION",
36
+ "shape": [
37
+ 7
38
+ ]
39
+ }
40
+ },
41
+ "device": "cuda",
42
+ "use_amp": false,
43
+ "chunk_size": 5,
44
+ "n_action_steps": 5,
45
+ "max_state_dim": 32,
46
+ "max_action_dim": 32,
47
+ "resize_imgs_with_padding": [
48
+ 512,
49
+ 512
50
+ ],
51
+ "empty_cameras": 0,
52
+ "adapt_to_pi_aloha": false,
53
+ "use_delta_joint_actions_aloha": false,
54
+ "tokenizer_max_length": 48,
55
+ "num_steps": 10,
56
+ "use_cache": true,
57
+ "freeze_vision_encoder": true,
58
+ "train_expert_only": true,
59
+ "train_state_proj": true,
60
+ "optimizer_lr": 0.0001,
61
+ "optimizer_betas": [
62
+ 0.9,
63
+ 0.95
64
+ ],
65
+ "optimizer_eps": 1e-08,
66
+ "optimizer_weight_decay": 1e-10,
67
+ "optimizer_grad_clip_norm": 10,
68
+ "scheduler_warmup_steps": 1000,
69
+ "scheduler_decay_steps": 30000,
70
+ "scheduler_decay_lr": 2.5e-06,
71
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
72
+ "load_vlm_weights": false,
73
+ "add_image_special_tokens": false,
74
+ "attention_mode": "cross_attn",
75
+ "prefix_length": -1,
76
+ "pad_language_to": "longest",
77
+ "num_expert_layers": -1,
78
+ "num_vlm_layers": 16,
79
+ "self_attn_every_n_layers": 2,
80
+ "expert_width_multiplier": 0.75,
81
+ "min_period": 0.004,
82
+ "max_period": 4.0
83
+ }
smolvla_omy_2000_steps/checkpoints/002000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:392c8b3292992bf112b346682fc469f9211a29aca9b37296a58fb8258502af41
3
+ size 1197790032
smolvla_omy_2000_steps/checkpoints/002000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "omy_pnp_language",
4
+ "root": "./demo_data_language",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "revision": null,
64
+ "use_imagenet_stats": true,
65
+ "video_backend": "torchcodec"
66
+ },
67
+ "env": null,
68
+ "policy": {
69
+ "type": "smolvla",
70
+ "n_obs_steps": 1,
71
+ "normalization_mapping": {
72
+ "VISUAL": "IDENTITY",
73
+ "STATE": "MEAN_STD",
74
+ "ACTION": "MEAN_STD"
75
+ },
76
+ "input_features": {
77
+ "observation.image": {
78
+ "type": "VISUAL",
79
+ "shape": [
80
+ 3,
81
+ 256,
82
+ 256
83
+ ]
84
+ },
85
+ "observation.wrist_image": {
86
+ "type": "VISUAL",
87
+ "shape": [
88
+ 3,
89
+ 256,
90
+ 256
91
+ ]
92
+ },
93
+ "observation.state": {
94
+ "type": "STATE",
95
+ "shape": [
96
+ 6
97
+ ]
98
+ }
99
+ },
100
+ "output_features": {
101
+ "action": {
102
+ "type": "ACTION",
103
+ "shape": [
104
+ 7
105
+ ]
106
+ }
107
+ },
108
+ "device": "cuda",
109
+ "use_amp": false,
110
+ "chunk_size": 5,
111
+ "n_action_steps": 5,
112
+ "max_state_dim": 32,
113
+ "max_action_dim": 32,
114
+ "resize_imgs_with_padding": [
115
+ 512,
116
+ 512
117
+ ],
118
+ "empty_cameras": 0,
119
+ "adapt_to_pi_aloha": false,
120
+ "use_delta_joint_actions_aloha": false,
121
+ "tokenizer_max_length": 48,
122
+ "num_steps": 10,
123
+ "use_cache": true,
124
+ "freeze_vision_encoder": true,
125
+ "train_expert_only": true,
126
+ "train_state_proj": true,
127
+ "optimizer_lr": 0.0001,
128
+ "optimizer_betas": [
129
+ 0.9,
130
+ 0.95
131
+ ],
132
+ "optimizer_eps": 1e-08,
133
+ "optimizer_weight_decay": 1e-10,
134
+ "optimizer_grad_clip_norm": 10,
135
+ "scheduler_warmup_steps": 1000,
136
+ "scheduler_decay_steps": 30000,
137
+ "scheduler_decay_lr": 2.5e-06,
138
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
139
+ "load_vlm_weights": false,
140
+ "add_image_special_tokens": false,
141
+ "attention_mode": "cross_attn",
142
+ "prefix_length": -1,
143
+ "pad_language_to": "longest",
144
+ "num_expert_layers": -1,
145
+ "num_vlm_layers": 16,
146
+ "self_attn_every_n_layers": 2,
147
+ "expert_width_multiplier": 0.75,
148
+ "min_period": 0.004,
149
+ "max_period": 4.0
150
+ },
151
+ "output_dir": "ckpt/smolvla_omy",
152
+ "job_name": "smolvla_omy",
153
+ "resume": false,
154
+ "seed": 42,
155
+ "num_workers": 8,
156
+ "batch_size": 16,
157
+ "steps": 2000,
158
+ "eval_freq": -1,
159
+ "log_freq": 50,
160
+ "save_checkpoint": true,
161
+ "save_freq": 10000,
162
+ "use_policy_training_preset": true,
163
+ "optimizer": {
164
+ "type": "adamw",
165
+ "lr": 0.0001,
166
+ "weight_decay": 1e-10,
167
+ "grad_clip_norm": 10,
168
+ "betas": [
169
+ 0.9,
170
+ 0.95
171
+ ],
172
+ "eps": 1e-08
173
+ },
174
+ "scheduler": {
175
+ "type": "cosine_decay_with_warmup",
176
+ "num_warmup_steps": 1000,
177
+ "num_decay_steps": 30000,
178
+ "peak_lr": 0.0001,
179
+ "decay_lr": 2.5e-06
180
+ },
181
+ "eval": {
182
+ "n_episodes": 50,
183
+ "batch_size": 50,
184
+ "use_async_envs": false
185
+ },
186
+ "wandb": {
187
+ "enable": true,
188
+ "disable_artifact": true,
189
+ "project": "smolvla_omy",
190
+ "entity": "ai-liber",
191
+ "notes": null,
192
+ "run_id": null,
193
+ "mode": null
194
+ }
195
+ }
smolvla_omy_2000_steps/checkpoints/002000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,526 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 9.893469553577303e-05,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "foreach": null,
12
+ "maximize": false,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "initial_lr": 0.0001,
17
+ "params": [
18
+ 0,
19
+ 1,
20
+ 2,
21
+ 3,
22
+ 4,
23
+ 5,
24
+ 6,
25
+ 7,
26
+ 8,
27
+ 9,
28
+ 10,
29
+ 11,
30
+ 12,
31
+ 13,
32
+ 14,
33
+ 15,
34
+ 16,
35
+ 17,
36
+ 18,
37
+ 19,
38
+ 20,
39
+ 21,
40
+ 22,
41
+ 23,
42
+ 24,
43
+ 25,
44
+ 26,
45
+ 27,
46
+ 28,
47
+ 29,
48
+ 30,
49
+ 31,
50
+ 32,
51
+ 33,
52
+ 34,
53
+ 35,
54
+ 36,
55
+ 37,
56
+ 38,
57
+ 39,
58
+ 40,
59
+ 41,
60
+ 42,
61
+ 43,
62
+ 44,
63
+ 45,
64
+ 46,
65
+ 47,
66
+ 48,
67
+ 49,
68
+ 50,
69
+ 51,
70
+ 52,
71
+ 53,
72
+ 54,
73
+ 55,
74
+ 56,
75
+ 57,
76
+ 58,
77
+ 59,
78
+ 60,
79
+ 61,
80
+ 62,
81
+ 63,
82
+ 64,
83
+ 65,
84
+ 66,
85
+ 67,
86
+ 68,
87
+ 69,
88
+ 70,
89
+ 71,
90
+ 72,
91
+ 73,
92
+ 74,
93
+ 75,
94
+ 76,
95
+ 77,
96
+ 78,
97
+ 79,
98
+ 80,
99
+ 81,
100
+ 82,
101
+ 83,
102
+ 84,
103
+ 85,
104
+ 86,
105
+ 87,
106
+ 88,
107
+ 89,
108
+ 90,
109
+ 91,
110
+ 92,
111
+ 93,
112
+ 94,
113
+ 95,
114
+ 96,
115
+ 97,
116
+ 98,
117
+ 99,
118
+ 100,
119
+ 101,
120
+ 102,
121
+ 103,
122
+ 104,
123
+ 105,
124
+ 106,
125
+ 107,
126
+ 108,
127
+ 109,
128
+ 110,
129
+ 111,
130
+ 112,
131
+ 113,
132
+ 114,
133
+ 115,
134
+ 116,
135
+ 117,
136
+ 118,
137
+ 119,
138
+ 120,
139
+ 121,
140
+ 122,
141
+ 123,
142
+ 124,
143
+ 125,
144
+ 126,
145
+ 127,
146
+ 128,
147
+ 129,
148
+ 130,
149
+ 131,
150
+ 132,
151
+ 133,
152
+ 134,
153
+ 135,
154
+ 136,
155
+ 137,
156
+ 138,
157
+ 139,
158
+ 140,
159
+ 141,
160
+ 142,
161
+ 143,
162
+ 144,
163
+ 145,
164
+ 146,
165
+ 147,
166
+ 148,
167
+ 149,
168
+ 150,
169
+ 151,
170
+ 152,
171
+ 153,
172
+ 154,
173
+ 155,
174
+ 156,
175
+ 157,
176
+ 158,
177
+ 159,
178
+ 160,
179
+ 161,
180
+ 162,
181
+ 163,
182
+ 164,
183
+ 165,
184
+ 166,
185
+ 167,
186
+ 168,
187
+ 169,
188
+ 170,
189
+ 171,
190
+ 172,
191
+ 173,
192
+ 174,
193
+ 175,
194
+ 176,
195
+ 177,
196
+ 178,
197
+ 179,
198
+ 180,
199
+ 181,
200
+ 182,
201
+ 183,
202
+ 184,
203
+ 185,
204
+ 186,
205
+ 187,
206
+ 188,
207
+ 189,
208
+ 190,
209
+ 191,
210
+ 192,
211
+ 193,
212
+ 194,
213
+ 195,
214
+ 196,
215
+ 197,
216
+ 198,
217
+ 199,
218
+ 200,
219
+ 201,
220
+ 202,
221
+ 203,
222
+ 204,
223
+ 205,
224
+ 206,
225
+ 207,
226
+ 208,
227
+ 209,
228
+ 210,
229
+ 211,
230
+ 212,
231
+ 213,
232
+ 214,
233
+ 215,
234
+ 216,
235
+ 217,
236
+ 218,
237
+ 219,
238
+ 220,
239
+ 221,
240
+ 222,
241
+ 223,
242
+ 224,
243
+ 225,
244
+ 226,
245
+ 227,
246
+ 228,
247
+ 229,
248
+ 230,
249
+ 231,
250
+ 232,
251
+ 233,
252
+ 234,
253
+ 235,
254
+ 236,
255
+ 237,
256
+ 238,
257
+ 239,
258
+ 240,
259
+ 241,
260
+ 242,
261
+ 243,
262
+ 244,
263
+ 245,
264
+ 246,
265
+ 247,
266
+ 248,
267
+ 249,
268
+ 250,
269
+ 251,
270
+ 252,
271
+ 253,
272
+ 254,
273
+ 255,
274
+ 256,
275
+ 257,
276
+ 258,
277
+ 259,
278
+ 260,
279
+ 261,
280
+ 262,
281
+ 263,
282
+ 264,
283
+ 265,
284
+ 266,
285
+ 267,
286
+ 268,
287
+ 269,
288
+ 270,
289
+ 271,
290
+ 272,
291
+ 273,
292
+ 274,
293
+ 275,
294
+ 276,
295
+ 277,
296
+ 278,
297
+ 279,
298
+ 280,
299
+ 281,
300
+ 282,
301
+ 283,
302
+ 284,
303
+ 285,
304
+ 286,
305
+ 287,
306
+ 288,
307
+ 289,
308
+ 290,
309
+ 291,
310
+ 292,
311
+ 293,
312
+ 294,
313
+ 295,
314
+ 296,
315
+ 297,
316
+ 298,
317
+ 299,
318
+ 300,
319
+ 301,
320
+ 302,
321
+ 303,
322
+ 304,
323
+ 305,
324
+ 306,
325
+ 307,
326
+ 308,
327
+ 309,
328
+ 310,
329
+ 311,
330
+ 312,
331
+ 313,
332
+ 314,
333
+ 315,
334
+ 316,
335
+ 317,
336
+ 318,
337
+ 319,
338
+ 320,
339
+ 321,
340
+ 322,
341
+ 323,
342
+ 324,
343
+ 325,
344
+ 326,
345
+ 327,
346
+ 328,
347
+ 329,
348
+ 330,
349
+ 331,
350
+ 332,
351
+ 333,
352
+ 334,
353
+ 335,
354
+ 336,
355
+ 337,
356
+ 338,
357
+ 339,
358
+ 340,
359
+ 341,
360
+ 342,
361
+ 343,
362
+ 344,
363
+ 345,
364
+ 346,
365
+ 347,
366
+ 348,
367
+ 349,
368
+ 350,
369
+ 351,
370
+ 352,
371
+ 353,
372
+ 354,
373
+ 355,
374
+ 356,
375
+ 357,
376
+ 358,
377
+ 359,
378
+ 360,
379
+ 361,
380
+ 362,
381
+ 363,
382
+ 364,
383
+ 365,
384
+ 366,
385
+ 367,
386
+ 368,
387
+ 369,
388
+ 370,
389
+ 371,
390
+ 372,
391
+ 373,
392
+ 374,
393
+ 375,
394
+ 376,
395
+ 377,
396
+ 378,
397
+ 379,
398
+ 380,
399
+ 381,
400
+ 382,
401
+ 383,
402
+ 384,
403
+ 385,
404
+ 386,
405
+ 387,
406
+ 388,
407
+ 389,
408
+ 390,
409
+ 391,
410
+ 392,
411
+ 393,
412
+ 394,
413
+ 395,
414
+ 396,
415
+ 397,
416
+ 398,
417
+ 399,
418
+ 400,
419
+ 401,
420
+ 402,
421
+ 403,
422
+ 404,
423
+ 405,
424
+ 406,
425
+ 407,
426
+ 408,
427
+ 409,
428
+ 410,
429
+ 411,
430
+ 412,
431
+ 413,
432
+ 414,
433
+ 415,
434
+ 416,
435
+ 417,
436
+ 418,
437
+ 419,
438
+ 420,
439
+ 421,
440
+ 422,
441
+ 423,
442
+ 424,
443
+ 425,
444
+ 426,
445
+ 427,
446
+ 428,
447
+ 429,
448
+ 430,
449
+ 431,
450
+ 432,
451
+ 433,
452
+ 434,
453
+ 435,
454
+ 436,
455
+ 437,
456
+ 438,
457
+ 439,
458
+ 440,
459
+ 441,
460
+ 442,
461
+ 443,
462
+ 444,
463
+ 445,
464
+ 446,
465
+ 447,
466
+ 448,
467
+ 449,
468
+ 450,
469
+ 451,
470
+ 452,
471
+ 453,
472
+ 454,
473
+ 455,
474
+ 456,
475
+ 457,
476
+ 458,
477
+ 459,
478
+ 460,
479
+ 461,
480
+ 462,
481
+ 463,
482
+ 464,
483
+ 465,
484
+ 466,
485
+ 467,
486
+ 468,
487
+ 469,
488
+ 470,
489
+ 471,
490
+ 472,
491
+ 473,
492
+ 474,
493
+ 475,
494
+ 476,
495
+ 477,
496
+ 478,
497
+ 479,
498
+ 480,
499
+ 481,
500
+ 482,
501
+ 483,
502
+ 484,
503
+ 485,
504
+ 486,
505
+ 487,
506
+ 488,
507
+ 489,
508
+ 490,
509
+ 491,
510
+ 492,
511
+ 493,
512
+ 494,
513
+ 495,
514
+ 496,
515
+ 497,
516
+ 498,
517
+ 499,
518
+ 500,
519
+ 501,
520
+ 502,
521
+ 503,
522
+ 504,
523
+ 505
524
+ ]
525
+ }
526
+ ]
smolvla_omy_2000_steps/checkpoints/002000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:841e1fea0d64c05cdf1969c6274cba7a592c53080fa0dc014ba397adfde48677
3
+ size 412659164
smolvla_omy_2000_steps/checkpoints/002000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1408160b155291236d7d16ec03ad17e4cb8484650560e2bea39128b5919b5f1
3
+ size 15708
smolvla_omy_2000_steps/checkpoints/002000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 2000,
6
+ "verbose": false,
7
+ "_step_count": 2001,
8
+ "_get_lr_called_within_step": false,
9
+ "_last_lr": [
10
+ 9.893469553577303e-05
11
+ ],
12
+ "lr_lambdas": [
13
+ null
14
+ ]
15
+ }
smolvla_omy_2000_steps/checkpoints/002000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 2000
3
+ }
smolvla_omy_2000_steps/wandb/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-06T15:12:53.085119529+08:00","level":"INFO","msg":"stream: starting","core version":"0.22.3"}
2
+ {"time":"2025-11-06T15:12:53.65199406+08:00","level":"INFO","msg":"stream: created new stream","id":"p7vw7943"}
3
+ {"time":"2025-11-06T15:12:53.652165503+08:00","level":"INFO","msg":"handler: started","stream_id":"p7vw7943"}
4
+ {"time":"2025-11-06T15:12:53.652453415+08:00","level":"INFO","msg":"stream: started","id":"p7vw7943"}
5
+ {"time":"2025-11-06T15:12:53.652494895+08:00","level":"INFO","msg":"writer: started","stream_id":"p7vw7943"}
6
+ {"time":"2025-11-06T15:12:53.652495951+08:00","level":"INFO","msg":"sender: started","stream_id":"p7vw7943"}
7
+ {"time":"2025-11-06T15:50:50.741281958+08:00","level":"INFO","msg":"stream: closing","id":"p7vw7943"}
8
+ {"time":"2025-11-06T15:50:52.335321515+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2025-11-06T15:50:52.795892747+08:00","level":"INFO","msg":"handler: closed","stream_id":"p7vw7943"}
10
+ {"time":"2025-11-06T15:50:52.796103474+08:00","level":"INFO","msg":"sender: closed","stream_id":"p7vw7943"}
11
+ {"time":"2025-11-06T15:50:52.796134886+08:00","level":"INFO","msg":"stream: closed","id":"p7vw7943"}
smolvla_omy_2000_steps/wandb/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 15:12:52,857 INFO MainThread:343428 [wandb_setup.py:_flush():81] Current SDK version is 0.22.3
2
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Configure stats pid to 343428
3
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Loading settings from /root/.config/wandb/settings
4
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Loading settings from /workspace/liber/embodied_ai/lerobot-mujoco-tutorial/wandb/settings
5
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:setup_run_log_directory():706] Logging user logs to ckpt/smolvla_omy/wandb/run-20251106_151252-p7vw7943/logs/debug.log
7
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:setup_run_log_directory():707] Logging internal logs to ckpt/smolvla_omy/wandb/run-20251106_151252-p7vw7943/logs/debug-internal.log
8
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:init():833] calling init triggers
9
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:init():838] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'omy_pnp_language', 'root': './demo_data_language', 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec'}, 'env': None, 'policy': {'type': 'smolvla', 'n_obs_steps': 1, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'chunk_size': 5, 'n_action_steps': 5, 'max_state_dim': 32, 'max_action_dim': 32, 'resize_imgs_with_padding': [512, 512], 'empty_cameras': 0, 'adapt_to_pi_aloha': False, 'use_delta_joint_actions_aloha': False, 'tokenizer_max_length': 48, 'num_steps': 10, 'use_cache': True, 'freeze_vision_encoder': True, 'train_expert_only': True, 'train_state_proj': True, 'optimizer_lr': 0.0001, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 1e-10, 'optimizer_grad_clip_norm': 10, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'vlm_model_name': 'HuggingFaceTB/SmolVLM2-500M-Video-Instruct', 'load_vlm_weights': False, 'add_image_special_tokens': False, 'attention_mode': 'cross_attn', 'prefix_length': -1, 'pad_language_to': 'longest', 'num_expert_layers': -1, 'num_vlm_layers': 16, 'self_attn_every_n_layers': 2, 'expert_width_multiplier': 0.75, 'min_period': 0.004, 'max_period': 4.0}, 'output_dir': 'ckpt/smolvla_omy', 'job_name': 'smolvla_omy', 'resume': False, 'seed': 42, 'num_workers': 8, 'batch_size': 16, 'steps': 2000, 'eval_freq': -1, 'log_freq': 50, 'save_checkpoint': True, 'save_freq': 10000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 0.0001, 'weight_decay': 1e-10, 'grad_clip_norm': 10, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 0.0001, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': True, 'project': 'smolvla_omy', 'entity': 'ai-liber', 'notes': None, 'run_id': None, 'mode': None}, '_wandb': {}}
11
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:init():881] starting backend
12
+ 2025-11-06 15:12:53,069 INFO MainThread:343428 [wandb_init.py:init():884] sending inform_init request
13
+ 2025-11-06 15:12:53,079 INFO MainThread:343428 [wandb_init.py:init():892] backend started and connected
14
+ 2025-11-06 15:12:53,082 INFO MainThread:343428 [wandb_init.py:init():962] updated telemetry
15
+ 2025-11-06 15:12:53,094 INFO MainThread:343428 [wandb_init.py:init():986] communicating run to backend with 90.0 second timeout
16
+ 2025-11-06 15:12:54,336 INFO MainThread:343428 [wandb_init.py:init():1033] starting run threads in backend
17
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-11-06 15:12:54,689 INFO MainThread:343428 [wandb_init.py:init():1073] run started, returning control to user process
22
+ 2025-11-06 15:50:50,741 INFO wandb-AsyncioManager-main:343428 [service_client.py:_forward_responses():80] Reached EOF.
23
+ 2025-11-06 15:50:50,742 INFO wandb-AsyncioManager-main:343428 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles.
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/config.yaml ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.22.3
4
+ e:
5
+ s9gajpxnrqtnt5al3g989j4cneoydvqt:
6
+ args:
7
+ - --config_path
8
+ - smolvla_omy.yaml
9
+ codePath: train_model.py
10
+ codePathLocal: train_model.py
11
+ cpu_count: 40
12
+ cpu_count_logical: 80
13
+ cudaVersion: "12.4"
14
+ disk:
15
+ /:
16
+ total: "1676411871232"
17
+ used: "1480236986368"
18
+ executable: /opt/conda/envs/lerobot_mujoco_tutorial/bin/python
19
+ git:
20
+ commit: b11377a0ce7a3c32cefe260c6dce959b0db2bdda
21
+ remote: git@github.com:liberow/lerobot-mujoco-tutorial.git
22
+ gpu: Tesla V100-SXM2-32GB
23
+ gpu_count: 4
24
+ gpu_nvidia:
25
+ - architecture: Volta
26
+ cudaCores: 5120
27
+ memoryTotal: "34359738368"
28
+ name: Tesla V100-SXM2-32GB
29
+ uuid: GPU-b29b46dc-42a3-9fcc-be17-50b8b8972960
30
+ - architecture: Volta
31
+ cudaCores: 5120
32
+ memoryTotal: "34359738368"
33
+ name: Tesla V100-SXM2-32GB
34
+ uuid: GPU-368aa17c-5504-4d63-4df2-339d38a17fe3
35
+ - architecture: Volta
36
+ cudaCores: 5120
37
+ memoryTotal: "34359738368"
38
+ name: Tesla V100-SXM2-32GB
39
+ uuid: GPU-acf7b3cc-f154-087a-3a06-d57e023c09be
40
+ - architecture: Volta
41
+ cudaCores: 5120
42
+ memoryTotal: "34359738368"
43
+ name: Tesla V100-SXM2-32GB
44
+ uuid: GPU-86b115b3-0485-dd33-56f3-385a1140ae36
45
+ host: c81ab3b21da2
46
+ memory:
47
+ total: "134767624192"
48
+ os: Linux-6.8.0-51-generic-x86_64-with-glibc2.35
49
+ program: /workspace/liber/embodied_ai/lerobot-mujoco-tutorial/train_model.py
50
+ python: CPython 3.10.19
51
+ root: ckpt/smolvla_omy
52
+ startedAt: "2025-11-06T07:12:52.855818Z"
53
+ writerId: s9gajpxnrqtnt5al3g989j4cneoydvqt
54
+ m: []
55
+ python_version: 3.10.19
56
+ t:
57
+ "1":
58
+ - 1
59
+ - 41
60
+ - 49
61
+ - 51
62
+ "2":
63
+ - 1
64
+ - 11
65
+ - 41
66
+ - 49
67
+ - 51
68
+ - 71
69
+ "3":
70
+ - 13
71
+ - 15
72
+ - 16
73
+ - 61
74
+ "4": 3.10.19
75
+ "5": 0.22.3
76
+ "10":
77
+ - 21
78
+ "12": 0.22.3
79
+ "13": linux-x86_64
80
+ batch_size:
81
+ value: 16
82
+ dataset:
83
+ value:
84
+ episodes: null
85
+ image_transforms:
86
+ enable: false
87
+ max_num_transforms: 3
88
+ random_order: false
89
+ tfs:
90
+ brightness:
91
+ kwargs:
92
+ brightness:
93
+ - 0.8
94
+ - 1.2
95
+ type: ColorJitter
96
+ weight: 1
97
+ contrast:
98
+ kwargs:
99
+ contrast:
100
+ - 0.8
101
+ - 1.2
102
+ type: ColorJitter
103
+ weight: 1
104
+ hue:
105
+ kwargs:
106
+ hue:
107
+ - -0.05
108
+ - 0.05
109
+ type: ColorJitter
110
+ weight: 1
111
+ saturation:
112
+ kwargs:
113
+ saturation:
114
+ - 0.5
115
+ - 1.5
116
+ type: ColorJitter
117
+ weight: 1
118
+ sharpness:
119
+ kwargs:
120
+ sharpness:
121
+ - 0.5
122
+ - 1.5
123
+ type: SharpnessJitter
124
+ weight: 1
125
+ repo_id: omy_pnp_language
126
+ revision: null
127
+ root: ./demo_data_language
128
+ use_imagenet_stats: true
129
+ video_backend: torchcodec
130
+ env:
131
+ value: null
132
+ eval:
133
+ value:
134
+ batch_size: 50
135
+ n_episodes: 50
136
+ use_async_envs: false
137
+ eval_freq:
138
+ value: -1
139
+ job_name:
140
+ value: smolvla_omy
141
+ log_freq:
142
+ value: 50
143
+ num_workers:
144
+ value: 8
145
+ optimizer:
146
+ value:
147
+ betas:
148
+ - 0.9
149
+ - 0.95
150
+ eps: 1e-08
151
+ grad_clip_norm: 10
152
+ lr: 0.0001
153
+ type: adamw
154
+ weight_decay: 1e-10
155
+ output_dir:
156
+ value: ckpt/smolvla_omy
157
+ policy:
158
+ value:
159
+ adapt_to_pi_aloha: false
160
+ add_image_special_tokens: false
161
+ attention_mode: cross_attn
162
+ chunk_size: 5
163
+ device: cuda
164
+ empty_cameras: 0
165
+ expert_width_multiplier: 0.75
166
+ freeze_vision_encoder: true
167
+ load_vlm_weights: false
168
+ max_action_dim: 32
169
+ max_period: 4
170
+ max_state_dim: 32
171
+ min_period: 0.004
172
+ n_action_steps: 5
173
+ n_obs_steps: 1
174
+ normalization_mapping:
175
+ ACTION: MEAN_STD
176
+ STATE: MEAN_STD
177
+ VISUAL: IDENTITY
178
+ num_expert_layers: -1
179
+ num_steps: 10
180
+ num_vlm_layers: 16
181
+ optimizer_betas:
182
+ - 0.9
183
+ - 0.95
184
+ optimizer_eps: 1e-08
185
+ optimizer_grad_clip_norm: 10
186
+ optimizer_lr: 0.0001
187
+ optimizer_weight_decay: 1e-10
188
+ pad_language_to: longest
189
+ prefix_length: -1
190
+ resize_imgs_with_padding:
191
+ - 512
192
+ - 512
193
+ scheduler_decay_lr: 2.5e-06
194
+ scheduler_decay_steps: 30000
195
+ scheduler_warmup_steps: 1000
196
+ self_attn_every_n_layers: 2
197
+ tokenizer_max_length: 48
198
+ train_expert_only: true
199
+ train_state_proj: true
200
+ type: smolvla
201
+ use_amp: false
202
+ use_cache: true
203
+ use_delta_joint_actions_aloha: false
204
+ vlm_model_name: HuggingFaceTB/SmolVLM2-500M-Video-Instruct
205
+ resume:
206
+ value: false
207
+ save_checkpoint:
208
+ value: true
209
+ save_freq:
210
+ value: 10000
211
+ scheduler:
212
+ value:
213
+ decay_lr: 2.5e-06
214
+ num_decay_steps: 30000
215
+ num_warmup_steps: 1000
216
+ peak_lr: 0.0001
217
+ type: cosine_decay_with_warmup
218
+ seed:
219
+ value: 42
220
+ steps:
221
+ value: 2000
222
+ use_policy_training_preset:
223
+ value: true
224
+ wandb:
225
+ value:
226
+ disable_artifact: true
227
+ enable: true
228
+ entity: ai-liber
229
+ mode: null
230
+ notes: null
231
+ project: smolvla_omy
232
+ run_id: null
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/output.log ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Logs will be synced with wandb.
2
+ INFO 2025-11-06 15:12:54 ndb_utils.py:96 Track this run --> https://wandb.ai/ai-liber/smolvla_omy/runs/p7vw7943
3
+ INFO 2025-11-06 15:12:54 in_model.py:127 Creating dataset
4
+ Resolving data files: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 20/20 [00:00<00:00, 28397.45it/s]
5
+ INFO 2025-11-06 15:12:56 in_model.py:138 Creating policy
6
+ Reducing the number of VLM layers to 16 ...
7
+ model.safetensors: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 907M/907M [01:19<00:00, 11.5MB/s]
8
+ INFO 2025-11-06 15:15:28 in_model.py:148 Creating optimizer and scheduler
9
+ INFO 2025-11-06 15:15:28 in_model.py:160 Output dir: ckpt/smolvla_omy
10
+ INFO 2025-11-06 15:15:28 in_model.py:163 cfg.steps=2000 (2K)
11
+ INFO 2025-11-06 15:15:28 in_model.py:164 dataset.num_frames=2621 (3K)
12
+ INFO 2025-11-06 15:15:28 in_model.py:165 dataset.num_episodes=20
13
+ INFO 2025-11-06 15:15:28 in_model.py:166 num_learnable_params=99880992 (100M)
14
+ INFO 2025-11-06 15:15:28 in_model.py:167 num_total_params=450046216 (450M)
15
+ INFO 2025-11-06 15:15:28 in_model.py:206 Start offline training on a fixed dataset
16
+ INFO 2025-11-06 15:16:19 in_model.py:236 step:50 smpl:800 ep:6 epch:0.31 loss:0.306 grdn:2.807 lr:2.6e-06 updt_s:1.004 data_s:0.025
17
+ WARNING 2025-11-06 15:16:19 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
18
+ WARNING 2025-11-06 15:16:19 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
19
+ INFO 2025-11-06 15:17:09 in_model.py:236 step:100 smpl:2K ep:12 epch:0.61 loss:0.223 grdn:1.959 lr:7.6e-06 updt_s:0.986 data_s:0.000
20
+ WARNING 2025-11-06 15:17:09 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
21
+ WARNING 2025-11-06 15:17:09 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
22
+ INFO 2025-11-06 15:17:59 in_model.py:236 step:150 smpl:2K ep:18 epch:0.92 loss:0.166 grdn:1.444 lr:1.3e-05 updt_s:1.000 data_s:0.000
23
+ WARNING 2025-11-06 15:17:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
24
+ WARNING 2025-11-06 15:17:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
25
+ INFO 2025-11-06 15:18:53 in_model.py:236 step:200 smpl:3K ep:24 epch:1.22 loss:0.112 grdn:1.042 lr:1.8e-05 updt_s:1.040 data_s:0.030
26
+ WARNING 2025-11-06 15:18:53 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
27
+ WARNING 2025-11-06 15:18:53 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
28
+ INFO 2025-11-06 15:19:45 in_model.py:236 step:250 smpl:4K ep:31 epch:1.53 loss:0.091 grdn:0.947 lr:2.3e-05 updt_s:1.050 data_s:0.000
29
+ WARNING 2025-11-06 15:19:45 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
30
+ WARNING 2025-11-06 15:19:45 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
31
+ INFO 2025-11-06 15:20:38 in_model.py:236 step:300 smpl:5K ep:37 epch:1.83 loss:0.079 grdn:0.972 lr:2.8e-05 updt_s:1.053 data_s:0.000
32
+ WARNING 2025-11-06 15:20:38 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
33
+ WARNING 2025-11-06 15:20:38 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
34
+ INFO 2025-11-06 15:21:32 in_model.py:236 step:350 smpl:6K ep:43 epch:2.14 loss:0.068 grdn:0.944 lr:3.3e-05 updt_s:1.045 data_s:0.022
35
+ WARNING 2025-11-06 15:21:32 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
36
+ WARNING 2025-11-06 15:21:32 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
37
+ INFO 2025-11-06 15:22:25 in_model.py:236 step:400 smpl:6K ep:49 epch:2.44 loss:0.062 grdn:0.921 lr:3.8e-05 updt_s:1.052 data_s:0.000
38
+ WARNING 2025-11-06 15:22:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
39
+ WARNING 2025-11-06 15:22:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
40
+ INFO 2025-11-06 15:23:17 in_model.py:236 step:450 smpl:7K ep:55 epch:2.75 loss:0.062 grdn:0.979 lr:4.3e-05 updt_s:1.051 data_s:0.000
41
+ WARNING 2025-11-06 15:23:17 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
42
+ WARNING 2025-11-06 15:23:17 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
43
+ INFO 2025-11-06 15:24:11 in_model.py:236 step:500 smpl:8K ep:61 epch:3.05 loss:0.055 grdn:0.987 lr:4.8e-05 updt_s:1.044 data_s:0.022
44
+ WARNING 2025-11-06 15:24:11 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
45
+ WARNING 2025-11-06 15:24:11 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
46
+ INFO 2025-11-06 15:25:03 in_model.py:236 step:550 smpl:9K ep:67 epch:3.36 loss:0.046 grdn:0.843 lr:5.3e-05 updt_s:1.049 data_s:0.000
47
+ WARNING 2025-11-06 15:25:03 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
48
+ WARNING 2025-11-06 15:25:03 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
49
+ INFO 2025-11-06 15:25:56 in_model.py:236 step:600 smpl:10K ep:73 epch:3.66 loss:0.050 grdn:0.955 lr:5.8e-05 updt_s:1.056 data_s:0.000
50
+ WARNING 2025-11-06 15:25:56 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
51
+ WARNING 2025-11-06 15:25:56 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
52
+ INFO 2025-11-06 15:26:49 in_model.py:236 step:650 smpl:10K ep:79 epch:3.97 loss:0.050 grdn:1.030 lr:6.3e-05 updt_s:1.056 data_s:0.000
53
+ WARNING 2025-11-06 15:26:49 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
54
+ WARNING 2025-11-06 15:26:49 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
55
+ INFO 2025-11-06 15:27:43 in_model.py:236 step:700 smpl:11K ep:85 epch:4.27 loss:0.044 grdn:0.878 lr:6.8e-05 updt_s:1.047 data_s:0.020
56
+ WARNING 2025-11-06 15:27:43 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
57
+ WARNING 2025-11-06 15:27:43 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
58
+ INFO 2025-11-06 15:28:36 in_model.py:236 step:750 smpl:12K ep:92 epch:4.58 loss:0.048 grdn:0.911 lr:7.3e-05 updt_s:1.055 data_s:0.000
59
+ WARNING 2025-11-06 15:28:36 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
60
+ WARNING 2025-11-06 15:28:36 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
61
+ INFO 2025-11-06 15:29:29 in_model.py:236 step:800 smpl:13K ep:98 epch:4.88 loss:0.050 grdn:0.946 lr:7.8e-05 updt_s:1.055 data_s:0.000
62
+ WARNING 2025-11-06 15:29:29 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
63
+ WARNING 2025-11-06 15:29:29 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
64
+ INFO 2025-11-06 15:30:23 in_model.py:236 step:850 smpl:14K ep:104 epch:5.19 loss:0.045 grdn:0.904 lr:8.3e-05 updt_s:1.047 data_s:0.021
65
+ WARNING 2025-11-06 15:30:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
66
+ WARNING 2025-11-06 15:30:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
67
+ INFO 2025-11-06 15:31:15 in_model.py:236 step:900 smpl:14K ep:110 epch:5.49 loss:0.056 grdn:1.043 lr:8.8e-05 updt_s:1.054 data_s:0.000
68
+ WARNING 2025-11-06 15:31:15 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
69
+ WARNING 2025-11-06 15:31:15 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
70
+ INFO 2025-11-06 15:32:09 in_model.py:236 step:950 smpl:15K ep:116 epch:5.80 loss:0.056 grdn:1.051 lr:9.3e-05 updt_s:1.056 data_s:0.000
71
+ WARNING 2025-11-06 15:32:09 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
72
+ WARNING 2025-11-06 15:32:09 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
73
+ INFO 2025-11-06 15:33:02 in_model.py:236 step:1K smpl:16K ep:122 epch:6.10 loss:0.051 grdn:0.937 lr:9.8e-05 updt_s:1.047 data_s:0.021
74
+ WARNING 2025-11-06 15:33:02 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
75
+ WARNING 2025-11-06 15:33:02 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
76
+ INFO 2025-11-06 15:33:55 in_model.py:236 step:1K smpl:17K ep:128 epch:6.41 loss:0.047 grdn:0.883 lr:1.0e-04 updt_s:1.053 data_s:0.000
77
+ WARNING 2025-11-06 15:33:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
78
+ WARNING 2025-11-06 15:33:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
79
+ INFO 2025-11-06 15:34:48 in_model.py:236 step:1K smpl:18K ep:134 epch:6.71 loss:0.056 grdn:0.951 lr:1.0e-04 updt_s:1.057 data_s:0.000
80
+ WARNING 2025-11-06 15:34:48 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
81
+ WARNING 2025-11-06 15:34:48 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
82
+ INFO 2025-11-06 15:35:42 in_model.py:236 step:1K smpl:18K ep:140 epch:7.02 loss:0.055 grdn:0.955 lr:1.0e-04 updt_s:1.051 data_s:0.020
83
+ WARNING 2025-11-06 15:35:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
84
+ WARNING 2025-11-06 15:35:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
85
+ INFO 2025-11-06 15:36:34 in_model.py:236 step:1K smpl:19K ep:147 epch:7.33 loss:0.049 grdn:0.953 lr:1.0e-04 updt_s:1.050 data_s:0.000
86
+ WARNING 2025-11-06 15:36:34 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
87
+ WARNING 2025-11-06 15:36:34 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
88
+ INFO 2025-11-06 15:37:27 in_model.py:236 step:1K smpl:20K ep:153 epch:7.63 loss:0.045 grdn:0.868 lr:1.0e-04 updt_s:1.052 data_s:0.000
89
+ WARNING 2025-11-06 15:37:27 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
90
+ WARNING 2025-11-06 15:37:27 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
91
+ INFO 2025-11-06 15:38:20 in_model.py:236 step:1K smpl:21K ep:159 epch:7.94 loss:0.053 grdn:0.922 lr:1.0e-04 updt_s:1.058 data_s:0.000
92
+ WARNING 2025-11-06 15:38:20 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
93
+ WARNING 2025-11-06 15:38:20 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
94
+ INFO 2025-11-06 15:39:14 in_model.py:236 step:1K smpl:22K ep:165 epch:8.24 loss:0.054 grdn:0.953 lr:1.0e-04 updt_s:1.048 data_s:0.021
95
+ WARNING 2025-11-06 15:39:14 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
96
+ WARNING 2025-11-06 15:39:14 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
97
+ INFO 2025-11-06 15:40:07 in_model.py:236 step:1K smpl:22K ep:171 epch:8.55 loss:0.044 grdn:0.778 lr:9.9e-05 updt_s:1.059 data_s:0.000
98
+ WARNING 2025-11-06 15:40:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
99
+ WARNING 2025-11-06 15:40:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
100
+ INFO 2025-11-06 15:41:00 in_model.py:236 step:1K smpl:23K ep:177 epch:8.85 loss:0.044 grdn:0.855 lr:9.9e-05 updt_s:1.054 data_s:0.000
101
+ WARNING 2025-11-06 15:41:00 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
102
+ WARNING 2025-11-06 15:41:00 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
103
+ INFO 2025-11-06 15:41:54 in_model.py:236 step:2K smpl:24K ep:183 epch:9.16 loss:0.045 grdn:0.835 lr:9.9e-05 updt_s:1.055 data_s:0.021
104
+ WARNING 2025-11-06 15:41:54 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
105
+ WARNING 2025-11-06 15:41:54 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
106
+ INFO 2025-11-06 15:42:47 in_model.py:236 step:2K smpl:25K ep:189 epch:9.46 loss:0.043 grdn:0.803 lr:9.9e-05 updt_s:1.063 data_s:0.000
107
+ WARNING 2025-11-06 15:42:47 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
108
+ WARNING 2025-11-06 15:42:47 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
109
+ INFO 2025-11-06 15:43:41 in_model.py:236 step:2K smpl:26K ep:195 epch:9.77 loss:0.047 grdn:0.870 lr:9.9e-05 updt_s:1.065 data_s:0.000
110
+ WARNING 2025-11-06 15:43:41 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
111
+ WARNING 2025-11-06 15:43:41 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
112
+ INFO 2025-11-06 15:44:35 in_model.py:236 step:2K smpl:26K ep:201 epch:10.07 loss:0.042 grdn:0.784 lr:9.9e-05 updt_s:1.052 data_s:0.022
113
+ WARNING 2025-11-06 15:44:35 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
114
+ WARNING 2025-11-06 15:44:35 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
115
+ INFO 2025-11-06 15:45:27 in_model.py:236 step:2K smpl:27K ep:208 epch:10.38 loss:0.043 grdn:0.775 lr:9.9e-05 updt_s:1.050 data_s:0.000
116
+ WARNING 2025-11-06 15:45:27 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
117
+ WARNING 2025-11-06 15:45:27 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
118
+ INFO 2025-11-06 15:46:20 in_model.py:236 step:2K smpl:28K ep:214 epch:10.68 loss:0.042 grdn:0.799 lr:9.9e-05 updt_s:1.056 data_s:0.000
119
+ WARNING 2025-11-06 15:46:20 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
120
+ WARNING 2025-11-06 15:46:20 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
121
+ INFO 2025-11-06 15:47:13 in_model.py:236 step:2K smpl:29K ep:220 epch:10.99 loss:0.038 grdn:0.714 lr:9.9e-05 updt_s:1.053 data_s:0.000
122
+ WARNING 2025-11-06 15:47:13 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
123
+ WARNING 2025-11-06 15:47:13 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
124
+ INFO 2025-11-06 15:48:07 in_model.py:236 step:2K smpl:30K ep:226 epch:11.29 loss:0.040 grdn:0.753 lr:9.9e-05 updt_s:1.047 data_s:0.021
125
+ WARNING 2025-11-06 15:48:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
126
+ WARNING 2025-11-06 15:48:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
127
+ INFO 2025-11-06 15:49:00 in_model.py:236 step:2K smpl:30K ep:232 epch:11.60 loss:0.038 grdn:0.699 lr:9.9e-05 updt_s:1.059 data_s:0.000
128
+ WARNING 2025-11-06 15:49:00 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
129
+ WARNING 2025-11-06 15:49:00 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
130
+ INFO 2025-11-06 15:49:53 in_model.py:236 step:2K smpl:31K ep:238 epch:11.90 loss:0.038 grdn:0.660 lr:9.9e-05 updt_s:1.056 data_s:0.000
131
+ WARNING 2025-11-06 15:49:53 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
132
+ WARNING 2025-11-06 15:49:53 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
133
+ INFO 2025-11-06 15:50:47 in_model.py:236 step:2K smpl:32K ep:244 epch:12.21 loss:0.034 grdn:0.633 lr:9.9e-05 updt_s:1.048 data_s:0.021
134
+ WARNING 2025-11-06 15:50:47 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
135
+ WARNING 2025-11-06 15:50:47 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
136
+ INFO 2025-11-06 15:50:47 in_model.py:245 Checkpoint policy after step 2000
137
+ INFO 2025-11-06 15:50:50 in_model.py:287 End of training
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/requirements.txt ADDED
@@ -0,0 +1,172 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ pyparsing==3.2.5
2
+ nvidia-cuda-cupti-cu12==12.4.127
3
+ pip==25.2
4
+ setuptools==80.9.0
5
+ gitdb==4.0.12
6
+ networkx==3.4.2
7
+ propcache==0.4.1
8
+ aiohappyeyeballs==2.6.1
9
+ aiosignal==1.4.0
10
+ Werkzeug==3.1.3
11
+ nvidia-nvtx-cu12==12.4.127
12
+ typing-inspect==0.9.0
13
+ urllib3==2.5.0
14
+ imageio-ffmpeg==0.6.0
15
+ num2words==0.5.14
16
+ PySocks==1.7.1
17
+ datasets==3.4.1
18
+ PyScreeze==1.0.1
19
+ tokenizers==0.21.4
20
+ yarl==1.22.0
21
+ MouseInfo==0.1.3
22
+ mypy_extensions==1.1.0
23
+ pyzmq==27.1.0
24
+ beautifulsoup4==4.14.2
25
+ importlib_metadata==8.7.0
26
+ evdev==1.9.2
27
+ sentry-sdk==2.43.0
28
+ wcwidth==0.2.14
29
+ typing_extensions==4.15.0
30
+ certifi==2025.10.5
31
+ importlib_resources==6.5.2
32
+ hf-xet==1.2.0
33
+ frozenlist==1.8.0
34
+ annotated-types==0.7.0
35
+ PyGetWindow==0.0.9
36
+ draccus==0.10.0
37
+ rerun-sdk==0.26.2
38
+ cmake==4.1.2
39
+ PyAutoGUI==0.9.54
40
+ gdown==5.2.0
41
+ pyarrow==22.0.0
42
+ nvidia-cusparse-cu12==12.3.1.170
43
+ torch==2.6.0
44
+ cycler==0.12.1
45
+ av==16.0.1
46
+ pytz==2025.2
47
+ tzdata==2025.2
48
+ zipp==3.23.0
49
+ prompt_toolkit==3.0.52
50
+ itsdangerous==2.2.0
51
+ lerobot==0.1.0
52
+ smmap==5.0.2
53
+ pyserial==3.5
54
+ charset-normalizer==3.4.4
55
+ numpy==2.2.6
56
+ psutil==7.1.3
57
+ zarr==2.18.3
58
+ matplotlib==3.10.7
59
+ llvmlite==0.45.1
60
+ torchaudio==2.6.0
61
+ huggingface-hub==0.36.0
62
+ MarkupSafe==3.0.3
63
+ packaging==25.0
64
+ PyYAML==6.0.3
65
+ Flask==3.1.2
66
+ hf_transfer==0.1.9
67
+ pydantic_core==2.41.4
68
+ pandas==2.3.3
69
+ mergedeep==1.3.4
70
+ h5py==3.15.1
71
+ transformers==4.50.3
72
+ torchcodec==0.8.1
73
+ termcolor==3.2.0
74
+ toml==0.10.2
75
+ imageio==2.37.0
76
+ nvidia-cuda-nvrtc-cu12==12.4.127
77
+ asciitree==0.3.3
78
+ antlr4-python3-runtime==4.9.3
79
+ jsonlines==4.0.0
80
+ scipy==1.15.3
81
+ pyperclip==1.11.0
82
+ pillow==12.0.0
83
+ click==8.3.0
84
+ torchvision==0.21.0
85
+ cffi==2.0.0
86
+ etils==1.13.0
87
+ Farama-Notifications==0.0.4
88
+ nvidia-cufft-cu12==11.2.1.3
89
+ kiwisolver==1.4.9
90
+ triton==3.2.0
91
+ fonttools==4.60.1
92
+ pycparser==2.23
93
+ absl-py==2.3.1
94
+ requests==2.32.5
95
+ six==1.17.0
96
+ nvidia-nccl-cu12==2.21.5
97
+ pytweening==1.2.0
98
+ docopt==0.6.2
99
+ contourpy==1.3.2
100
+ nvidia-cusolver-cu12==11.6.1.9
101
+ pynput==1.8.1
102
+ Jinja2==3.1.6
103
+ einops==0.8.1
104
+ numcodecs==0.13.1
105
+ async-timeout==5.0.1
106
+ blinker==1.9.0
107
+ pydantic==2.12.3
108
+ xxhash==3.6.0
109
+ cloudpickle==3.1.2
110
+ diffusers==0.35.2
111
+ PyOpenGL==3.1.10
112
+ glfw==2.10.0
113
+ soupsieve==2.8
114
+ GitPython==3.1.45
115
+ fsspec==2024.12.0
116
+ aiohttp==3.13.2
117
+ opencv-python-headless==4.12.0.88
118
+ gymnasium==0.29.1
119
+ python3-xlib==0.15
120
+ mpmath==1.3.0
121
+ nvidia-curand-cu12==10.3.5.147
122
+ wandb==0.22.3
123
+ nvidia-cublas-cu12==12.4.5.8
124
+ nvidia-cuda-runtime-cu12==12.4.127
125
+ python-xlib==0.33
126
+ multiprocess==0.70.16
127
+ pfzy==0.3.4
128
+ orderly-set==5.5.0
129
+ nvidia-nvjitlink-cu12==12.4.127
130
+ accelerate==1.11.0
131
+ regex==2025.11.3
132
+ omegaconf==2.3.0
133
+ safetensors==0.5.3
134
+ filelock==3.20.0
135
+ numba==0.62.1
136
+ nvidia-cusparselt-cu12==0.6.2
137
+ dill==0.3.8
138
+ pyyaml-include==1.4.1
139
+ platformdirs==4.5.0
140
+ multidict==6.7.0
141
+ protobuf==6.33.0
142
+ attrs==25.4.0
143
+ inquirerpy==0.3.4
144
+ PyMsgBox==2.0.1
145
+ pymunk==6.11.1
146
+ nvidia-cudnn-cu12==9.1.0.70
147
+ typing-inspection==0.4.2
148
+ fasteners==0.20
149
+ PyRect==0.2.0
150
+ idna==3.11
151
+ deepdiff==8.6.1
152
+ wheel==0.45.1
153
+ sympy==1.13.1
154
+ python-dateutil==2.9.0.post0
155
+ mujoco==3.1.6
156
+ tqdm==4.67.1
157
+ backports.tarfile==1.2.0
158
+ typeguard==4.3.0
159
+ zipp==3.19.2
160
+ inflect==7.3.1
161
+ more-itertools==10.3.0
162
+ tomli==2.0.1
163
+ jaraco.collections==5.1.0
164
+ jaraco.functools==4.0.1
165
+ typing_extensions==4.12.2
166
+ importlib_metadata==8.0.0
167
+ jaraco.text==3.12.1
168
+ platformdirs==4.2.2
169
+ jaraco.context==5.3.0
170
+ wheel==0.45.1
171
+ packaging==24.2
172
+ autocommand==2.2.2
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/wandb-metadata.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-51-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.10.19",
4
+ "startedAt": "2025-11-06T07:12:52.855818Z",
5
+ "args": [
6
+ "--config_path",
7
+ "smolvla_omy.yaml"
8
+ ],
9
+ "program": "/workspace/liber/embodied_ai/lerobot-mujoco-tutorial/train_model.py",
10
+ "codePath": "train_model.py",
11
+ "codePathLocal": "train_model.py",
12
+ "git": {
13
+ "remote": "git@github.com:liberow/lerobot-mujoco-tutorial.git",
14
+ "commit": "b11377a0ce7a3c32cefe260c6dce959b0db2bdda"
15
+ },
16
+ "root": "ckpt/smolvla_omy",
17
+ "host": "c81ab3b21da2",
18
+ "executable": "/opt/conda/envs/lerobot_mujoco_tutorial/bin/python",
19
+ "cpu_count": 40,
20
+ "cpu_count_logical": 80,
21
+ "gpu": "Tesla V100-SXM2-32GB",
22
+ "gpu_count": 4,
23
+ "disk": {
24
+ "/": {
25
+ "total": "1676411871232",
26
+ "used": "1480236986368"
27
+ }
28
+ },
29
+ "memory": {
30
+ "total": "134767624192"
31
+ },
32
+ "gpu_nvidia": [
33
+ {
34
+ "name": "Tesla V100-SXM2-32GB",
35
+ "memoryTotal": "34359738368",
36
+ "cudaCores": 5120,
37
+ "architecture": "Volta",
38
+ "uuid": "GPU-b29b46dc-42a3-9fcc-be17-50b8b8972960"
39
+ },
40
+ {
41
+ "name": "Tesla V100-SXM2-32GB",
42
+ "memoryTotal": "34359738368",
43
+ "cudaCores": 5120,
44
+ "architecture": "Volta",
45
+ "uuid": "GPU-368aa17c-5504-4d63-4df2-339d38a17fe3"
46
+ },
47
+ {
48
+ "name": "Tesla V100-SXM2-32GB",
49
+ "memoryTotal": "34359738368",
50
+ "cudaCores": 5120,
51
+ "architecture": "Volta",
52
+ "uuid": "GPU-acf7b3cc-f154-087a-3a06-d57e023c09be"
53
+ },
54
+ {
55
+ "name": "Tesla V100-SXM2-32GB",
56
+ "memoryTotal": "34359738368",
57
+ "cudaCores": 5120,
58
+ "architecture": "Volta",
59
+ "uuid": "GPU-86b115b3-0485-dd33-56f3-385a1140ae36"
60
+ }
61
+ ],
62
+ "cudaVersion": "12.4",
63
+ "writerId": "s9gajpxnrqtnt5al3g989j4cneoydvqt"
64
+ }
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_timestamp":1.7624154471798701e+09,"train/epochs":12.209080503624572,"_wandb":{"runtime":2276},"train/lr":9.896048857046275e-05,"train/samples":32000,"_runtime":2276.402947135,"train/update_s":1.0481160825677216,"_step":2000,"train/steps":2000,"train/grad_norm":0.6333963936567306,"train/episodes":244.1816100724914,"train/loss":0.025371486321091652,"train/dataloading_s":0.02116153893992305}
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/logs/debug-core.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-06T15:12:52.895741661+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp0l06am8l/port-343428.txt","pid":343428,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-11-06T15:12:52.897386616+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":343428}
3
+ {"time":"2025-11-06T15:12:52.897390339+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-343428-344296-4240599004/socket","Net":"unix"}}
4
+ {"time":"2025-11-06T15:12:53.068133217+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-11-06T15:12:53.084869085+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"p7vw7943","id":"1(@)"}
6
+ {"time":"2025-11-06T15:12:53.652476859+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"p7vw7943","id":"1(@)"}
7
+ {"time":"2025-11-06T15:50:50.741198323+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
8
+ {"time":"2025-11-06T15:50:50.741293234+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"}
9
+ {"time":"2025-11-06T15:50:50.741322423+08:00","level":"INFO","msg":"server is shutting down"}
10
+ {"time":"2025-11-06T15:50:50.741481914+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
11
+ {"time":"2025-11-06T15:50:50.741595273+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-343428-344296-4240599004/socket","Net":"unix"}}
12
+ {"time":"2025-11-06T15:50:52.796486185+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
13
+ {"time":"2025-11-06T15:50:52.796535792+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
14
+ {"time":"2025-11-06T15:50:52.796557312+08:00","level":"INFO","msg":"server is closed"}
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/logs/debug-internal.log ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-06T15:12:53.085119529+08:00","level":"INFO","msg":"stream: starting","core version":"0.22.3"}
2
+ {"time":"2025-11-06T15:12:53.65199406+08:00","level":"INFO","msg":"stream: created new stream","id":"p7vw7943"}
3
+ {"time":"2025-11-06T15:12:53.652165503+08:00","level":"INFO","msg":"handler: started","stream_id":"p7vw7943"}
4
+ {"time":"2025-11-06T15:12:53.652453415+08:00","level":"INFO","msg":"stream: started","id":"p7vw7943"}
5
+ {"time":"2025-11-06T15:12:53.652494895+08:00","level":"INFO","msg":"writer: started","stream_id":"p7vw7943"}
6
+ {"time":"2025-11-06T15:12:53.652495951+08:00","level":"INFO","msg":"sender: started","stream_id":"p7vw7943"}
7
+ {"time":"2025-11-06T15:50:50.741281958+08:00","level":"INFO","msg":"stream: closing","id":"p7vw7943"}
8
+ {"time":"2025-11-06T15:50:52.335321515+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
9
+ {"time":"2025-11-06T15:50:52.795892747+08:00","level":"INFO","msg":"handler: closed","stream_id":"p7vw7943"}
10
+ {"time":"2025-11-06T15:50:52.796103474+08:00","level":"INFO","msg":"sender: closed","stream_id":"p7vw7943"}
11
+ {"time":"2025-11-06T15:50:52.796134886+08:00","level":"INFO","msg":"stream: closed","id":"p7vw7943"}
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/logs/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-06 15:12:52,857 INFO MainThread:343428 [wandb_setup.py:_flush():81] Current SDK version is 0.22.3
2
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Configure stats pid to 343428
3
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Loading settings from /root/.config/wandb/settings
4
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Loading settings from /workspace/liber/embodied_ai/lerobot-mujoco-tutorial/wandb/settings
5
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:setup_run_log_directory():706] Logging user logs to ckpt/smolvla_omy/wandb/run-20251106_151252-p7vw7943/logs/debug.log
7
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:setup_run_log_directory():707] Logging internal logs to ckpt/smolvla_omy/wandb/run-20251106_151252-p7vw7943/logs/debug-internal.log
8
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:init():833] calling init triggers
9
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:init():838] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'omy_pnp_language', 'root': './demo_data_language', 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec'}, 'env': None, 'policy': {'type': 'smolvla', 'n_obs_steps': 1, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'chunk_size': 5, 'n_action_steps': 5, 'max_state_dim': 32, 'max_action_dim': 32, 'resize_imgs_with_padding': [512, 512], 'empty_cameras': 0, 'adapt_to_pi_aloha': False, 'use_delta_joint_actions_aloha': False, 'tokenizer_max_length': 48, 'num_steps': 10, 'use_cache': True, 'freeze_vision_encoder': True, 'train_expert_only': True, 'train_state_proj': True, 'optimizer_lr': 0.0001, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 1e-10, 'optimizer_grad_clip_norm': 10, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'vlm_model_name': 'HuggingFaceTB/SmolVLM2-500M-Video-Instruct', 'load_vlm_weights': False, 'add_image_special_tokens': False, 'attention_mode': 'cross_attn', 'prefix_length': -1, 'pad_language_to': 'longest', 'num_expert_layers': -1, 'num_vlm_layers': 16, 'self_attn_every_n_layers': 2, 'expert_width_multiplier': 0.75, 'min_period': 0.004, 'max_period': 4.0}, 'output_dir': 'ckpt/smolvla_omy', 'job_name': 'smolvla_omy', 'resume': False, 'seed': 42, 'num_workers': 8, 'batch_size': 16, 'steps': 2000, 'eval_freq': -1, 'log_freq': 50, 'save_checkpoint': True, 'save_freq': 10000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 0.0001, 'weight_decay': 1e-10, 'grad_clip_norm': 10, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 0.0001, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': True, 'project': 'smolvla_omy', 'entity': 'ai-liber', 'notes': None, 'run_id': None, 'mode': None}, '_wandb': {}}
11
+ 2025-11-06 15:12:52,858 INFO MainThread:343428 [wandb_init.py:init():881] starting backend
12
+ 2025-11-06 15:12:53,069 INFO MainThread:343428 [wandb_init.py:init():884] sending inform_init request
13
+ 2025-11-06 15:12:53,079 INFO MainThread:343428 [wandb_init.py:init():892] backend started and connected
14
+ 2025-11-06 15:12:53,082 INFO MainThread:343428 [wandb_init.py:init():962] updated telemetry
15
+ 2025-11-06 15:12:53,094 INFO MainThread:343428 [wandb_init.py:init():986] communicating run to backend with 90.0 second timeout
16
+ 2025-11-06 15:12:54,336 INFO MainThread:343428 [wandb_init.py:init():1033] starting run threads in backend
17
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-11-06 15:12:54,684 INFO MainThread:343428 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-11-06 15:12:54,689 INFO MainThread:343428 [wandb_init.py:init():1073] run started, returning control to user process
22
+ 2025-11-06 15:50:50,741 INFO wandb-AsyncioManager-main:343428 [service_client.py:_forward_responses():80] Reached EOF.
23
+ 2025-11-06 15:50:50,742 INFO wandb-AsyncioManager-main:343428 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles.
smolvla_omy_2000_steps/wandb/run-20251106_151252-p7vw7943/run-p7vw7943.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03f8032f5860ee1d4d5c37feb52286d1469d23d14b6bb370b766d431d10a92f6
3
+ size 418963