Liberow commited on
Commit
15f7011
·
verified ·
1 Parent(s): b82d3fe

Upload folder using huggingface_hub

Browse files
smolvla_omy/checkpoints/020000/pretrained_model/config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "normalization_mapping": {
5
+ "VISUAL": "IDENTITY",
6
+ "STATE": "MEAN_STD",
7
+ "ACTION": "MEAN_STD"
8
+ },
9
+ "input_features": {
10
+ "observation.image": {
11
+ "type": "VISUAL",
12
+ "shape": [
13
+ 3,
14
+ 256,
15
+ 256
16
+ ]
17
+ },
18
+ "observation.wrist_image": {
19
+ "type": "VISUAL",
20
+ "shape": [
21
+ 3,
22
+ 256,
23
+ 256
24
+ ]
25
+ },
26
+ "observation.state": {
27
+ "type": "STATE",
28
+ "shape": [
29
+ 6
30
+ ]
31
+ }
32
+ },
33
+ "output_features": {
34
+ "action": {
35
+ "type": "ACTION",
36
+ "shape": [
37
+ 7
38
+ ]
39
+ }
40
+ },
41
+ "device": "cuda",
42
+ "use_amp": false,
43
+ "chunk_size": 5,
44
+ "n_action_steps": 5,
45
+ "max_state_dim": 32,
46
+ "max_action_dim": 32,
47
+ "resize_imgs_with_padding": [
48
+ 512,
49
+ 512
50
+ ],
51
+ "empty_cameras": 0,
52
+ "adapt_to_pi_aloha": false,
53
+ "use_delta_joint_actions_aloha": false,
54
+ "tokenizer_max_length": 48,
55
+ "num_steps": 10,
56
+ "use_cache": true,
57
+ "freeze_vision_encoder": true,
58
+ "train_expert_only": true,
59
+ "train_state_proj": true,
60
+ "optimizer_lr": 0.0001,
61
+ "optimizer_betas": [
62
+ 0.9,
63
+ 0.95
64
+ ],
65
+ "optimizer_eps": 1e-08,
66
+ "optimizer_weight_decay": 1e-10,
67
+ "optimizer_grad_clip_norm": 10,
68
+ "scheduler_warmup_steps": 1000,
69
+ "scheduler_decay_steps": 30000,
70
+ "scheduler_decay_lr": 2.5e-06,
71
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
72
+ "load_vlm_weights": false,
73
+ "add_image_special_tokens": false,
74
+ "attention_mode": "cross_attn",
75
+ "prefix_length": -1,
76
+ "pad_language_to": "longest",
77
+ "num_expert_layers": -1,
78
+ "num_vlm_layers": 16,
79
+ "self_attn_every_n_layers": 2,
80
+ "expert_width_multiplier": 0.75,
81
+ "min_period": 0.004,
82
+ "max_period": 4.0
83
+ }
smolvla_omy/checkpoints/020000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c4d65699809dec9913a8a8d4fadfbeac34d66644ea961d1b390f5734e4407354
3
+ size 1197790032
smolvla_omy/checkpoints/020000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,195 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "omy_pnp_language",
4
+ "root": "./demo_data_language",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "revision": null,
64
+ "use_imagenet_stats": true,
65
+ "video_backend": "torchcodec"
66
+ },
67
+ "env": null,
68
+ "policy": {
69
+ "type": "smolvla",
70
+ "n_obs_steps": 1,
71
+ "normalization_mapping": {
72
+ "VISUAL": "IDENTITY",
73
+ "STATE": "MEAN_STD",
74
+ "ACTION": "MEAN_STD"
75
+ },
76
+ "input_features": {
77
+ "observation.image": {
78
+ "type": "VISUAL",
79
+ "shape": [
80
+ 3,
81
+ 256,
82
+ 256
83
+ ]
84
+ },
85
+ "observation.wrist_image": {
86
+ "type": "VISUAL",
87
+ "shape": [
88
+ 3,
89
+ 256,
90
+ 256
91
+ ]
92
+ },
93
+ "observation.state": {
94
+ "type": "STATE",
95
+ "shape": [
96
+ 6
97
+ ]
98
+ }
99
+ },
100
+ "output_features": {
101
+ "action": {
102
+ "type": "ACTION",
103
+ "shape": [
104
+ 7
105
+ ]
106
+ }
107
+ },
108
+ "device": "cuda",
109
+ "use_amp": false,
110
+ "chunk_size": 5,
111
+ "n_action_steps": 5,
112
+ "max_state_dim": 32,
113
+ "max_action_dim": 32,
114
+ "resize_imgs_with_padding": [
115
+ 512,
116
+ 512
117
+ ],
118
+ "empty_cameras": 0,
119
+ "adapt_to_pi_aloha": false,
120
+ "use_delta_joint_actions_aloha": false,
121
+ "tokenizer_max_length": 48,
122
+ "num_steps": 10,
123
+ "use_cache": true,
124
+ "freeze_vision_encoder": true,
125
+ "train_expert_only": true,
126
+ "train_state_proj": true,
127
+ "optimizer_lr": 0.0001,
128
+ "optimizer_betas": [
129
+ 0.9,
130
+ 0.95
131
+ ],
132
+ "optimizer_eps": 1e-08,
133
+ "optimizer_weight_decay": 1e-10,
134
+ "optimizer_grad_clip_norm": 10,
135
+ "scheduler_warmup_steps": 1000,
136
+ "scheduler_decay_steps": 30000,
137
+ "scheduler_decay_lr": 2.5e-06,
138
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
139
+ "load_vlm_weights": false,
140
+ "add_image_special_tokens": false,
141
+ "attention_mode": "cross_attn",
142
+ "prefix_length": -1,
143
+ "pad_language_to": "longest",
144
+ "num_expert_layers": -1,
145
+ "num_vlm_layers": 16,
146
+ "self_attn_every_n_layers": 2,
147
+ "expert_width_multiplier": 0.75,
148
+ "min_period": 0.004,
149
+ "max_period": 4.0
150
+ },
151
+ "output_dir": "ckpt/smolvla_omy",
152
+ "job_name": "smolvla_omy",
153
+ "resume": false,
154
+ "seed": 42,
155
+ "num_workers": 8,
156
+ "batch_size": 16,
157
+ "steps": 20000,
158
+ "eval_freq": -1,
159
+ "log_freq": 50,
160
+ "save_checkpoint": true,
161
+ "save_freq": 5000,
162
+ "use_policy_training_preset": true,
163
+ "optimizer": {
164
+ "type": "adamw",
165
+ "lr": 0.0001,
166
+ "weight_decay": 1e-10,
167
+ "grad_clip_norm": 10,
168
+ "betas": [
169
+ 0.9,
170
+ 0.95
171
+ ],
172
+ "eps": 1e-08
173
+ },
174
+ "scheduler": {
175
+ "type": "cosine_decay_with_warmup",
176
+ "num_warmup_steps": 1000,
177
+ "num_decay_steps": 30000,
178
+ "peak_lr": 0.0001,
179
+ "decay_lr": 2.5e-06
180
+ },
181
+ "eval": {
182
+ "n_episodes": 50,
183
+ "batch_size": 50,
184
+ "use_async_envs": false
185
+ },
186
+ "wandb": {
187
+ "enable": true,
188
+ "disable_artifact": true,
189
+ "project": "smolvla_omy",
190
+ "entity": "ai-liber",
191
+ "notes": null,
192
+ "run_id": null,
193
+ "mode": null
194
+ }
195
+ }
smolvla_omy/checkpoints/020000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,526 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.6875000000000013e-05,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "foreach": null,
12
+ "maximize": false,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "initial_lr": 0.0001,
17
+ "params": [
18
+ 0,
19
+ 1,
20
+ 2,
21
+ 3,
22
+ 4,
23
+ 5,
24
+ 6,
25
+ 7,
26
+ 8,
27
+ 9,
28
+ 10,
29
+ 11,
30
+ 12,
31
+ 13,
32
+ 14,
33
+ 15,
34
+ 16,
35
+ 17,
36
+ 18,
37
+ 19,
38
+ 20,
39
+ 21,
40
+ 22,
41
+ 23,
42
+ 24,
43
+ 25,
44
+ 26,
45
+ 27,
46
+ 28,
47
+ 29,
48
+ 30,
49
+ 31,
50
+ 32,
51
+ 33,
52
+ 34,
53
+ 35,
54
+ 36,
55
+ 37,
56
+ 38,
57
+ 39,
58
+ 40,
59
+ 41,
60
+ 42,
61
+ 43,
62
+ 44,
63
+ 45,
64
+ 46,
65
+ 47,
66
+ 48,
67
+ 49,
68
+ 50,
69
+ 51,
70
+ 52,
71
+ 53,
72
+ 54,
73
+ 55,
74
+ 56,
75
+ 57,
76
+ 58,
77
+ 59,
78
+ 60,
79
+ 61,
80
+ 62,
81
+ 63,
82
+ 64,
83
+ 65,
84
+ 66,
85
+ 67,
86
+ 68,
87
+ 69,
88
+ 70,
89
+ 71,
90
+ 72,
91
+ 73,
92
+ 74,
93
+ 75,
94
+ 76,
95
+ 77,
96
+ 78,
97
+ 79,
98
+ 80,
99
+ 81,
100
+ 82,
101
+ 83,
102
+ 84,
103
+ 85,
104
+ 86,
105
+ 87,
106
+ 88,
107
+ 89,
108
+ 90,
109
+ 91,
110
+ 92,
111
+ 93,
112
+ 94,
113
+ 95,
114
+ 96,
115
+ 97,
116
+ 98,
117
+ 99,
118
+ 100,
119
+ 101,
120
+ 102,
121
+ 103,
122
+ 104,
123
+ 105,
124
+ 106,
125
+ 107,
126
+ 108,
127
+ 109,
128
+ 110,
129
+ 111,
130
+ 112,
131
+ 113,
132
+ 114,
133
+ 115,
134
+ 116,
135
+ 117,
136
+ 118,
137
+ 119,
138
+ 120,
139
+ 121,
140
+ 122,
141
+ 123,
142
+ 124,
143
+ 125,
144
+ 126,
145
+ 127,
146
+ 128,
147
+ 129,
148
+ 130,
149
+ 131,
150
+ 132,
151
+ 133,
152
+ 134,
153
+ 135,
154
+ 136,
155
+ 137,
156
+ 138,
157
+ 139,
158
+ 140,
159
+ 141,
160
+ 142,
161
+ 143,
162
+ 144,
163
+ 145,
164
+ 146,
165
+ 147,
166
+ 148,
167
+ 149,
168
+ 150,
169
+ 151,
170
+ 152,
171
+ 153,
172
+ 154,
173
+ 155,
174
+ 156,
175
+ 157,
176
+ 158,
177
+ 159,
178
+ 160,
179
+ 161,
180
+ 162,
181
+ 163,
182
+ 164,
183
+ 165,
184
+ 166,
185
+ 167,
186
+ 168,
187
+ 169,
188
+ 170,
189
+ 171,
190
+ 172,
191
+ 173,
192
+ 174,
193
+ 175,
194
+ 176,
195
+ 177,
196
+ 178,
197
+ 179,
198
+ 180,
199
+ 181,
200
+ 182,
201
+ 183,
202
+ 184,
203
+ 185,
204
+ 186,
205
+ 187,
206
+ 188,
207
+ 189,
208
+ 190,
209
+ 191,
210
+ 192,
211
+ 193,
212
+ 194,
213
+ 195,
214
+ 196,
215
+ 197,
216
+ 198,
217
+ 199,
218
+ 200,
219
+ 201,
220
+ 202,
221
+ 203,
222
+ 204,
223
+ 205,
224
+ 206,
225
+ 207,
226
+ 208,
227
+ 209,
228
+ 210,
229
+ 211,
230
+ 212,
231
+ 213,
232
+ 214,
233
+ 215,
234
+ 216,
235
+ 217,
236
+ 218,
237
+ 219,
238
+ 220,
239
+ 221,
240
+ 222,
241
+ 223,
242
+ 224,
243
+ 225,
244
+ 226,
245
+ 227,
246
+ 228,
247
+ 229,
248
+ 230,
249
+ 231,
250
+ 232,
251
+ 233,
252
+ 234,
253
+ 235,
254
+ 236,
255
+ 237,
256
+ 238,
257
+ 239,
258
+ 240,
259
+ 241,
260
+ 242,
261
+ 243,
262
+ 244,
263
+ 245,
264
+ 246,
265
+ 247,
266
+ 248,
267
+ 249,
268
+ 250,
269
+ 251,
270
+ 252,
271
+ 253,
272
+ 254,
273
+ 255,
274
+ 256,
275
+ 257,
276
+ 258,
277
+ 259,
278
+ 260,
279
+ 261,
280
+ 262,
281
+ 263,
282
+ 264,
283
+ 265,
284
+ 266,
285
+ 267,
286
+ 268,
287
+ 269,
288
+ 270,
289
+ 271,
290
+ 272,
291
+ 273,
292
+ 274,
293
+ 275,
294
+ 276,
295
+ 277,
296
+ 278,
297
+ 279,
298
+ 280,
299
+ 281,
300
+ 282,
301
+ 283,
302
+ 284,
303
+ 285,
304
+ 286,
305
+ 287,
306
+ 288,
307
+ 289,
308
+ 290,
309
+ 291,
310
+ 292,
311
+ 293,
312
+ 294,
313
+ 295,
314
+ 296,
315
+ 297,
316
+ 298,
317
+ 299,
318
+ 300,
319
+ 301,
320
+ 302,
321
+ 303,
322
+ 304,
323
+ 305,
324
+ 306,
325
+ 307,
326
+ 308,
327
+ 309,
328
+ 310,
329
+ 311,
330
+ 312,
331
+ 313,
332
+ 314,
333
+ 315,
334
+ 316,
335
+ 317,
336
+ 318,
337
+ 319,
338
+ 320,
339
+ 321,
340
+ 322,
341
+ 323,
342
+ 324,
343
+ 325,
344
+ 326,
345
+ 327,
346
+ 328,
347
+ 329,
348
+ 330,
349
+ 331,
350
+ 332,
351
+ 333,
352
+ 334,
353
+ 335,
354
+ 336,
355
+ 337,
356
+ 338,
357
+ 339,
358
+ 340,
359
+ 341,
360
+ 342,
361
+ 343,
362
+ 344,
363
+ 345,
364
+ 346,
365
+ 347,
366
+ 348,
367
+ 349,
368
+ 350,
369
+ 351,
370
+ 352,
371
+ 353,
372
+ 354,
373
+ 355,
374
+ 356,
375
+ 357,
376
+ 358,
377
+ 359,
378
+ 360,
379
+ 361,
380
+ 362,
381
+ 363,
382
+ 364,
383
+ 365,
384
+ 366,
385
+ 367,
386
+ 368,
387
+ 369,
388
+ 370,
389
+ 371,
390
+ 372,
391
+ 373,
392
+ 374,
393
+ 375,
394
+ 376,
395
+ 377,
396
+ 378,
397
+ 379,
398
+ 380,
399
+ 381,
400
+ 382,
401
+ 383,
402
+ 384,
403
+ 385,
404
+ 386,
405
+ 387,
406
+ 388,
407
+ 389,
408
+ 390,
409
+ 391,
410
+ 392,
411
+ 393,
412
+ 394,
413
+ 395,
414
+ 396,
415
+ 397,
416
+ 398,
417
+ 399,
418
+ 400,
419
+ 401,
420
+ 402,
421
+ 403,
422
+ 404,
423
+ 405,
424
+ 406,
425
+ 407,
426
+ 408,
427
+ 409,
428
+ 410,
429
+ 411,
430
+ 412,
431
+ 413,
432
+ 414,
433
+ 415,
434
+ 416,
435
+ 417,
436
+ 418,
437
+ 419,
438
+ 420,
439
+ 421,
440
+ 422,
441
+ 423,
442
+ 424,
443
+ 425,
444
+ 426,
445
+ 427,
446
+ 428,
447
+ 429,
448
+ 430,
449
+ 431,
450
+ 432,
451
+ 433,
452
+ 434,
453
+ 435,
454
+ 436,
455
+ 437,
456
+ 438,
457
+ 439,
458
+ 440,
459
+ 441,
460
+ 442,
461
+ 443,
462
+ 444,
463
+ 445,
464
+ 446,
465
+ 447,
466
+ 448,
467
+ 449,
468
+ 450,
469
+ 451,
470
+ 452,
471
+ 453,
472
+ 454,
473
+ 455,
474
+ 456,
475
+ 457,
476
+ 458,
477
+ 459,
478
+ 460,
479
+ 461,
480
+ 462,
481
+ 463,
482
+ 464,
483
+ 465,
484
+ 466,
485
+ 467,
486
+ 468,
487
+ 469,
488
+ 470,
489
+ 471,
490
+ 472,
491
+ 473,
492
+ 474,
493
+ 475,
494
+ 476,
495
+ 477,
496
+ 478,
497
+ 479,
498
+ 480,
499
+ 481,
500
+ 482,
501
+ 483,
502
+ 484,
503
+ 485,
504
+ 486,
505
+ 487,
506
+ 488,
507
+ 489,
508
+ 490,
509
+ 491,
510
+ 492,
511
+ 493,
512
+ 494,
513
+ 495,
514
+ 496,
515
+ 497,
516
+ 498,
517
+ 499,
518
+ 500,
519
+ 501,
520
+ 502,
521
+ 503,
522
+ 504,
523
+ 505
524
+ ]
525
+ }
526
+ ]
smolvla_omy/checkpoints/020000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81b799f264abc0b6313d19f76ef35293ee8d31f9398f95be430935be49a2569c
3
+ size 412659164
smolvla_omy/checkpoints/020000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cea11deb7d2f3ac86ae2dea69aada4ca48ae69943294e53c8335c73d1141fb3
3
+ size 15708
smolvla_omy/checkpoints/020000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 20000,
6
+ "verbose": false,
7
+ "_step_count": 20001,
8
+ "_get_lr_called_within_step": false,
9
+ "_last_lr": [
10
+ 2.6875000000000013e-05
11
+ ],
12
+ "lr_lambdas": [
13
+ null
14
+ ]
15
+ }
smolvla_omy/checkpoints/020000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 20000
3
+ }
smolvla_omy/wandb/debug-internal.log CHANGED
@@ -7,3 +7,10 @@
7
  {"time":"2025-11-06T18:28:51.825582909+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
8
  {"time":"2025-11-06T19:15:51.940995974+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
9
  {"time":"2025-11-06T20:24:22.8723501+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
 
 
 
 
 
 
 
 
7
  {"time":"2025-11-06T18:28:51.825582909+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
8
  {"time":"2025-11-06T19:15:51.940995974+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
9
  {"time":"2025-11-06T20:24:22.8723501+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
10
+ {"time":"2025-11-06T21:50:57.568308958+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/ai-liber/smolvla_omy/8gs9of7r/file_stream\": dial tcp: lookup api.wandb.ai on 222.201.54.123:53: server misbehaving"}
11
+ {"time":"2025-11-06T21:51:00.931430036+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/ai-liber/smolvla_omy/8gs9of7r/file_stream\": dial tcp: lookup api.wandb.ai on 222.201.54.123:53: server misbehaving"}
12
+ {"time":"2025-11-06T22:09:29.4723197+08:00","level":"INFO","msg":"stream: closing","id":"8gs9of7r"}
13
+ {"time":"2025-11-06T22:09:30.974870957+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
14
+ {"time":"2025-11-06T22:09:31.653721675+08:00","level":"INFO","msg":"handler: closed","stream_id":"8gs9of7r"}
15
+ {"time":"2025-11-06T22:09:31.653947554+08:00","level":"INFO","msg":"sender: closed","stream_id":"8gs9of7r"}
16
+ {"time":"2025-11-06T22:09:31.653974862+08:00","level":"INFO","msg":"stream: closed","id":"8gs9of7r"}
smolvla_omy/wandb/debug.log CHANGED
@@ -19,3 +19,5 @@ config: {'dataset': {'repo_id': 'omy_pnp_language', 'root': './demo_data_languag
19
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2446] Redirects installed.
21
  2025-11-06 16:21:51,416 INFO MainThread:528220 [wandb_init.py:init():1073] run started, returning control to user process
 
 
 
19
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2446] Redirects installed.
21
  2025-11-06 16:21:51,416 INFO MainThread:528220 [wandb_init.py:init():1073] run started, returning control to user process
22
+ 2025-11-06 22:09:29,472 INFO wandb-AsyncioManager-main:528220 [service_client.py:_forward_responses():80] Reached EOF.
23
+ 2025-11-06 22:09:29,472 INFO wandb-AsyncioManager-main:528220 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles.
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/config.yaml ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.22.3
4
+ e:
5
+ qj4c6zn4buu7s32sv5nl108gvz7rnfa0:
6
+ args:
7
+ - --config_path
8
+ - smolvla_omy.yaml
9
+ codePath: train_model.py
10
+ codePathLocal: train_model.py
11
+ cpu_count: 40
12
+ cpu_count_logical: 80
13
+ cudaVersion: "12.4"
14
+ disk:
15
+ /:
16
+ total: "1676411871232"
17
+ used: "1481895833600"
18
+ executable: /opt/conda/envs/lerobot_mujoco_tutorial/bin/python
19
+ git:
20
+ commit: b11377a0ce7a3c32cefe260c6dce959b0db2bdda
21
+ remote: git@github.com:liberow/lerobot-mujoco-tutorial.git
22
+ gpu: Tesla V100-SXM2-32GB
23
+ gpu_count: 4
24
+ gpu_nvidia:
25
+ - architecture: Volta
26
+ cudaCores: 5120
27
+ memoryTotal: "34359738368"
28
+ name: Tesla V100-SXM2-32GB
29
+ uuid: GPU-b29b46dc-42a3-9fcc-be17-50b8b8972960
30
+ - architecture: Volta
31
+ cudaCores: 5120
32
+ memoryTotal: "34359738368"
33
+ name: Tesla V100-SXM2-32GB
34
+ uuid: GPU-368aa17c-5504-4d63-4df2-339d38a17fe3
35
+ - architecture: Volta
36
+ cudaCores: 5120
37
+ memoryTotal: "34359738368"
38
+ name: Tesla V100-SXM2-32GB
39
+ uuid: GPU-acf7b3cc-f154-087a-3a06-d57e023c09be
40
+ - architecture: Volta
41
+ cudaCores: 5120
42
+ memoryTotal: "34359738368"
43
+ name: Tesla V100-SXM2-32GB
44
+ uuid: GPU-86b115b3-0485-dd33-56f3-385a1140ae36
45
+ host: c81ab3b21da2
46
+ memory:
47
+ total: "134767624192"
48
+ os: Linux-6.8.0-51-generic-x86_64-with-glibc2.35
49
+ program: /workspace/liber/embodied_ai/lerobot-mujoco-tutorial/train_model.py
50
+ python: CPython 3.10.19
51
+ root: ckpt/smolvla_omy
52
+ startedAt: "2025-11-06T08:21:49.583849Z"
53
+ writerId: qj4c6zn4buu7s32sv5nl108gvz7rnfa0
54
+ m: []
55
+ python_version: 3.10.19
56
+ t:
57
+ "1":
58
+ - 1
59
+ - 41
60
+ - 49
61
+ - 51
62
+ "2":
63
+ - 1
64
+ - 11
65
+ - 41
66
+ - 49
67
+ - 51
68
+ - 71
69
+ "3":
70
+ - 13
71
+ - 15
72
+ - 16
73
+ - 61
74
+ "4": 3.10.19
75
+ "5": 0.22.3
76
+ "10":
77
+ - 21
78
+ "12": 0.22.3
79
+ "13": linux-x86_64
80
+ batch_size:
81
+ value: 16
82
+ dataset:
83
+ value:
84
+ episodes: null
85
+ image_transforms:
86
+ enable: false
87
+ max_num_transforms: 3
88
+ random_order: false
89
+ tfs:
90
+ brightness:
91
+ kwargs:
92
+ brightness:
93
+ - 0.8
94
+ - 1.2
95
+ type: ColorJitter
96
+ weight: 1
97
+ contrast:
98
+ kwargs:
99
+ contrast:
100
+ - 0.8
101
+ - 1.2
102
+ type: ColorJitter
103
+ weight: 1
104
+ hue:
105
+ kwargs:
106
+ hue:
107
+ - -0.05
108
+ - 0.05
109
+ type: ColorJitter
110
+ weight: 1
111
+ saturation:
112
+ kwargs:
113
+ saturation:
114
+ - 0.5
115
+ - 1.5
116
+ type: ColorJitter
117
+ weight: 1
118
+ sharpness:
119
+ kwargs:
120
+ sharpness:
121
+ - 0.5
122
+ - 1.5
123
+ type: SharpnessJitter
124
+ weight: 1
125
+ repo_id: omy_pnp_language
126
+ revision: null
127
+ root: ./demo_data_language
128
+ use_imagenet_stats: true
129
+ video_backend: torchcodec
130
+ env:
131
+ value: null
132
+ eval:
133
+ value:
134
+ batch_size: 50
135
+ n_episodes: 50
136
+ use_async_envs: false
137
+ eval_freq:
138
+ value: -1
139
+ job_name:
140
+ value: smolvla_omy
141
+ log_freq:
142
+ value: 50
143
+ num_workers:
144
+ value: 8
145
+ optimizer:
146
+ value:
147
+ betas:
148
+ - 0.9
149
+ - 0.95
150
+ eps: 1e-08
151
+ grad_clip_norm: 10
152
+ lr: 0.0001
153
+ type: adamw
154
+ weight_decay: 1e-10
155
+ output_dir:
156
+ value: ckpt/smolvla_omy
157
+ policy:
158
+ value:
159
+ adapt_to_pi_aloha: false
160
+ add_image_special_tokens: false
161
+ attention_mode: cross_attn
162
+ chunk_size: 5
163
+ device: cuda
164
+ empty_cameras: 0
165
+ expert_width_multiplier: 0.75
166
+ freeze_vision_encoder: true
167
+ load_vlm_weights: false
168
+ max_action_dim: 32
169
+ max_period: 4
170
+ max_state_dim: 32
171
+ min_period: 0.004
172
+ n_action_steps: 5
173
+ n_obs_steps: 1
174
+ normalization_mapping:
175
+ ACTION: MEAN_STD
176
+ STATE: MEAN_STD
177
+ VISUAL: IDENTITY
178
+ num_expert_layers: -1
179
+ num_steps: 10
180
+ num_vlm_layers: 16
181
+ optimizer_betas:
182
+ - 0.9
183
+ - 0.95
184
+ optimizer_eps: 1e-08
185
+ optimizer_grad_clip_norm: 10
186
+ optimizer_lr: 0.0001
187
+ optimizer_weight_decay: 1e-10
188
+ pad_language_to: longest
189
+ prefix_length: -1
190
+ resize_imgs_with_padding:
191
+ - 512
192
+ - 512
193
+ scheduler_decay_lr: 2.5e-06
194
+ scheduler_decay_steps: 30000
195
+ scheduler_warmup_steps: 1000
196
+ self_attn_every_n_layers: 2
197
+ tokenizer_max_length: 48
198
+ train_expert_only: true
199
+ train_state_proj: true
200
+ type: smolvla
201
+ use_amp: false
202
+ use_cache: true
203
+ use_delta_joint_actions_aloha: false
204
+ vlm_model_name: HuggingFaceTB/SmolVLM2-500M-Video-Instruct
205
+ resume:
206
+ value: false
207
+ save_checkpoint:
208
+ value: true
209
+ save_freq:
210
+ value: 5000
211
+ scheduler:
212
+ value:
213
+ decay_lr: 2.5e-06
214
+ num_decay_steps: 30000
215
+ num_warmup_steps: 1000
216
+ peak_lr: 0.0001
217
+ type: cosine_decay_with_warmup
218
+ seed:
219
+ value: 42
220
+ steps:
221
+ value: 20000
222
+ use_policy_training_preset:
223
+ value: true
224
+ wandb:
225
+ value:
226
+ disable_artifact: true
227
+ enable: true
228
+ entity: ai-liber
229
+ mode: null
230
+ notes: null
231
+ project: smolvla_omy
232
+ run_id: null
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/output.log CHANGED
@@ -981,3 +981,239 @@ WARNING 2025-11-06 21:03:04 db_utils.py:117 WandB logging of key "losses_after_r
981
  INFO 2025-11-06 21:03:55 in_model.py:236 step:16K smpl:258K ep:2K epch:98.28 loss:0.008 grdn:0.249 lr:4.6e-05 updt_s:0.995 data_s:0.022
982
  WARNING 2025-11-06 21:03:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
983
  WARNING 2025-11-06 21:03:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
981
  INFO 2025-11-06 21:03:55 in_model.py:236 step:16K smpl:258K ep:2K epch:98.28 loss:0.008 grdn:0.249 lr:4.6e-05 updt_s:0.995 data_s:0.022
982
  WARNING 2025-11-06 21:03:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
983
  WARNING 2025-11-06 21:03:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
984
+ INFO 2025-11-06 21:04:45 in_model.py:236 step:16K smpl:258K ep:2K epch:98.59 loss:0.009 grdn:0.263 lr:4.6e-05 updt_s:0.996 data_s:0.000
985
+ WARNING 2025-11-06 21:04:45 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
986
+ WARNING 2025-11-06 21:04:45 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
987
+ INFO 2025-11-06 21:05:35 in_model.py:236 step:16K smpl:259K ep:2K epch:98.89 loss:0.009 grdn:0.263 lr:4.5e-05 updt_s:0.996 data_s:0.000
988
+ WARNING 2025-11-06 21:05:35 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
989
+ WARNING 2025-11-06 21:05:35 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
990
+ INFO 2025-11-06 21:06:26 in_model.py:236 step:16K smpl:260K ep:2K epch:99.20 loss:0.009 grdn:0.279 lr:4.5e-05 updt_s:0.994 data_s:0.022
991
+ WARNING 2025-11-06 21:06:26 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
992
+ WARNING 2025-11-06 21:06:26 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
993
+ INFO 2025-11-06 21:07:16 in_model.py:236 step:16K smpl:261K ep:2K epch:99.50 loss:0.009 grdn:0.248 lr:4.5e-05 updt_s:0.999 data_s:0.000
994
+ WARNING 2025-11-06 21:07:16 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
995
+ WARNING 2025-11-06 21:07:16 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
996
+ INFO 2025-11-06 21:08:06 in_model.py:236 step:16K smpl:262K ep:2K epch:99.81 loss:0.008 grdn:0.231 lr:4.5e-05 updt_s:1.001 data_s:0.000
997
+ WARNING 2025-11-06 21:08:06 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
998
+ WARNING 2025-11-06 21:08:06 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
999
+ INFO 2025-11-06 21:08:57 in_model.py:236 step:16K smpl:262K ep:2K epch:100.11 loss:0.008 grdn:0.249 lr:4.4e-05 updt_s:0.997 data_s:0.000
1000
+ WARNING 2025-11-06 21:08:57 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1001
+ WARNING 2025-11-06 21:08:57 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1002
+ INFO 2025-11-06 21:09:48 in_model.py:236 step:16K smpl:263K ep:2K epch:100.42 loss:0.009 grdn:0.273 lr:4.4e-05 updt_s:1.000 data_s:0.025
1003
+ WARNING 2025-11-06 21:09:48 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1004
+ WARNING 2025-11-06 21:09:48 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1005
+ INFO 2025-11-06 21:10:38 in_model.py:236 step:16K smpl:264K ep:2K epch:100.72 loss:0.009 grdn:0.252 lr:4.4e-05 updt_s:1.000 data_s:0.000
1006
+ WARNING 2025-11-06 21:10:38 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1007
+ WARNING 2025-11-06 21:10:38 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1008
+ INFO 2025-11-06 21:11:28 in_model.py:236 step:17K smpl:265K ep:2K epch:101.03 loss:0.008 grdn:0.246 lr:4.3e-05 updt_s:1.001 data_s:0.000
1009
+ WARNING 2025-11-06 21:11:28 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1010
+ WARNING 2025-11-06 21:11:28 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1011
+ INFO 2025-11-06 21:12:20 in_model.py:236 step:17K smpl:266K ep:2K epch:101.34 loss:0.008 grdn:0.273 lr:4.3e-05 updt_s:0.994 data_s:0.025
1012
+ WARNING 2025-11-06 21:12:20 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1013
+ WARNING 2025-11-06 21:12:20 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1014
+ INFO 2025-11-06 21:13:10 in_model.py:236 step:17K smpl:266K ep:2K epch:101.64 loss:0.007 grdn:0.213 lr:4.3e-05 updt_s:0.997 data_s:0.000
1015
+ WARNING 2025-11-06 21:13:10 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1016
+ WARNING 2025-11-06 21:13:10 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1017
+ INFO 2025-11-06 21:14:00 in_model.py:236 step:17K smpl:267K ep:2K epch:101.95 loss:0.010 grdn:0.284 lr:4.3e-05 updt_s:0.997 data_s:0.000
1018
+ WARNING 2025-11-06 21:14:00 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1019
+ WARNING 2025-11-06 21:14:00 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1020
+ INFO 2025-11-06 21:14:51 in_model.py:236 step:17K smpl:268K ep:2K epch:102.25 loss:0.009 grdn:0.281 lr:4.2e-05 updt_s:0.993 data_s:0.024
1021
+ WARNING 2025-11-06 21:14:51 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1022
+ WARNING 2025-11-06 21:14:51 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1023
+ INFO 2025-11-06 21:15:41 in_model.py:236 step:17K smpl:269K ep:2K epch:102.56 loss:0.009 grdn:0.227 lr:4.2e-05 updt_s:0.996 data_s:0.000
1024
+ WARNING 2025-11-06 21:15:41 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1025
+ WARNING 2025-11-06 21:15:41 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1026
+ INFO 2025-11-06 21:16:31 in_model.py:236 step:17K smpl:270K ep:2K epch:102.86 loss:0.010 grdn:0.310 lr:4.2e-05 updt_s:0.995 data_s:0.000
1027
+ WARNING 2025-11-06 21:16:31 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1028
+ WARNING 2025-11-06 21:16:31 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1029
+ INFO 2025-11-06 21:17:22 in_model.py:236 step:17K smpl:270K ep:2K epch:103.17 loss:0.009 grdn:0.269 lr:4.2e-05 updt_s:0.993 data_s:0.023
1030
+ WARNING 2025-11-06 21:17:22 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1031
+ WARNING 2025-11-06 21:17:22 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1032
+ INFO 2025-11-06 21:18:12 in_model.py:236 step:17K smpl:271K ep:2K epch:103.47 loss:0.010 grdn:0.273 lr:4.1e-05 updt_s:0.997 data_s:0.000
1033
+ WARNING 2025-11-06 21:18:12 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1034
+ WARNING 2025-11-06 21:18:12 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1035
+ INFO 2025-11-06 21:19:02 in_model.py:236 step:17K smpl:272K ep:2K epch:103.78 loss:0.008 grdn:0.260 lr:4.1e-05 updt_s:0.997 data_s:0.000
1036
+ WARNING 2025-11-06 21:19:02 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1037
+ WARNING 2025-11-06 21:19:02 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1038
+ INFO 2025-11-06 21:19:52 in_model.py:236 step:17K smpl:273K ep:2K epch:104.08 loss:0.009 grdn:0.291 lr:4.1e-05 updt_s:0.996 data_s:0.000
1039
+ WARNING 2025-11-06 21:19:52 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1040
+ WARNING 2025-11-06 21:19:52 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1041
+ INFO 2025-11-06 21:20:43 in_model.py:236 step:17K smpl:274K ep:2K epch:104.39 loss:0.009 grdn:0.265 lr:4.1e-05 updt_s:0.995 data_s:0.022
1042
+ WARNING 2025-11-06 21:20:43 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1043
+ WARNING 2025-11-06 21:20:43 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1044
+ INFO 2025-11-06 21:21:33 in_model.py:236 step:17K smpl:274K ep:2K epch:104.69 loss:0.007 grdn:0.220 lr:4.0e-05 updt_s:0.998 data_s:0.000
1045
+ WARNING 2025-11-06 21:21:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1046
+ WARNING 2025-11-06 21:21:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1047
+ INFO 2025-11-06 21:22:23 in_model.py:236 step:17K smpl:275K ep:2K epch:105.00 loss:0.009 grdn:0.267 lr:4.0e-05 updt_s:1.000 data_s:0.000
1048
+ WARNING 2025-11-06 21:22:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1049
+ WARNING 2025-11-06 21:22:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1050
+ INFO 2025-11-06 21:23:14 in_model.py:236 step:17K smpl:276K ep:2K epch:105.30 loss:0.010 grdn:0.271 lr:4.0e-05 updt_s:1.000 data_s:0.023
1051
+ WARNING 2025-11-06 21:23:14 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1052
+ WARNING 2025-11-06 21:23:14 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1053
+ INFO 2025-11-06 21:24:05 in_model.py:236 step:17K smpl:277K ep:2K epch:105.61 loss:0.007 grdn:0.213 lr:4.0e-05 updt_s:1.001 data_s:0.000
1054
+ WARNING 2025-11-06 21:24:05 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1055
+ WARNING 2025-11-06 21:24:05 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1056
+ INFO 2025-11-06 21:24:55 in_model.py:236 step:17K smpl:278K ep:2K epch:105.91 loss:0.009 grdn:0.278 lr:3.9e-05 updt_s:1.004 data_s:0.000
1057
+ WARNING 2025-11-06 21:24:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1058
+ WARNING 2025-11-06 21:24:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1059
+ INFO 2025-11-06 21:25:46 in_model.py:236 step:17K smpl:278K ep:2K epch:106.22 loss:0.010 grdn:0.279 lr:3.9e-05 updt_s:0.999 data_s:0.021
1060
+ WARNING 2025-11-06 21:25:46 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1061
+ WARNING 2025-11-06 21:25:46 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1062
+ INFO 2025-11-06 21:26:37 in_model.py:236 step:17K smpl:279K ep:2K epch:106.52 loss:0.009 grdn:0.259 lr:3.9e-05 updt_s:1.003 data_s:0.000
1063
+ WARNING 2025-11-06 21:26:37 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1064
+ WARNING 2025-11-06 21:26:37 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1065
+ INFO 2025-11-06 21:27:27 in_model.py:236 step:18K smpl:280K ep:2K epch:106.83 loss:0.008 grdn:0.260 lr:3.9e-05 updt_s:1.003 data_s:0.000
1066
+ WARNING 2025-11-06 21:27:27 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1067
+ WARNING 2025-11-06 21:27:27 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1068
+ INFO 2025-11-06 21:28:18 in_model.py:236 step:18K smpl:281K ep:2K epch:107.13 loss:0.009 grdn:0.232 lr:3.9e-05 updt_s:0.999 data_s:0.021
1069
+ WARNING 2025-11-06 21:28:18 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1070
+ WARNING 2025-11-06 21:28:18 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1071
+ INFO 2025-11-06 21:29:08 in_model.py:236 step:18K smpl:282K ep:2K epch:107.44 loss:0.008 grdn:0.261 lr:3.8e-05 updt_s:1.000 data_s:0.000
1072
+ WARNING 2025-11-06 21:29:08 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1073
+ WARNING 2025-11-06 21:29:08 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1074
+ INFO 2025-11-06 21:29:59 in_model.py:236 step:18K smpl:282K ep:2K epch:107.75 loss:0.008 grdn:0.263 lr:3.8e-05 updt_s:0.999 data_s:0.000
1075
+ WARNING 2025-11-06 21:29:59 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1076
+ WARNING 2025-11-06 21:29:59 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1077
+ INFO 2025-11-06 21:30:49 in_model.py:236 step:18K smpl:283K ep:2K epch:108.05 loss:0.009 grdn:0.243 lr:3.8e-05 updt_s:1.001 data_s:0.000
1078
+ WARNING 2025-11-06 21:30:49 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1079
+ WARNING 2025-11-06 21:30:49 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1080
+ INFO 2025-11-06 21:31:40 in_model.py:236 step:18K smpl:284K ep:2K epch:108.36 loss:0.009 grdn:0.262 lr:3.8e-05 updt_s:0.998 data_s:0.021
1081
+ WARNING 2025-11-06 21:31:40 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1082
+ WARNING 2025-11-06 21:31:40 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1083
+ INFO 2025-11-06 21:32:30 in_model.py:236 step:18K smpl:285K ep:2K epch:108.66 loss:0.007 grdn:0.230 lr:3.7e-05 updt_s:1.001 data_s:0.000
1084
+ WARNING 2025-11-06 21:32:30 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1085
+ WARNING 2025-11-06 21:32:30 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1086
+ INFO 2025-11-06 21:33:20 in_model.py:236 step:18K smpl:286K ep:2K epch:108.97 loss:0.008 grdn:0.233 lr:3.7e-05 updt_s:0.999 data_s:0.000
1087
+ WARNING 2025-11-06 21:33:20 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1088
+ WARNING 2025-11-06 21:33:20 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1089
+ INFO 2025-11-06 21:34:11 in_model.py:236 step:18K smpl:286K ep:2K epch:109.27 loss:0.007 grdn:0.240 lr:3.7e-05 updt_s:0.994 data_s:0.021
1090
+ WARNING 2025-11-06 21:34:11 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1091
+ WARNING 2025-11-06 21:34:11 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1092
+ INFO 2025-11-06 21:35:01 in_model.py:236 step:18K smpl:287K ep:2K epch:109.58 loss:0.009 grdn:0.239 lr:3.7e-05 updt_s:0.998 data_s:0.000
1093
+ WARNING 2025-11-06 21:35:01 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1094
+ WARNING 2025-11-06 21:35:01 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1095
+ INFO 2025-11-06 21:35:51 in_model.py:236 step:18K smpl:288K ep:2K epch:109.88 loss:0.009 grdn:0.291 lr:3.6e-05 updt_s:0.997 data_s:0.000
1096
+ WARNING 2025-11-06 21:35:51 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1097
+ WARNING 2025-11-06 21:35:51 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1098
+ INFO 2025-11-06 21:36:42 in_model.py:236 step:18K smpl:289K ep:2K epch:110.19 loss:0.009 grdn:0.259 lr:3.6e-05 updt_s:0.994 data_s:0.022
1099
+ WARNING 2025-11-06 21:36:42 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1100
+ WARNING 2025-11-06 21:36:42 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1101
+ INFO 2025-11-06 21:37:33 in_model.py:236 step:18K smpl:290K ep:2K epch:110.49 loss:0.007 grdn:0.213 lr:3.6e-05 updt_s:0.999 data_s:0.000
1102
+ WARNING 2025-11-06 21:37:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1103
+ WARNING 2025-11-06 21:37:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1104
+ INFO 2025-11-06 21:38:23 in_model.py:236 step:18K smpl:290K ep:2K epch:110.80 loss:0.008 grdn:0.264 lr:3.6e-05 updt_s:0.997 data_s:0.000
1105
+ WARNING 2025-11-06 21:38:23 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1106
+ WARNING 2025-11-06 21:38:23 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1107
+ INFO 2025-11-06 21:39:13 in_model.py:236 step:18K smpl:291K ep:2K epch:111.10 loss:0.007 grdn:0.230 lr:3.5e-05 updt_s:0.998 data_s:0.000
1108
+ WARNING 2025-11-06 21:39:13 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1109
+ WARNING 2025-11-06 21:39:13 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1110
+ INFO 2025-11-06 21:40:04 in_model.py:236 step:18K smpl:292K ep:2K epch:111.41 loss:0.008 grdn:0.236 lr:3.5e-05 updt_s:0.994 data_s:0.021
1111
+ WARNING 2025-11-06 21:40:04 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1112
+ WARNING 2025-11-06 21:40:04 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1113
+ INFO 2025-11-06 21:40:54 in_model.py:236 step:18K smpl:293K ep:2K epch:111.71 loss:0.008 grdn:0.235 lr:3.5e-05 updt_s:0.997 data_s:0.000
1114
+ WARNING 2025-11-06 21:40:54 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1115
+ WARNING 2025-11-06 21:40:54 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1116
+ INFO 2025-11-06 21:41:44 in_model.py:236 step:18K smpl:294K ep:2K epch:112.02 loss:0.007 grdn:0.214 lr:3.5e-05 updt_s:0.997 data_s:0.000
1117
+ WARNING 2025-11-06 21:41:44 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1118
+ WARNING 2025-11-06 21:41:44 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1119
+ INFO 2025-11-06 21:42:35 in_model.py:236 step:18K smpl:294K ep:2K epch:112.32 loss:0.010 grdn:0.257 lr:3.4e-05 updt_s:0.993 data_s:0.021
1120
+ WARNING 2025-11-06 21:42:35 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1121
+ WARNING 2025-11-06 21:42:35 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1122
+ INFO 2025-11-06 21:43:25 in_model.py:236 step:18K smpl:295K ep:2K epch:112.63 loss:0.007 grdn:0.212 lr:3.4e-05 updt_s:0.997 data_s:0.000
1123
+ WARNING 2025-11-06 21:43:25 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1124
+ WARNING 2025-11-06 21:43:25 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1125
+ INFO 2025-11-06 21:44:15 in_model.py:236 step:18K smpl:296K ep:2K epch:112.93 loss:0.007 grdn:0.222 lr:3.4e-05 updt_s:0.997 data_s:0.000
1126
+ WARNING 2025-11-06 21:44:15 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1127
+ WARNING 2025-11-06 21:44:15 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1128
+ INFO 2025-11-06 21:45:06 in_model.py:236 step:19K smpl:297K ep:2K epch:113.24 loss:0.008 grdn:0.240 lr:3.4e-05 updt_s:0.993 data_s:0.022
1129
+ WARNING 2025-11-06 21:45:06 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1130
+ WARNING 2025-11-06 21:45:06 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1131
+ INFO 2025-11-06 21:45:56 in_model.py:236 step:19K smpl:298K ep:2K epch:113.54 loss:0.008 grdn:0.227 lr:3.3e-05 updt_s:0.994 data_s:0.000
1132
+ WARNING 2025-11-06 21:45:56 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1133
+ WARNING 2025-11-06 21:45:56 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1134
+ INFO 2025-11-06 21:46:45 in_model.py:236 step:19K smpl:298K ep:2K epch:113.85 loss:0.009 grdn:0.245 lr:3.3e-05 updt_s:0.995 data_s:0.000
1135
+ WARNING 2025-11-06 21:46:46 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1136
+ WARNING 2025-11-06 21:46:46 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1137
+ INFO 2025-11-06 21:47:36 in_model.py:236 step:19K smpl:299K ep:2K epch:114.15 loss:0.007 grdn:0.248 lr:3.3e-05 updt_s:0.994 data_s:0.021
1138
+ WARNING 2025-11-06 21:47:36 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1139
+ WARNING 2025-11-06 21:47:36 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1140
+ INFO 2025-11-06 21:48:26 in_model.py:236 step:19K smpl:300K ep:2K epch:114.46 loss:0.008 grdn:0.231 lr:3.3e-05 updt_s:0.996 data_s:0.000
1141
+ WARNING 2025-11-06 21:48:26 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1142
+ WARNING 2025-11-06 21:48:26 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1143
+ INFO 2025-11-06 21:49:17 in_model.py:236 step:19K smpl:301K ep:2K epch:114.77 loss:0.010 grdn:0.285 lr:3.2e-05 updt_s:0.998 data_s:0.000
1144
+ WARNING 2025-11-06 21:49:17 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1145
+ WARNING 2025-11-06 21:49:17 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1146
+ INFO 2025-11-06 21:50:07 in_model.py:236 step:19K smpl:302K ep:2K epch:115.07 loss:0.008 grdn:0.242 lr:3.2e-05 updt_s:0.999 data_s:0.000
1147
+ WARNING 2025-11-06 21:50:07 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1148
+ WARNING 2025-11-06 21:50:07 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1149
+ INFO 2025-11-06 21:50:58 in_model.py:236 step:19K smpl:302K ep:2K epch:115.38 loss:0.009 grdn:0.217 lr:3.2e-05 updt_s:0.995 data_s:0.022
1150
+ WARNING 2025-11-06 21:50:58 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1151
+ WARNING 2025-11-06 21:50:58 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1152
+ INFO 2025-11-06 21:51:48 in_model.py:236 step:19K smpl:303K ep:2K epch:115.68 loss:0.007 grdn:0.218 lr:3.2e-05 updt_s:0.999 data_s:0.000
1153
+ WARNING 2025-11-06 21:51:48 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1154
+ WARNING 2025-11-06 21:51:48 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1155
+ INFO 2025-11-06 21:52:38 in_model.py:236 step:19K smpl:304K ep:2K epch:115.99 loss:0.007 grdn:0.236 lr:3.2e-05 updt_s:0.998 data_s:0.000
1156
+ WARNING 2025-11-06 21:52:38 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1157
+ WARNING 2025-11-06 21:52:38 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1158
+ INFO 2025-11-06 21:53:29 in_model.py:236 step:19K smpl:305K ep:2K epch:116.29 loss:0.008 grdn:0.254 lr:3.1e-05 updt_s:0.995 data_s:0.021
1159
+ WARNING 2025-11-06 21:53:29 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1160
+ WARNING 2025-11-06 21:53:29 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1161
+ INFO 2025-11-06 21:54:19 in_model.py:236 step:19K smpl:306K ep:2K epch:116.60 loss:0.007 grdn:0.224 lr:3.1e-05 updt_s:0.999 data_s:0.000
1162
+ WARNING 2025-11-06 21:54:19 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1163
+ WARNING 2025-11-06 21:54:19 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1164
+ INFO 2025-11-06 21:55:09 in_model.py:236 step:19K smpl:306K ep:2K epch:116.90 loss:0.008 grdn:0.241 lr:3.1e-05 updt_s:0.999 data_s:0.000
1165
+ WARNING 2025-11-06 21:55:09 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1166
+ WARNING 2025-11-06 21:55:09 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1167
+ INFO 2025-11-06 21:56:00 in_model.py:236 step:19K smpl:307K ep:2K epch:117.21 loss:0.007 grdn:0.192 lr:3.1e-05 updt_s:0.997 data_s:0.021
1168
+ WARNING 2025-11-06 21:56:00 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1169
+ WARNING 2025-11-06 21:56:00 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1170
+ INFO 2025-11-06 21:56:51 in_model.py:236 step:19K smpl:308K ep:2K epch:117.51 loss:0.009 grdn:0.299 lr:3.0e-05 updt_s:0.999 data_s:0.000
1171
+ WARNING 2025-11-06 21:56:51 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1172
+ WARNING 2025-11-06 21:56:51 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1173
+ INFO 2025-11-06 21:57:41 in_model.py:236 step:19K smpl:309K ep:2K epch:117.82 loss:0.007 grdn:0.240 lr:3.0e-05 updt_s:0.997 data_s:0.000
1174
+ WARNING 2025-11-06 21:57:41 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1175
+ WARNING 2025-11-06 21:57:41 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1176
+ INFO 2025-11-06 21:58:31 in_model.py:236 step:19K smpl:310K ep:2K epch:118.12 loss:0.008 grdn:0.235 lr:3.0e-05 updt_s:0.997 data_s:0.000
1177
+ WARNING 2025-11-06 21:58:31 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1178
+ WARNING 2025-11-06 21:58:31 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1179
+ INFO 2025-11-06 21:59:22 in_model.py:236 step:19K smpl:310K ep:2K epch:118.43 loss:0.007 grdn:0.230 lr:3.0e-05 updt_s:0.994 data_s:0.022
1180
+ WARNING 2025-11-06 21:59:22 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1181
+ WARNING 2025-11-06 21:59:22 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1182
+ INFO 2025-11-06 22:00:12 in_model.py:236 step:19K smpl:311K ep:2K epch:118.73 loss:0.007 grdn:0.232 lr:2.9e-05 updt_s:0.999 data_s:0.000
1183
+ WARNING 2025-11-06 22:00:12 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1184
+ WARNING 2025-11-06 22:00:12 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1185
+ INFO 2025-11-06 22:01:02 in_model.py:236 step:20K smpl:312K ep:2K epch:119.04 loss:0.008 grdn:0.233 lr:2.9e-05 updt_s:1.001 data_s:0.000
1186
+ WARNING 2025-11-06 22:01:02 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1187
+ WARNING 2025-11-06 22:01:02 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1188
+ INFO 2025-11-06 22:01:53 in_model.py:236 step:20K smpl:313K ep:2K epch:119.34 loss:0.007 grdn:0.230 lr:2.9e-05 updt_s:0.995 data_s:0.023
1189
+ WARNING 2025-11-06 22:01:53 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1190
+ WARNING 2025-11-06 22:01:53 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1191
+ INFO 2025-11-06 22:02:43 in_model.py:236 step:20K smpl:314K ep:2K epch:119.65 loss:0.007 grdn:0.215 lr:2.9e-05 updt_s:0.996 data_s:0.000
1192
+ WARNING 2025-11-06 22:02:43 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1193
+ WARNING 2025-11-06 22:02:43 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1194
+ INFO 2025-11-06 22:03:33 in_model.py:236 step:20K smpl:314K ep:2K epch:119.95 loss:0.008 grdn:0.236 lr:2.9e-05 updt_s:0.997 data_s:0.000
1195
+ WARNING 2025-11-06 22:03:33 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1196
+ WARNING 2025-11-06 22:03:33 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1197
+ INFO 2025-11-06 22:04:24 in_model.py:236 step:20K smpl:315K ep:2K epch:120.26 loss:0.007 grdn:0.199 lr:2.8e-05 updt_s:0.993 data_s:0.022
1198
+ WARNING 2025-11-06 22:04:24 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1199
+ WARNING 2025-11-06 22:04:24 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1200
+ INFO 2025-11-06 22:05:14 in_model.py:236 step:20K smpl:316K ep:2K epch:120.56 loss:0.005 grdn:0.182 lr:2.8e-05 updt_s:0.997 data_s:0.000
1201
+ WARNING 2025-11-06 22:05:14 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1202
+ WARNING 2025-11-06 22:05:14 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1203
+ INFO 2025-11-06 22:06:04 in_model.py:236 step:20K smpl:317K ep:2K epch:120.87 loss:0.007 grdn:0.227 lr:2.8e-05 updt_s:0.999 data_s:0.000
1204
+ WARNING 2025-11-06 22:06:04 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1205
+ WARNING 2025-11-06 22:06:04 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1206
+ INFO 2025-11-06 22:06:55 in_model.py:236 step:20K smpl:318K ep:2K epch:121.18 loss:0.007 grdn:0.188 lr:2.8e-05 updt_s:0.993 data_s:0.022
1207
+ WARNING 2025-11-06 22:06:55 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1208
+ WARNING 2025-11-06 22:06:55 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1209
+ INFO 2025-11-06 22:07:45 in_model.py:236 step:20K smpl:318K ep:2K epch:121.48 loss:0.007 grdn:0.226 lr:2.7e-05 updt_s:0.994 data_s:0.000
1210
+ WARNING 2025-11-06 22:07:45 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1211
+ WARNING 2025-11-06 22:07:45 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1212
+ INFO 2025-11-06 22:08:36 in_model.py:236 step:20K smpl:319K ep:2K epch:121.79 loss:0.008 grdn:0.219 lr:2.7e-05 updt_s:1.003 data_s:0.000
1213
+ WARNING 2025-11-06 22:08:36 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1214
+ WARNING 2025-11-06 22:08:36 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1215
+ INFO 2025-11-06 22:09:26 in_model.py:236 step:20K smpl:320K ep:2K epch:122.09 loss:0.007 grdn:0.210 lr:2.7e-05 updt_s:1.000 data_s:0.000
1216
+ WARNING 2025-11-06 22:09:26 db_utils.py:117 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1217
+ WARNING 2025-11-06 22:09:26 db_utils.py:117 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
1218
+ INFO 2025-11-06 22:09:26 in_model.py:245 Checkpoint policy after step 20000
1219
+ INFO 2025-11-06 22:09:29 in_model.py:287 End of training
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"train/update_s":1.0000515630748124,"train/samples":320000,"train/lr":2.698342564486797e-05,"train/steps":20000,"_wandb":{"runtime":20858},"_runtime":20858.415510728,"train/dataloading_s":0.0004111651517450809,"train/grad_norm":0.2096482476592064,"train/episodes":2441.816100724914,"train/epochs":122.09080503624571,"train/loss":0.00800122320652008,"_timestamp":1.762438166308964e+09,"_step":20000}
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-core.log CHANGED
@@ -4,3 +4,11 @@
4
  {"time":"2025-11-06T16:21:49.795665893+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
  {"time":"2025-11-06T16:21:49.810358292+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"8gs9of7r","id":"1(@)"}
6
  {"time":"2025-11-06T16:21:50.381980279+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"8gs9of7r","id":"1(@)"}
 
 
 
 
 
 
 
 
 
4
  {"time":"2025-11-06T16:21:49.795665893+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
  {"time":"2025-11-06T16:21:49.810358292+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"8gs9of7r","id":"1(@)"}
6
  {"time":"2025-11-06T16:21:50.381980279+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"8gs9of7r","id":"1(@)"}
7
+ {"time":"2025-11-06T22:09:29.472181794+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
8
+ {"time":"2025-11-06T22:09:29.472334092+08:00","level":"INFO","msg":"server is shutting down"}
9
+ {"time":"2025-11-06T22:09:29.472330091+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2025-11-06T22:09:29.472456362+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
11
+ {"time":"2025-11-06T22:09:29.472561842+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-528220-528551-3638943209/socket","Net":"unix"}}
12
+ {"time":"2025-11-06T22:09:31.654792202+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
13
+ {"time":"2025-11-06T22:09:31.654841186+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
14
+ {"time":"2025-11-06T22:09:31.654868995+08:00","level":"INFO","msg":"server is closed"}
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug-internal.log CHANGED
@@ -7,3 +7,10 @@
7
  {"time":"2025-11-06T18:28:51.825582909+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
8
  {"time":"2025-11-06T19:15:51.940995974+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
9
  {"time":"2025-11-06T20:24:22.8723501+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
 
 
 
 
 
 
 
 
7
  {"time":"2025-11-06T18:28:51.825582909+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
8
  {"time":"2025-11-06T19:15:51.940995974+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
9
  {"time":"2025-11-06T20:24:22.8723501+08:00","level":"INFO","msg":"api: retrying HTTP error","status":500,"url":"https://api.wandb.ai/graphql","body":"{\"errors\":[{\"message\":\"context deadline exceeded\",\"path\":[\"project\",\"run\"]}],\"data\":{\"project\":{\"run\":null}}}"}
10
+ {"time":"2025-11-06T21:50:57.568308958+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/ai-liber/smolvla_omy/8gs9of7r/file_stream\": dial tcp: lookup api.wandb.ai on 222.201.54.123:53: server misbehaving"}
11
+ {"time":"2025-11-06T21:51:00.931430036+08:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/ai-liber/smolvla_omy/8gs9of7r/file_stream\": dial tcp: lookup api.wandb.ai on 222.201.54.123:53: server misbehaving"}
12
+ {"time":"2025-11-06T22:09:29.4723197+08:00","level":"INFO","msg":"stream: closing","id":"8gs9of7r"}
13
+ {"time":"2025-11-06T22:09:30.974870957+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
14
+ {"time":"2025-11-06T22:09:31.653721675+08:00","level":"INFO","msg":"handler: closed","stream_id":"8gs9of7r"}
15
+ {"time":"2025-11-06T22:09:31.653947554+08:00","level":"INFO","msg":"sender: closed","stream_id":"8gs9of7r"}
16
+ {"time":"2025-11-06T22:09:31.653974862+08:00","level":"INFO","msg":"stream: closed","id":"8gs9of7r"}
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/logs/debug.log CHANGED
@@ -19,3 +19,5 @@ config: {'dataset': {'repo_id': 'omy_pnp_language', 'root': './demo_data_languag
19
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2446] Redirects installed.
21
  2025-11-06 16:21:51,416 INFO MainThread:528220 [wandb_init.py:init():1073] run started, returning control to user process
 
 
 
19
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
  2025-11-06 16:21:51,410 INFO MainThread:528220 [wandb_run.py:_redirect():2446] Redirects installed.
21
  2025-11-06 16:21:51,416 INFO MainThread:528220 [wandb_init.py:init():1073] run started, returning control to user process
22
+ 2025-11-06 22:09:29,472 INFO wandb-AsyncioManager-main:528220 [service_client.py:_forward_responses():80] Reached EOF.
23
+ 2025-11-06 22:09:29,472 INFO wandb-AsyncioManager-main:528220 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles.
smolvla_omy/wandb/run-20251106_162149-8gs9of7r/run-8gs9of7r.wandb CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84e401863c9a52b1e0868f3395d571a429a67c767c565ab9b4c9e9d619e3ebc2
3
- size 2752512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1f0139e9e948e37049a676a47e814137ea2dd8a668f98b3517fc13284a90818
3
+ size 3501036