Lakesenberg commited on
Commit
c10ead5
·
verified ·
1 Parent(s): 48b9461

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20251216_174551-tckp70s8/run-tckp70s8.wandb filter=lfs diff=lfs merge=lfs -text
checkpoints/020000/pretrained_model/config.json ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "normalization_mapping": {
5
+ "VISUAL": "IDENTITY",
6
+ "STATE": "MEAN_STD",
7
+ "ACTION": "MEAN_STD"
8
+ },
9
+ "input_features": {
10
+ "observation.images.image": {
11
+ "type": "VISUAL",
12
+ "shape": [
13
+ 3,
14
+ 256,
15
+ 256
16
+ ]
17
+ },
18
+ "observation.images.wrist_image": {
19
+ "type": "VISUAL",
20
+ "shape": [
21
+ 3,
22
+ 256,
23
+ 256
24
+ ]
25
+ },
26
+ "observation.state": {
27
+ "type": "STATE",
28
+ "shape": [
29
+ 8
30
+ ]
31
+ }
32
+ },
33
+ "output_features": {
34
+ "action": {
35
+ "type": "ACTION",
36
+ "shape": [
37
+ 7
38
+ ]
39
+ }
40
+ },
41
+ "device": "cuda",
42
+ "use_amp": false,
43
+ "push_to_hub": true,
44
+ "repo_id": "username/libero_smolvla",
45
+ "private": null,
46
+ "tags": null,
47
+ "license": null,
48
+ "chunk_size": 50,
49
+ "n_action_steps": 50,
50
+ "max_state_dim": 32,
51
+ "max_action_dim": 32,
52
+ "resize_imgs_with_padding": [
53
+ 512,
54
+ 512
55
+ ],
56
+ "empty_cameras": 0,
57
+ "adapt_to_pi_aloha": false,
58
+ "use_delta_joint_actions_aloha": false,
59
+ "tokenizer_max_length": 48,
60
+ "num_steps": 10,
61
+ "use_cache": true,
62
+ "freeze_vision_encoder": true,
63
+ "train_expert_only": true,
64
+ "train_state_proj": true,
65
+ "optimizer_lr": 0.0001,
66
+ "optimizer_betas": [
67
+ 0.9,
68
+ 0.95
69
+ ],
70
+ "optimizer_eps": 1e-08,
71
+ "optimizer_weight_decay": 1e-10,
72
+ "optimizer_grad_clip_norm": 10,
73
+ "scheduler_warmup_steps": 1000,
74
+ "scheduler_decay_steps": 30000,
75
+ "scheduler_decay_lr": 2.5e-06,
76
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
77
+ "load_vlm_weights": true,
78
+ "add_image_special_tokens": false,
79
+ "attention_mode": "cross_attn",
80
+ "prefix_length": -1,
81
+ "pad_language_to": "longest",
82
+ "num_expert_layers": -1,
83
+ "num_vlm_layers": 16,
84
+ "self_attn_every_n_layers": 2,
85
+ "expert_width_multiplier": 0.75,
86
+ "min_period": 0.004,
87
+ "max_period": 4.0
88
+ }
checkpoints/020000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:761fc8cc537084fdbab16784812dd0f2405e86da499b713abe3d6cb40e0ffe9c
3
+ size 906713328
checkpoints/020000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,200 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "Lakesenberg/libero_spatial_image",
4
+ "root": "/data/libero_spatial_image",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ }
61
+ }
62
+ },
63
+ "revision": null,
64
+ "use_imagenet_stats": true,
65
+ "video_backend": "torchcodec"
66
+ },
67
+ "env": null,
68
+ "policy": {
69
+ "type": "smolvla",
70
+ "n_obs_steps": 1,
71
+ "normalization_mapping": {
72
+ "VISUAL": "IDENTITY",
73
+ "STATE": "MEAN_STD",
74
+ "ACTION": "MEAN_STD"
75
+ },
76
+ "input_features": {
77
+ "observation.images.image": {
78
+ "type": "VISUAL",
79
+ "shape": [
80
+ 3,
81
+ 256,
82
+ 256
83
+ ]
84
+ },
85
+ "observation.images.wrist_image": {
86
+ "type": "VISUAL",
87
+ "shape": [
88
+ 3,
89
+ 256,
90
+ 256
91
+ ]
92
+ },
93
+ "observation.state": {
94
+ "type": "STATE",
95
+ "shape": [
96
+ 8
97
+ ]
98
+ }
99
+ },
100
+ "output_features": {
101
+ "action": {
102
+ "type": "ACTION",
103
+ "shape": [
104
+ 7
105
+ ]
106
+ }
107
+ },
108
+ "device": "cuda",
109
+ "use_amp": false,
110
+ "push_to_hub": true,
111
+ "repo_id": "username/libero_smolvla",
112
+ "private": null,
113
+ "tags": null,
114
+ "license": null,
115
+ "chunk_size": 50,
116
+ "n_action_steps": 50,
117
+ "max_state_dim": 32,
118
+ "max_action_dim": 32,
119
+ "resize_imgs_with_padding": [
120
+ 512,
121
+ 512
122
+ ],
123
+ "empty_cameras": 0,
124
+ "adapt_to_pi_aloha": false,
125
+ "use_delta_joint_actions_aloha": false,
126
+ "tokenizer_max_length": 48,
127
+ "num_steps": 10,
128
+ "use_cache": true,
129
+ "freeze_vision_encoder": true,
130
+ "train_expert_only": true,
131
+ "train_state_proj": true,
132
+ "optimizer_lr": 0.0001,
133
+ "optimizer_betas": [
134
+ 0.9,
135
+ 0.95
136
+ ],
137
+ "optimizer_eps": 1e-08,
138
+ "optimizer_weight_decay": 1e-10,
139
+ "optimizer_grad_clip_norm": 10,
140
+ "scheduler_warmup_steps": 1000,
141
+ "scheduler_decay_steps": 30000,
142
+ "scheduler_decay_lr": 2.5e-06,
143
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
144
+ "load_vlm_weights": true,
145
+ "add_image_special_tokens": false,
146
+ "attention_mode": "cross_attn",
147
+ "prefix_length": -1,
148
+ "pad_language_to": "longest",
149
+ "num_expert_layers": -1,
150
+ "num_vlm_layers": 16,
151
+ "self_attn_every_n_layers": 2,
152
+ "expert_width_multiplier": 0.75,
153
+ "min_period": 0.004,
154
+ "max_period": 4.0
155
+ },
156
+ "output_dir": "libero_smolvla_a2c2",
157
+ "job_name": "libero_smolvla_a2c2",
158
+ "resume": false,
159
+ "seed": 1000,
160
+ "num_workers": 4,
161
+ "batch_size": 64,
162
+ "steps": 100000,
163
+ "eval_freq": 20000,
164
+ "log_freq": 200,
165
+ "save_checkpoint": true,
166
+ "save_freq": 20000,
167
+ "use_policy_training_preset": true,
168
+ "optimizer": {
169
+ "type": "adamw",
170
+ "lr": 0.0001,
171
+ "weight_decay": 1e-10,
172
+ "grad_clip_norm": 10,
173
+ "betas": [
174
+ 0.9,
175
+ 0.95
176
+ ],
177
+ "eps": 1e-08
178
+ },
179
+ "scheduler": {
180
+ "type": "cosine_decay_with_warmup",
181
+ "num_warmup_steps": 1000,
182
+ "num_decay_steps": 30000,
183
+ "peak_lr": 0.0001,
184
+ "decay_lr": 2.5e-06
185
+ },
186
+ "eval": {
187
+ "n_episodes": 50,
188
+ "batch_size": 50,
189
+ "use_async_envs": false
190
+ },
191
+ "wandb": {
192
+ "enable": true,
193
+ "disable_artifact": false,
194
+ "project": "lerobot",
195
+ "entity": null,
196
+ "notes": null,
197
+ "run_id": "tckp70s8",
198
+ "mode": null
199
+ }
200
+ }
checkpoints/020000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,527 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.6875000000000013e-05,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "maximize": false,
12
+ "foreach": null,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "decoupled_weight_decay": true,
17
+ "initial_lr": 0.0001,
18
+ "params": [
19
+ 0,
20
+ 1,
21
+ 2,
22
+ 3,
23
+ 4,
24
+ 5,
25
+ 6,
26
+ 7,
27
+ 8,
28
+ 9,
29
+ 10,
30
+ 11,
31
+ 12,
32
+ 13,
33
+ 14,
34
+ 15,
35
+ 16,
36
+ 17,
37
+ 18,
38
+ 19,
39
+ 20,
40
+ 21,
41
+ 22,
42
+ 23,
43
+ 24,
44
+ 25,
45
+ 26,
46
+ 27,
47
+ 28,
48
+ 29,
49
+ 30,
50
+ 31,
51
+ 32,
52
+ 33,
53
+ 34,
54
+ 35,
55
+ 36,
56
+ 37,
57
+ 38,
58
+ 39,
59
+ 40,
60
+ 41,
61
+ 42,
62
+ 43,
63
+ 44,
64
+ 45,
65
+ 46,
66
+ 47,
67
+ 48,
68
+ 49,
69
+ 50,
70
+ 51,
71
+ 52,
72
+ 53,
73
+ 54,
74
+ 55,
75
+ 56,
76
+ 57,
77
+ 58,
78
+ 59,
79
+ 60,
80
+ 61,
81
+ 62,
82
+ 63,
83
+ 64,
84
+ 65,
85
+ 66,
86
+ 67,
87
+ 68,
88
+ 69,
89
+ 70,
90
+ 71,
91
+ 72,
92
+ 73,
93
+ 74,
94
+ 75,
95
+ 76,
96
+ 77,
97
+ 78,
98
+ 79,
99
+ 80,
100
+ 81,
101
+ 82,
102
+ 83,
103
+ 84,
104
+ 85,
105
+ 86,
106
+ 87,
107
+ 88,
108
+ 89,
109
+ 90,
110
+ 91,
111
+ 92,
112
+ 93,
113
+ 94,
114
+ 95,
115
+ 96,
116
+ 97,
117
+ 98,
118
+ 99,
119
+ 100,
120
+ 101,
121
+ 102,
122
+ 103,
123
+ 104,
124
+ 105,
125
+ 106,
126
+ 107,
127
+ 108,
128
+ 109,
129
+ 110,
130
+ 111,
131
+ 112,
132
+ 113,
133
+ 114,
134
+ 115,
135
+ 116,
136
+ 117,
137
+ 118,
138
+ 119,
139
+ 120,
140
+ 121,
141
+ 122,
142
+ 123,
143
+ 124,
144
+ 125,
145
+ 126,
146
+ 127,
147
+ 128,
148
+ 129,
149
+ 130,
150
+ 131,
151
+ 132,
152
+ 133,
153
+ 134,
154
+ 135,
155
+ 136,
156
+ 137,
157
+ 138,
158
+ 139,
159
+ 140,
160
+ 141,
161
+ 142,
162
+ 143,
163
+ 144,
164
+ 145,
165
+ 146,
166
+ 147,
167
+ 148,
168
+ 149,
169
+ 150,
170
+ 151,
171
+ 152,
172
+ 153,
173
+ 154,
174
+ 155,
175
+ 156,
176
+ 157,
177
+ 158,
178
+ 159,
179
+ 160,
180
+ 161,
181
+ 162,
182
+ 163,
183
+ 164,
184
+ 165,
185
+ 166,
186
+ 167,
187
+ 168,
188
+ 169,
189
+ 170,
190
+ 171,
191
+ 172,
192
+ 173,
193
+ 174,
194
+ 175,
195
+ 176,
196
+ 177,
197
+ 178,
198
+ 179,
199
+ 180,
200
+ 181,
201
+ 182,
202
+ 183,
203
+ 184,
204
+ 185,
205
+ 186,
206
+ 187,
207
+ 188,
208
+ 189,
209
+ 190,
210
+ 191,
211
+ 192,
212
+ 193,
213
+ 194,
214
+ 195,
215
+ 196,
216
+ 197,
217
+ 198,
218
+ 199,
219
+ 200,
220
+ 201,
221
+ 202,
222
+ 203,
223
+ 204,
224
+ 205,
225
+ 206,
226
+ 207,
227
+ 208,
228
+ 209,
229
+ 210,
230
+ 211,
231
+ 212,
232
+ 213,
233
+ 214,
234
+ 215,
235
+ 216,
236
+ 217,
237
+ 218,
238
+ 219,
239
+ 220,
240
+ 221,
241
+ 222,
242
+ 223,
243
+ 224,
244
+ 225,
245
+ 226,
246
+ 227,
247
+ 228,
248
+ 229,
249
+ 230,
250
+ 231,
251
+ 232,
252
+ 233,
253
+ 234,
254
+ 235,
255
+ 236,
256
+ 237,
257
+ 238,
258
+ 239,
259
+ 240,
260
+ 241,
261
+ 242,
262
+ 243,
263
+ 244,
264
+ 245,
265
+ 246,
266
+ 247,
267
+ 248,
268
+ 249,
269
+ 250,
270
+ 251,
271
+ 252,
272
+ 253,
273
+ 254,
274
+ 255,
275
+ 256,
276
+ 257,
277
+ 258,
278
+ 259,
279
+ 260,
280
+ 261,
281
+ 262,
282
+ 263,
283
+ 264,
284
+ 265,
285
+ 266,
286
+ 267,
287
+ 268,
288
+ 269,
289
+ 270,
290
+ 271,
291
+ 272,
292
+ 273,
293
+ 274,
294
+ 275,
295
+ 276,
296
+ 277,
297
+ 278,
298
+ 279,
299
+ 280,
300
+ 281,
301
+ 282,
302
+ 283,
303
+ 284,
304
+ 285,
305
+ 286,
306
+ 287,
307
+ 288,
308
+ 289,
309
+ 290,
310
+ 291,
311
+ 292,
312
+ 293,
313
+ 294,
314
+ 295,
315
+ 296,
316
+ 297,
317
+ 298,
318
+ 299,
319
+ 300,
320
+ 301,
321
+ 302,
322
+ 303,
323
+ 304,
324
+ 305,
325
+ 306,
326
+ 307,
327
+ 308,
328
+ 309,
329
+ 310,
330
+ 311,
331
+ 312,
332
+ 313,
333
+ 314,
334
+ 315,
335
+ 316,
336
+ 317,
337
+ 318,
338
+ 319,
339
+ 320,
340
+ 321,
341
+ 322,
342
+ 323,
343
+ 324,
344
+ 325,
345
+ 326,
346
+ 327,
347
+ 328,
348
+ 329,
349
+ 330,
350
+ 331,
351
+ 332,
352
+ 333,
353
+ 334,
354
+ 335,
355
+ 336,
356
+ 337,
357
+ 338,
358
+ 339,
359
+ 340,
360
+ 341,
361
+ 342,
362
+ 343,
363
+ 344,
364
+ 345,
365
+ 346,
366
+ 347,
367
+ 348,
368
+ 349,
369
+ 350,
370
+ 351,
371
+ 352,
372
+ 353,
373
+ 354,
374
+ 355,
375
+ 356,
376
+ 357,
377
+ 358,
378
+ 359,
379
+ 360,
380
+ 361,
381
+ 362,
382
+ 363,
383
+ 364,
384
+ 365,
385
+ 366,
386
+ 367,
387
+ 368,
388
+ 369,
389
+ 370,
390
+ 371,
391
+ 372,
392
+ 373,
393
+ 374,
394
+ 375,
395
+ 376,
396
+ 377,
397
+ 378,
398
+ 379,
399
+ 380,
400
+ 381,
401
+ 382,
402
+ 383,
403
+ 384,
404
+ 385,
405
+ 386,
406
+ 387,
407
+ 388,
408
+ 389,
409
+ 390,
410
+ 391,
411
+ 392,
412
+ 393,
413
+ 394,
414
+ 395,
415
+ 396,
416
+ 397,
417
+ 398,
418
+ 399,
419
+ 400,
420
+ 401,
421
+ 402,
422
+ 403,
423
+ 404,
424
+ 405,
425
+ 406,
426
+ 407,
427
+ 408,
428
+ 409,
429
+ 410,
430
+ 411,
431
+ 412,
432
+ 413,
433
+ 414,
434
+ 415,
435
+ 416,
436
+ 417,
437
+ 418,
438
+ 419,
439
+ 420,
440
+ 421,
441
+ 422,
442
+ 423,
443
+ 424,
444
+ 425,
445
+ 426,
446
+ 427,
447
+ 428,
448
+ 429,
449
+ 430,
450
+ 431,
451
+ 432,
452
+ 433,
453
+ 434,
454
+ 435,
455
+ 436,
456
+ 437,
457
+ 438,
458
+ 439,
459
+ 440,
460
+ 441,
461
+ 442,
462
+ 443,
463
+ 444,
464
+ 445,
465
+ 446,
466
+ 447,
467
+ 448,
468
+ 449,
469
+ 450,
470
+ 451,
471
+ 452,
472
+ 453,
473
+ 454,
474
+ 455,
475
+ 456,
476
+ 457,
477
+ 458,
478
+ 459,
479
+ 460,
480
+ 461,
481
+ 462,
482
+ 463,
483
+ 464,
484
+ 465,
485
+ 466,
486
+ 467,
487
+ 468,
488
+ 469,
489
+ 470,
490
+ 471,
491
+ 472,
492
+ 473,
493
+ 474,
494
+ 475,
495
+ 476,
496
+ 477,
497
+ 478,
498
+ 479,
499
+ 480,
500
+ 481,
501
+ 482,
502
+ 483,
503
+ 484,
504
+ 485,
505
+ 486,
506
+ 487,
507
+ 488,
508
+ 489,
509
+ 490,
510
+ 491,
511
+ 492,
512
+ 493,
513
+ 494,
514
+ 495,
515
+ 496,
516
+ 497,
517
+ 498,
518
+ 499,
519
+ 500,
520
+ 501,
521
+ 502,
522
+ 503,
523
+ 504,
524
+ 505
525
+ ]
526
+ }
527
+ ]
checkpoints/020000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:670eaf7d35bff7eccad5763c8aae80a7fc1b1f13222e830b0caa568b568bd249
3
+ size 412659164
checkpoints/020000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:827033df27498fef6440a808882eb2ba906912452537d63a6a0a004b272fceb4
3
+ size 15708
checkpoints/020000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 20000,
6
+ "_step_count": 20001,
7
+ "_is_initial": false,
8
+ "_get_lr_called_within_step": false,
9
+ "_last_lr": [
10
+ 2.6875000000000013e-05
11
+ ],
12
+ "lr_lambdas": [
13
+ null
14
+ ]
15
+ }
checkpoints/020000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 20000
3
+ }
wandb/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Current SDK version is 0.23.1
2
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Configure stats pid to 2741015
3
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
4
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Loading settings from /code/lq/a2c2-libero/wandb/settings
5
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Loading settings from environment variables
6
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_init.py:setup_run_log_directory():714] Logging user logs to libero_smolvla_a2c2/wandb/run-20251216_174551-tckp70s8/logs/debug.log
7
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_init.py:setup_run_log_directory():715] Logging internal logs to libero_smolvla_a2c2/wandb/run-20251216_174551-tckp70s8/logs/debug-internal.log
8
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_init.py:init():841] calling init triggers
9
+ 2025-12-16 17:45:51,386 INFO MainThread:2741015 [wandb_init.py:init():846] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'Lakesenberg/libero_spatial_image', 'root': '/data/libero_spatial_image', 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec'}, 'env': None, 'policy': {'type': 'smolvla', 'n_obs_steps': 1, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': True, 'repo_id': 'username/libero_smolvla', 'private': None, 'tags': None, 'license': None, 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'resize_imgs_with_padding': [512, 512], 'empty_cameras': 0, 'adapt_to_pi_aloha': False, 'use_delta_joint_actions_aloha': False, 'tokenizer_max_length': 48, 'num_steps': 10, 'use_cache': True, 'freeze_vision_encoder': True, 'train_expert_only': True, 'train_state_proj': True, 'optimizer_lr': 0.0001, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 1e-10, 'optimizer_grad_clip_norm': 10, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'vlm_model_name': 'HuggingFaceTB/SmolVLM2-500M-Video-Instruct', 'load_vlm_weights': True, 'add_image_special_tokens': False, 'attention_mode': 'cross_attn', 'prefix_length': -1, 'pad_language_to': 'longest', 'num_expert_layers': -1, 'num_vlm_layers': 16, 'self_attn_every_n_layers': 2, 'expert_width_multiplier': 0.75, 'min_period': 0.004, 'max_period': 4.0}, 'output_dir': 'libero_smolvla_a2c2', 'job_name': 'libero_smolvla_a2c2', 'resume': False, 'seed': 1000, 'num_workers': 4, 'batch_size': 64, 'steps': 100000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 0.0001, 'weight_decay': 1e-10, 'grad_clip_norm': 10, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 0.0001, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, '_wandb': {}}
11
+ 2025-12-16 17:45:51,386 INFO MainThread:2741015 [wandb_init.py:init():889] starting backend
12
+ 2025-12-16 17:45:51,636 INFO MainThread:2741015 [wandb_init.py:init():892] sending inform_init request
13
+ 2025-12-16 17:45:51,642 INFO MainThread:2741015 [wandb_init.py:init():900] backend started and connected
14
+ 2025-12-16 17:45:51,646 INFO MainThread:2741015 [wandb_init.py:init():970] updated telemetry
15
+ 2025-12-16 17:45:51,654 INFO MainThread:2741015 [wandb_init.py:init():994] communicating run to backend with 90.0 second timeout
16
+ 2025-12-16 17:45:52,908 INFO MainThread:2741015 [wandb_init.py:init():1041] starting run threads in backend
17
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_console_start():2521] atexit reg
18
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_redirect():2369] redirect: wrap_raw
19
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_redirect():2438] Wrapping output streams.
20
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_redirect():2461] Redirects installed.
21
+ 2025-12-16 17:45:53,079 INFO MainThread:2741015 [wandb_init.py:init():1081] run started, returning control to user process
22
+ 2025-12-17 00:37:38,942 INFO wandb-AsyncioManager-main:2741015 [service_client.py:_forward_responses():80] Reached EOF.
23
+ 2025-12-17 00:37:38,943 INFO wandb-AsyncioManager-main:2741015 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles.
wandb/run-20251216_174551-tckp70s8/files/config.yaml ADDED
@@ -0,0 +1,272 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.23.1
4
+ e:
5
+ 5kiukxb0n4qxfbottrvf5zv3vh1llpjw:
6
+ args:
7
+ - --policy.type=smolvla
8
+ - --policy.load_vlm_weights
9
+ - "True"
10
+ - --dataset.repo_id
11
+ - Lakesenberg/libero_spatial_image
12
+ - --dataset.root
13
+ - /data/libero_spatial_image
14
+ - --batch_size=64
15
+ - --steps=100000
16
+ - --policy.repo_id
17
+ - username/libero_smolvla
18
+ - --output_dir
19
+ - libero_smolvla_a2c2
20
+ - --job_name
21
+ - libero_smolvla_a2c2
22
+ - --wandb.enable
23
+ - "true"
24
+ codePath: src/lerobot/scripts/train.py
25
+ codePathLocal: src/lerobot/scripts/train.py
26
+ cpu_count: 128
27
+ cpu_count_logical: 128
28
+ cudaVersion: "12.8"
29
+ disk:
30
+ /:
31
+ total: "2362542452736"
32
+ used: "2217778720768"
33
+ executable: /code/lq/a2c2-libero/.venv/bin/python
34
+ git:
35
+ commit: 54dd088302a0ef3f50c4add3ec927ab94d76a406
36
+ remote: https://github.com/k1000dai/a2c2-libero.git
37
+ gpu: NVIDIA A100-SXM4-80GB
38
+ gpu_count: 8
39
+ gpu_nvidia:
40
+ - architecture: Ampere
41
+ cudaCores: 6912
42
+ memoryTotal: "85899345920"
43
+ name: NVIDIA A100-SXM4-80GB
44
+ uuid: GPU-d36e98cb-a1e9-190c-4f5b-4de233402cf5
45
+ - architecture: Ampere
46
+ cudaCores: 6912
47
+ memoryTotal: "85899345920"
48
+ name: NVIDIA A100-SXM4-80GB
49
+ uuid: GPU-a5ea4ad0-8a8b-34af-3b05-58a0ac4c9fd0
50
+ - architecture: Ampere
51
+ cudaCores: 6912
52
+ memoryTotal: "85899345920"
53
+ name: NVIDIA A100-SXM4-80GB
54
+ uuid: GPU-948a62a9-3401-457e-bba7-dc2781e19b1b
55
+ - architecture: Ampere
56
+ cudaCores: 6912
57
+ memoryTotal: "85899345920"
58
+ name: NVIDIA A100-SXM4-80GB
59
+ uuid: GPU-5314fe46-34bf-c69c-dccc-976f70896178
60
+ - architecture: Ampere
61
+ cudaCores: 6912
62
+ memoryTotal: "85899345920"
63
+ name: NVIDIA A100-SXM4-80GB
64
+ uuid: GPU-59048bb2-8115-237d-2ad3-a1f99b67c906
65
+ - architecture: Ampere
66
+ cudaCores: 6912
67
+ memoryTotal: "85899345920"
68
+ name: NVIDIA A100-SXM4-80GB
69
+ uuid: GPU-a3f470ff-de9b-6067-d889-5e5d5e2a4860
70
+ - architecture: Ampere
71
+ cudaCores: 6912
72
+ memoryTotal: "85899345920"
73
+ name: NVIDIA A100-SXM4-80GB
74
+ uuid: GPU-74a6ef3f-1b1c-5880-8fee-4441da7ac27a
75
+ - architecture: Ampere
76
+ cudaCores: 6912
77
+ memoryTotal: "85899345920"
78
+ name: NVIDIA A100-SXM4-80GB
79
+ uuid: GPU-380c1f88-df1e-2984-c9fc-f424eb38064f
80
+ host: GPU01
81
+ memory:
82
+ total: "1082019954688"
83
+ os: Linux-5.15.0-139-generic-x86_64-with-glibc2.35
84
+ program: /code/lq/a2c2-libero/src/lerobot/scripts/train.py
85
+ python: CPython 3.10.12
86
+ root: libero_smolvla_a2c2
87
+ startedAt: "2025-12-16T09:45:51.384094Z"
88
+ writerId: 5kiukxb0n4qxfbottrvf5zv3vh1llpjw
89
+ m: []
90
+ python_version: 3.10.12
91
+ t:
92
+ "1":
93
+ - 1
94
+ - 41
95
+ - 49
96
+ - 51
97
+ "2":
98
+ - 1
99
+ - 11
100
+ - 41
101
+ - 49
102
+ - 51
103
+ - 71
104
+ "3":
105
+ - 13
106
+ - 15
107
+ - 16
108
+ - 61
109
+ "4": 3.10.12
110
+ "5": 0.23.1
111
+ "10":
112
+ - 21
113
+ "12": 0.23.1
114
+ "13": linux-x86_64
115
+ batch_size:
116
+ value: 64
117
+ dataset:
118
+ value:
119
+ episodes: null
120
+ image_transforms:
121
+ enable: false
122
+ max_num_transforms: 3
123
+ random_order: false
124
+ tfs:
125
+ brightness:
126
+ kwargs:
127
+ brightness:
128
+ - 0.8
129
+ - 1.2
130
+ type: ColorJitter
131
+ weight: 1
132
+ contrast:
133
+ kwargs:
134
+ contrast:
135
+ - 0.8
136
+ - 1.2
137
+ type: ColorJitter
138
+ weight: 1
139
+ hue:
140
+ kwargs:
141
+ hue:
142
+ - -0.05
143
+ - 0.05
144
+ type: ColorJitter
145
+ weight: 1
146
+ saturation:
147
+ kwargs:
148
+ saturation:
149
+ - 0.5
150
+ - 1.5
151
+ type: ColorJitter
152
+ weight: 1
153
+ sharpness:
154
+ kwargs:
155
+ sharpness:
156
+ - 0.5
157
+ - 1.5
158
+ type: SharpnessJitter
159
+ weight: 1
160
+ repo_id: Lakesenberg/libero_spatial_image
161
+ revision: null
162
+ root: /data/libero_spatial_image
163
+ use_imagenet_stats: true
164
+ video_backend: torchcodec
165
+ env:
166
+ value: null
167
+ eval:
168
+ value:
169
+ batch_size: 50
170
+ n_episodes: 50
171
+ use_async_envs: false
172
+ eval_freq:
173
+ value: 20000
174
+ job_name:
175
+ value: libero_smolvla_a2c2
176
+ log_freq:
177
+ value: 200
178
+ num_workers:
179
+ value: 4
180
+ optimizer:
181
+ value:
182
+ betas:
183
+ - 0.9
184
+ - 0.95
185
+ eps: 1e-08
186
+ grad_clip_norm: 10
187
+ lr: 0.0001
188
+ type: adamw
189
+ weight_decay: 1e-10
190
+ output_dir:
191
+ value: libero_smolvla_a2c2
192
+ policy:
193
+ value:
194
+ adapt_to_pi_aloha: false
195
+ add_image_special_tokens: false
196
+ attention_mode: cross_attn
197
+ chunk_size: 50
198
+ device: cuda
199
+ empty_cameras: 0
200
+ expert_width_multiplier: 0.75
201
+ freeze_vision_encoder: true
202
+ license: null
203
+ load_vlm_weights: true
204
+ max_action_dim: 32
205
+ max_period: 4
206
+ max_state_dim: 32
207
+ min_period: 0.004
208
+ n_action_steps: 50
209
+ n_obs_steps: 1
210
+ normalization_mapping:
211
+ ACTION: MEAN_STD
212
+ STATE: MEAN_STD
213
+ VISUAL: IDENTITY
214
+ num_expert_layers: -1
215
+ num_steps: 10
216
+ num_vlm_layers: 16
217
+ optimizer_betas:
218
+ - 0.9
219
+ - 0.95
220
+ optimizer_eps: 1e-08
221
+ optimizer_grad_clip_norm: 10
222
+ optimizer_lr: 0.0001
223
+ optimizer_weight_decay: 1e-10
224
+ pad_language_to: longest
225
+ prefix_length: -1
226
+ private: null
227
+ push_to_hub: true
228
+ repo_id: username/libero_smolvla
229
+ resize_imgs_with_padding:
230
+ - 512
231
+ - 512
232
+ scheduler_decay_lr: 2.5e-06
233
+ scheduler_decay_steps: 30000
234
+ scheduler_warmup_steps: 1000
235
+ self_attn_every_n_layers: 2
236
+ tags: null
237
+ tokenizer_max_length: 48
238
+ train_expert_only: true
239
+ train_state_proj: true
240
+ type: smolvla
241
+ use_amp: false
242
+ use_cache: true
243
+ use_delta_joint_actions_aloha: false
244
+ vlm_model_name: HuggingFaceTB/SmolVLM2-500M-Video-Instruct
245
+ resume:
246
+ value: false
247
+ save_checkpoint:
248
+ value: true
249
+ save_freq:
250
+ value: 20000
251
+ scheduler:
252
+ value:
253
+ decay_lr: 2.5e-06
254
+ num_decay_steps: 30000
255
+ num_warmup_steps: 1000
256
+ peak_lr: 0.0001
257
+ type: cosine_decay_with_warmup
258
+ seed:
259
+ value: 1000
260
+ steps:
261
+ value: 100000
262
+ use_policy_training_preset:
263
+ value: true
264
+ wandb:
265
+ value:
266
+ disable_artifact: false
267
+ enable: true
268
+ entity: null
269
+ mode: null
270
+ notes: null
271
+ project: lerobot
272
+ run_id: null
wandb/run-20251216_174551-tckp70s8/files/output.log ADDED
@@ -0,0 +1,527 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Logs will be synced with wandb.
2
+ INFO 2025-12-16 17:45:53 db_utils.py:103 Track this run --> https://wandb.ai/1143892929-makermods/lerobot/runs/tckp70s8
3
+ INFO 2025-12-16 17:45:53 ts/train.py:127 Creating dataset
4
+ WARNING 2025-12-16 17:45:53 ts/utils.py:302
5
+ The dataset you requested (Lakesenberg/libero_spatial_image) is in 2.0 format.
6
+ While current version of LeRobot is backward-compatible with it, the version of your dataset still uses global
7
+ stats instead of per-episode stats. Update your dataset stats to the new format using this command:
8
+ ```
9
+ python -m lerobot.datasets.v21.convert_dataset_v20_to_v21 --repo-id=Lakesenberg/libero_spatial_image
10
+ ```
11
+
12
+ If you encounter a problem, contact LeRobot maintainers on [Discord](https://discord.com/invite/s3KuuzsPFb)
13
+ or open an [issue on GitHub](https://github.com/huggingface/lerobot/issues/new/choose).
14
+
15
+ WARNING 2025-12-16 17:45:53 ts/utils.py:302
16
+ The dataset you requested (Lakesenberg/libero_spatial_image) is in 2.0 format.
17
+ While current version of LeRobot is backward-compatible with it, the version of your dataset still uses global
18
+ stats instead of per-episode stats. Update your dataset stats to the new format using this command:
19
+ ```
20
+ python -m lerobot.datasets.v21.convert_dataset_v20_to_v21 --repo-id=Lakesenberg/libero_spatial_image
21
+ ```
22
+
23
+ If you encounter a problem, contact LeRobot maintainers on [Discord](https://discord.com/invite/s3KuuzsPFb)
24
+ or open an [issue on GitHub](https://github.com/huggingface/lerobot/issues/new/choose).
25
+
26
+ Resolving data files: 100%|██████████████████████████████████████████████████████████████████████| 432/432 [00:00<00:00, 54142.69it/s]
27
+ INFO 2025-12-16 17:45:55 ts/train.py:138 Creating policy
28
+ Loading HuggingFaceTB/SmolVLM2-500M-Video-Instruct weights ...
29
+ INFO 2025-12-16 17:46:00 modeling.py:987 We will use 90% of the memory on device 0 for storing the model, and 10% for the buffer to avoid OOM. You can set `max_memory` in to a higher value to use more memory (at your own risk).
30
+ Reducing the number of VLM layers to 16 ...
31
+ INFO 2025-12-16 17:46:10 ts/train.py:144 Creating optimizer and scheduler
32
+ INFO 2025-12-16 17:46:10 ts/train.py:156 Output dir: libero_smolvla_a2c2
33
+ INFO 2025-12-16 17:46:10 ts/train.py:159 cfg.steps=100000 (100K)
34
+ INFO 2025-12-16 17:46:10 ts/train.py:160 dataset.num_frames=52970 (53K)
35
+ INFO 2025-12-16 17:46:10 ts/train.py:161 dataset.num_episodes=432
36
+ INFO 2025-12-16 17:46:10 ts/train.py:162 num_learnable_params=99880992 (100M)
37
+ INFO 2025-12-16 17:46:10 ts/train.py:163 num_total_params=450046220 (450M)
38
+ INFO 2025-12-16 17:46:10 ts/train.py:202 Start offline training on a fixed dataset
39
+ INFO 2025-12-16 17:51:39 ts/train.py:232 step:200 smpl:13K ep:104 epch:0.24 loss:1.142 grdn:2.823 lr:1.0e-05 updt_s:1.628 data_s:0.014
40
+ WARNING 2025-12-16 17:51:39 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
41
+ WARNING 2025-12-16 17:51:39 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
42
+ INFO 2025-12-16 17:57:05 ts/train.py:232 step:400 smpl:26K ep:209 epch:0.48 loss:0.446 grdn:2.192 lr:3.0e-05 updt_s:1.624 data_s:0.001
43
+ WARNING 2025-12-16 17:57:05 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
44
+ WARNING 2025-12-16 17:57:05 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
45
+ INFO 2025-12-16 18:02:30 ts/train.py:232 step:600 smpl:38K ep:313 epch:0.72 loss:0.281 grdn:3.079 lr:5.0e-05 updt_s:1.625 data_s:0.001
46
+ WARNING 2025-12-16 18:02:30 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
47
+ WARNING 2025-12-16 18:02:30 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
48
+ INFO 2025-12-16 18:07:57 ts/train.py:232 step:800 smpl:51K ep:418 epch:0.97 loss:0.229 grdn:3.066 lr:7.0e-05 updt_s:1.630 data_s:0.001
49
+ WARNING 2025-12-16 18:07:57 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
50
+ WARNING 2025-12-16 18:07:57 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
51
+ INFO 2025-12-16 18:13:26 ts/train.py:232 step:1K smpl:64K ep:522 epch:1.21 loss:0.209 grdn:2.633 lr:9.0e-05 updt_s:1.631 data_s:0.014
52
+ WARNING 2025-12-16 18:13:26 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
53
+ WARNING 2025-12-16 18:13:26 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
54
+ INFO 2025-12-16 18:18:53 ts/train.py:232 step:1K smpl:77K ep:626 epch:1.45 loss:0.185 grdn:1.838 lr:1.0e-04 updt_s:1.632 data_s:0.001
55
+ WARNING 2025-12-16 18:18:53 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
56
+ WARNING 2025-12-16 18:18:53 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
57
+ INFO 2025-12-16 18:24:20 ts/train.py:232 step:1K smpl:90K ep:731 epch:1.69 loss:0.170 grdn:1.381 lr:1.0e-04 updt_s:1.633 data_s:0.000
58
+ WARNING 2025-12-16 18:24:20 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
59
+ WARNING 2025-12-16 18:24:20 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
60
+ INFO 2025-12-16 18:29:48 ts/train.py:232 step:2K smpl:102K ep:835 epch:1.93 loss:0.159 grdn:1.057 lr:9.9e-05 updt_s:1.633 data_s:0.001
61
+ WARNING 2025-12-16 18:29:48 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
62
+ WARNING 2025-12-16 18:29:48 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
63
+ INFO 2025-12-16 18:35:17 ts/train.py:232 step:2K smpl:115K ep:940 epch:2.17 loss:0.147 grdn:0.864 lr:9.9e-05 updt_s:1.630 data_s:0.013
64
+ WARNING 2025-12-16 18:35:17 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
65
+ WARNING 2025-12-16 18:35:17 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
66
+ INFO 2025-12-16 18:40:44 ts/train.py:232 step:2K smpl:128K ep:1K epch:2.42 loss:0.138 grdn:0.759 lr:9.9e-05 updt_s:1.632 data_s:0.000
67
+ WARNING 2025-12-16 18:40:44 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
68
+ WARNING 2025-12-16 18:40:44 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
69
+ INFO 2025-12-16 18:46:11 ts/train.py:232 step:2K smpl:141K ep:1K epch:2.66 loss:0.133 grdn:0.672 lr:9.9e-05 updt_s:1.632 data_s:0.000
70
+ WARNING 2025-12-16 18:46:11 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
71
+ WARNING 2025-12-16 18:46:11 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
72
+ INFO 2025-12-16 18:51:38 ts/train.py:232 step:2K smpl:154K ep:1K epch:2.90 loss:0.127 grdn:0.589 lr:9.9e-05 updt_s:1.632 data_s:0.000
73
+ WARNING 2025-12-16 18:51:38 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
74
+ WARNING 2025-12-16 18:51:38 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
75
+ INFO 2025-12-16 18:57:07 ts/train.py:232 step:3K smpl:166K ep:1K epch:3.14 loss:0.124 grdn:0.563 lr:9.8e-05 updt_s:1.630 data_s:0.014
76
+ WARNING 2025-12-16 18:57:07 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
77
+ WARNING 2025-12-16 18:57:07 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
78
+ INFO 2025-12-16 19:02:34 ts/train.py:232 step:3K smpl:179K ep:1K epch:3.38 loss:0.120 grdn:0.516 lr:9.8e-05 updt_s:1.633 data_s:0.000
79
+ WARNING 2025-12-16 19:02:34 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
80
+ WARNING 2025-12-16 19:02:34 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
81
+ INFO 2025-12-16 19:08:01 ts/train.py:232 step:3K smpl:192K ep:2K epch:3.62 loss:0.118 grdn:0.487 lr:9.8e-05 updt_s:1.635 data_s:0.000
82
+ WARNING 2025-12-16 19:08:01 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
83
+ WARNING 2025-12-16 19:08:01 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
84
+ INFO 2025-12-16 19:13:29 ts/train.py:232 step:3K smpl:205K ep:2K epch:3.87 loss:0.114 grdn:0.448 lr:9.7e-05 updt_s:1.633 data_s:0.000
85
+ WARNING 2025-12-16 19:13:29 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
86
+ WARNING 2025-12-16 19:13:29 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
87
+ INFO 2025-12-16 19:18:58 ts/train.py:232 step:3K smpl:218K ep:2K epch:4.11 loss:0.112 grdn:0.431 lr:9.7e-05 updt_s:1.630 data_s:0.015
88
+ WARNING 2025-12-16 19:18:58 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
89
+ WARNING 2025-12-16 19:18:58 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
90
+ INFO 2025-12-16 19:24:25 ts/train.py:232 step:4K smpl:230K ep:2K epch:4.35 loss:0.110 grdn:0.427 lr:9.7e-05 updt_s:1.633 data_s:0.001
91
+ WARNING 2025-12-16 19:24:25 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
92
+ WARNING 2025-12-16 19:24:25 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
93
+ INFO 2025-12-16 19:29:52 ts/train.py:232 step:4K smpl:243K ep:2K epch:4.59 loss:0.108 grdn:0.414 lr:9.6e-05 updt_s:1.633 data_s:0.000
94
+ WARNING 2025-12-16 19:29:52 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
95
+ WARNING 2025-12-16 19:29:52 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
96
+ INFO 2025-12-16 19:35:18 ts/train.py:232 step:4K smpl:256K ep:2K epch:4.83 loss:0.106 grdn:0.401 lr:9.6e-05 updt_s:1.627 data_s:0.000
97
+ WARNING 2025-12-16 19:35:18 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
98
+ WARNING 2025-12-16 19:35:18 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
99
+ INFO 2025-12-16 19:40:48 ts/train.py:232 step:4K smpl:269K ep:2K epch:5.07 loss:0.105 grdn:0.381 lr:9.6e-05 updt_s:1.633 data_s:0.013
100
+ WARNING 2025-12-16 19:40:48 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
101
+ WARNING 2025-12-16 19:40:48 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
102
+ INFO 2025-12-16 19:46:15 ts/train.py:232 step:4K smpl:282K ep:2K epch:5.32 loss:0.102 grdn:0.381 lr:9.5e-05 updt_s:1.633 data_s:0.000
103
+ WARNING 2025-12-16 19:46:15 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
104
+ WARNING 2025-12-16 19:46:15 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
105
+ INFO 2025-12-16 19:51:42 ts/train.py:232 step:5K smpl:294K ep:2K epch:5.56 loss:0.103 grdn:0.383 lr:9.5e-05 updt_s:1.635 data_s:0.000
106
+ WARNING 2025-12-16 19:51:42 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
107
+ WARNING 2025-12-16 19:51:42 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
108
+ INFO 2025-12-16 19:57:10 ts/train.py:232 step:5K smpl:307K ep:3K epch:5.80 loss:0.100 grdn:0.354 lr:9.4e-05 updt_s:1.633 data_s:0.001
109
+ WARNING 2025-12-16 19:57:10 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
110
+ WARNING 2025-12-16 19:57:10 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
111
+ INFO 2025-12-16 20:02:39 ts/train.py:232 step:5K smpl:320K ep:3K epch:6.04 loss:0.099 grdn:0.365 lr:9.4e-05 updt_s:1.630 data_s:0.014
112
+ WARNING 2025-12-16 20:02:39 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
113
+ WARNING 2025-12-16 20:02:39 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
114
+ INFO 2025-12-16 20:08:06 ts/train.py:232 step:5K smpl:333K ep:3K epch:6.28 loss:0.096 grdn:0.350 lr:9.3e-05 updt_s:1.634 data_s:0.000
115
+ WARNING 2025-12-16 20:08:06 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
116
+ WARNING 2025-12-16 20:08:06 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
117
+ INFO 2025-12-16 20:13:34 ts/train.py:232 step:5K smpl:346K ep:3K epch:6.52 loss:0.095 grdn:0.352 lr:9.3e-05 updt_s:1.636 data_s:0.000
118
+ WARNING 2025-12-16 20:13:34 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
119
+ WARNING 2025-12-16 20:13:34 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
120
+ INFO 2025-12-16 20:19:01 ts/train.py:232 step:6K smpl:358K ep:3K epch:6.77 loss:0.093 grdn:0.342 lr:9.2e-05 updt_s:1.633 data_s:0.000
121
+ WARNING 2025-12-16 20:19:01 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
122
+ WARNING 2025-12-16 20:19:01 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
123
+ INFO 2025-12-16 20:24:31 ts/train.py:232 step:6K smpl:371K ep:3K epch:7.01 loss:0.092 grdn:0.329 lr:9.2e-05 updt_s:1.633 data_s:0.014
124
+ WARNING 2025-12-16 20:24:31 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
125
+ WARNING 2025-12-16 20:24:31 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
126
+ INFO 2025-12-16 20:29:58 ts/train.py:232 step:6K smpl:384K ep:3K epch:7.25 loss:0.089 grdn:0.323 lr:9.1e-05 updt_s:1.633 data_s:0.000
127
+ WARNING 2025-12-16 20:29:58 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
128
+ WARNING 2025-12-16 20:29:58 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
129
+ INFO 2025-12-16 20:35:25 ts/train.py:232 step:6K smpl:397K ep:3K epch:7.49 loss:0.088 grdn:0.315 lr:9.0e-05 updt_s:1.633 data_s:0.000
130
+ WARNING 2025-12-16 20:35:25 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
131
+ WARNING 2025-12-16 20:35:25 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
132
+ INFO 2025-12-16 20:40:52 ts/train.py:232 step:6K smpl:410K ep:3K epch:7.73 loss:0.086 grdn:0.331 lr:9.0e-05 updt_s:1.633 data_s:0.000
133
+ WARNING 2025-12-16 20:40:52 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
134
+ WARNING 2025-12-16 20:40:52 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
135
+ INFO 2025-12-16 20:46:20 ts/train.py:232 step:7K smpl:422K ep:3K epch:7.97 loss:0.086 grdn:0.328 lr:8.9e-05 updt_s:1.635 data_s:0.000
136
+ WARNING 2025-12-16 20:46:20 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
137
+ WARNING 2025-12-16 20:46:20 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
138
+ INFO 2025-12-16 20:51:49 ts/train.py:232 step:7K smpl:435K ep:4K epch:8.22 loss:0.082 grdn:0.321 lr:8.8e-05 updt_s:1.631 data_s:0.014
139
+ WARNING 2025-12-16 20:51:49 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
140
+ WARNING 2025-12-16 20:51:49 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
141
+ INFO 2025-12-16 20:57:16 ts/train.py:232 step:7K smpl:448K ep:4K epch:8.46 loss:0.081 grdn:0.314 lr:8.8e-05 updt_s:1.634 data_s:0.000
142
+ WARNING 2025-12-16 20:57:16 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
143
+ WARNING 2025-12-16 20:57:16 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
144
+ INFO 2025-12-16 21:02:43 ts/train.py:232 step:7K smpl:461K ep:4K epch:8.70 loss:0.080 grdn:0.312 lr:8.7e-05 updt_s:1.634 data_s:0.000
145
+ WARNING 2025-12-16 21:02:43 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
146
+ WARNING 2025-12-16 21:02:43 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
147
+ INFO 2025-12-16 21:08:11 ts/train.py:232 step:7K smpl:474K ep:4K epch:8.94 loss:0.079 grdn:0.311 lr:8.6e-05 updt_s:1.633 data_s:0.001
148
+ WARNING 2025-12-16 21:08:11 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
149
+ WARNING 2025-12-16 21:08:11 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
150
+ INFO 2025-12-16 21:13:40 ts/train.py:232 step:8K smpl:486K ep:4K epch:9.18 loss:0.076 grdn:0.305 lr:8.6e-05 updt_s:1.630 data_s:0.014
151
+ WARNING 2025-12-16 21:13:40 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
152
+ WARNING 2025-12-16 21:13:40 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
153
+ INFO 2025-12-16 21:19:07 ts/train.py:232 step:8K smpl:499K ep:4K epch:9.42 loss:0.074 grdn:0.306 lr:8.5e-05 updt_s:1.633 data_s:0.000
154
+ WARNING 2025-12-16 21:19:07 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
155
+ WARNING 2025-12-16 21:19:07 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
156
+ INFO 2025-12-16 21:24:34 ts/train.py:232 step:8K smpl:512K ep:4K epch:9.67 loss:0.073 grdn:0.295 lr:8.4e-05 updt_s:1.633 data_s:0.000
157
+ WARNING 2025-12-16 21:24:34 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
158
+ WARNING 2025-12-16 21:24:34 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
159
+ INFO 2025-12-16 21:30:01 ts/train.py:232 step:8K smpl:525K ep:4K epch:9.91 loss:0.072 grdn:0.302 lr:8.3e-05 updt_s:1.632 data_s:0.000
160
+ WARNING 2025-12-16 21:30:01 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
161
+ WARNING 2025-12-16 21:30:01 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
162
+ INFO 2025-12-16 21:35:30 ts/train.py:232 step:8K smpl:538K ep:4K epch:10.15 loss:0.070 grdn:0.295 lr:8.3e-05 updt_s:1.631 data_s:0.013
163
+ WARNING 2025-12-16 21:35:30 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
164
+ WARNING 2025-12-16 21:35:30 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
165
+ INFO 2025-12-16 21:40:57 ts/train.py:232 step:9K smpl:550K ep:4K epch:10.39 loss:0.068 grdn:0.294 lr:8.2e-05 updt_s:1.632 data_s:0.000
166
+ WARNING 2025-12-16 21:40:57 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
167
+ WARNING 2025-12-16 21:40:57 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
168
+ INFO 2025-12-16 21:46:25 ts/train.py:232 step:9K smpl:563K ep:5K epch:10.63 loss:0.066 grdn:0.286 lr:8.1e-05 updt_s:1.633 data_s:0.001
169
+ WARNING 2025-12-16 21:46:25 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
170
+ WARNING 2025-12-16 21:46:25 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
171
+ INFO 2025-12-16 21:51:51 ts/train.py:232 step:9K smpl:576K ep:5K epch:10.87 loss:0.065 grdn:0.290 lr:8.0e-05 updt_s:1.631 data_s:0.000
172
+ WARNING 2025-12-16 21:51:51 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
173
+ WARNING 2025-12-16 21:51:51 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
174
+ INFO 2025-12-16 21:57:21 ts/train.py:232 step:9K smpl:589K ep:5K epch:11.12 loss:0.063 grdn:0.285 lr:7.9e-05 updt_s:1.632 data_s:0.014
175
+ WARNING 2025-12-16 21:57:21 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
176
+ WARNING 2025-12-16 21:57:21 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
177
+ INFO 2025-12-16 22:02:48 ts/train.py:232 step:9K smpl:602K ep:5K epch:11.36 loss:0.061 grdn:0.288 lr:7.9e-05 updt_s:1.632 data_s:0.000
178
+ WARNING 2025-12-16 22:02:48 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
179
+ WARNING 2025-12-16 22:02:48 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
180
+ INFO 2025-12-16 22:08:15 ts/train.py:232 step:10K smpl:614K ep:5K epch:11.60 loss:0.059 grdn:0.276 lr:7.8e-05 updt_s:1.632 data_s:0.000
181
+ WARNING 2025-12-16 22:08:15 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
182
+ WARNING 2025-12-16 22:08:15 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
183
+ INFO 2025-12-16 22:13:42 ts/train.py:232 step:10K smpl:627K ep:5K epch:11.84 loss:0.059 grdn:0.280 lr:7.7e-05 updt_s:1.632 data_s:0.000
184
+ WARNING 2025-12-16 22:13:42 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
185
+ WARNING 2025-12-16 22:13:42 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
186
+ INFO 2025-12-16 22:19:11 ts/train.py:232 step:10K smpl:640K ep:5K epch:12.08 loss:0.057 grdn:0.277 lr:7.6e-05 updt_s:1.630 data_s:0.015
187
+ WARNING 2025-12-16 22:19:11 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
188
+ WARNING 2025-12-16 22:19:11 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
189
+ INFO 2025-12-16 22:24:39 ts/train.py:232 step:10K smpl:653K ep:5K epch:12.32 loss:0.055 grdn:0.270 lr:7.5e-05 updt_s:1.634 data_s:0.001
190
+ WARNING 2025-12-16 22:24:39 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
191
+ WARNING 2025-12-16 22:24:39 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
192
+ INFO 2025-12-16 22:30:06 ts/train.py:232 step:10K smpl:666K ep:5K epch:12.57 loss:0.054 grdn:0.270 lr:7.4e-05 updt_s:1.633 data_s:0.000
193
+ WARNING 2025-12-16 22:30:06 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
194
+ WARNING 2025-12-16 22:30:06 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
195
+ INFO 2025-12-16 22:35:33 ts/train.py:232 step:11K smpl:678K ep:6K epch:12.81 loss:0.053 grdn:0.267 lr:7.3e-05 updt_s:1.632 data_s:0.000
196
+ WARNING 2025-12-16 22:35:33 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
197
+ WARNING 2025-12-16 22:35:33 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
198
+ INFO 2025-12-16 22:41:02 ts/train.py:232 step:11K smpl:691K ep:6K epch:13.05 loss:0.052 grdn:0.272 lr:7.2e-05 updt_s:1.631 data_s:0.013
199
+ WARNING 2025-12-16 22:41:02 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
200
+ WARNING 2025-12-16 22:41:02 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
201
+ INFO 2025-12-16 22:46:29 ts/train.py:232 step:11K smpl:704K ep:6K epch:13.29 loss:0.051 grdn:0.268 lr:7.2e-05 updt_s:1.632 data_s:0.000
202
+ WARNING 2025-12-16 22:46:29 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
203
+ WARNING 2025-12-16 22:46:29 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
204
+ INFO 2025-12-16 22:51:56 ts/train.py:232 step:11K smpl:717K ep:6K epch:13.53 loss:0.049 grdn:0.263 lr:7.1e-05 updt_s:1.633 data_s:0.000
205
+ WARNING 2025-12-16 22:51:56 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
206
+ WARNING 2025-12-16 22:51:56 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
207
+ INFO 2025-12-16 22:57:23 ts/train.py:232 step:11K smpl:730K ep:6K epch:13.77 loss:0.048 grdn:0.257 lr:7.0e-05 updt_s:1.632 data_s:0.000
208
+ WARNING 2025-12-16 22:57:23 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
209
+ WARNING 2025-12-16 22:57:23 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
210
+ INFO 2025-12-16 23:02:53 ts/train.py:232 step:12K smpl:742K ep:6K epch:14.02 loss:0.047 grdn:0.256 lr:6.9e-05 updt_s:1.632 data_s:0.015
211
+ WARNING 2025-12-16 23:02:53 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
212
+ WARNING 2025-12-16 23:02:53 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
213
+ INFO 2025-12-16 23:08:20 ts/train.py:232 step:12K smpl:755K ep:6K epch:14.26 loss:0.046 grdn:0.252 lr:6.8e-05 updt_s:1.633 data_s:0.000
214
+ WARNING 2025-12-16 23:08:20 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
215
+ WARNING 2025-12-16 23:08:20 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
216
+ INFO 2025-12-16 23:13:47 ts/train.py:232 step:12K smpl:768K ep:6K epch:14.50 loss:0.046 grdn:0.256 lr:6.7e-05 updt_s:1.632 data_s:0.000
217
+ WARNING 2025-12-16 23:13:47 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
218
+ WARNING 2025-12-16 23:13:47 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
219
+ INFO 2025-12-16 23:19:14 ts/train.py:232 step:12K smpl:781K ep:6K epch:14.74 loss:0.044 grdn:0.255 lr:6.6e-05 updt_s:1.632 data_s:0.001
220
+ WARNING 2025-12-16 23:19:14 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
221
+ WARNING 2025-12-16 23:19:14 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
222
+ INFO 2025-12-16 23:24:28 ts/train.py:232 step:12K smpl:794K ep:6K epch:14.98 loss:0.044 grdn:0.253 lr:6.5e-05 updt_s:1.566 data_s:0.000
223
+ WARNING 2025-12-16 23:24:28 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
224
+ WARNING 2025-12-16 23:24:28 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
225
+ INFO 2025-12-16 23:26:25 ts/train.py:232 step:13K smpl:806K ep:7K epch:15.22 loss:0.042 grdn:0.255 lr:6.4e-05 updt_s:0.570 data_s:0.014
226
+ WARNING 2025-12-16 23:26:25 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
227
+ WARNING 2025-12-16 23:26:25 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
228
+ INFO 2025-12-16 23:28:19 ts/train.py:232 step:13K smpl:819K ep:7K epch:15.47 loss:0.042 grdn:0.248 lr:6.3e-05 updt_s:0.569 data_s:0.001
229
+ WARNING 2025-12-16 23:28:19 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
230
+ WARNING 2025-12-16 23:28:19 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
231
+ INFO 2025-12-16 23:30:14 ts/train.py:232 step:13K smpl:832K ep:7K epch:15.71 loss:0.041 grdn:0.248 lr:6.2e-05 updt_s:0.570 data_s:0.000
232
+ WARNING 2025-12-16 23:30:14 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
233
+ WARNING 2025-12-16 23:30:14 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
234
+ INFO 2025-12-16 23:32:09 ts/train.py:232 step:13K smpl:845K ep:7K epch:15.95 loss:0.041 grdn:0.247 lr:6.1e-05 updt_s:0.573 data_s:0.001
235
+ WARNING 2025-12-16 23:32:09 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
236
+ WARNING 2025-12-16 23:32:09 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
237
+ INFO 2025-12-16 23:34:06 ts/train.py:232 step:13K smpl:858K ep:7K epch:16.19 loss:0.039 grdn:0.240 lr:6.0e-05 updt_s:0.569 data_s:0.014
238
+ WARNING 2025-12-16 23:34:06 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
239
+ WARNING 2025-12-16 23:34:06 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
240
+ INFO 2025-12-16 23:36:01 ts/train.py:232 step:14K smpl:870K ep:7K epch:16.43 loss:0.039 grdn:0.243 lr:5.9e-05 updt_s:0.570 data_s:0.000
241
+ WARNING 2025-12-16 23:36:01 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
242
+ WARNING 2025-12-16 23:36:01 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
243
+ INFO 2025-12-16 23:37:54 ts/train.py:232 step:14K smpl:883K ep:7K epch:16.67 loss:0.039 grdn:0.247 lr:5.8e-05 updt_s:0.563 data_s:0.000
244
+ WARNING 2025-12-16 23:37:54 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
245
+ WARNING 2025-12-16 23:37:54 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
246
+ INFO 2025-12-16 23:39:49 ts/train.py:232 step:14K smpl:896K ep:7K epch:16.92 loss:0.038 grdn:0.246 lr:5.7e-05 updt_s:0.571 data_s:0.001
247
+ WARNING 2025-12-16 23:39:49 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
248
+ WARNING 2025-12-16 23:39:49 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
249
+ INFO 2025-12-16 23:41:46 ts/train.py:232 step:14K smpl:909K ep:7K epch:17.16 loss:0.037 grdn:0.239 lr:5.6e-05 updt_s:0.568 data_s:0.013
250
+ WARNING 2025-12-16 23:41:46 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
251
+ WARNING 2025-12-16 23:41:46 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
252
+ INFO 2025-12-16 23:43:40 ts/train.py:232 step:14K smpl:922K ep:8K epch:17.40 loss:0.037 grdn:0.237 lr:5.5e-05 updt_s:0.568 data_s:0.000
253
+ WARNING 2025-12-16 23:43:40 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
254
+ WARNING 2025-12-16 23:43:40 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
255
+ INFO 2025-12-16 23:45:35 ts/train.py:232 step:15K smpl:934K ep:8K epch:17.64 loss:0.036 grdn:0.237 lr:5.4e-05 updt_s:0.571 data_s:0.000
256
+ WARNING 2025-12-16 23:45:35 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
257
+ WARNING 2025-12-16 23:45:35 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
258
+ INFO 2025-12-16 23:47:28 ts/train.py:232 step:15K smpl:947K ep:8K epch:17.88 loss:0.036 grdn:0.238 lr:5.3e-05 updt_s:0.566 data_s:0.001
259
+ WARNING 2025-12-16 23:47:28 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
260
+ WARNING 2025-12-16 23:47:28 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
261
+ INFO 2025-12-16 23:49:25 ts/train.py:232 step:15K smpl:960K ep:8K epch:18.12 loss:0.035 grdn:0.240 lr:5.2e-05 updt_s:0.567 data_s:0.014
262
+ WARNING 2025-12-16 23:49:25 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
263
+ WARNING 2025-12-16 23:49:25 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
264
+ INFO 2025-12-16 23:51:20 ts/train.py:232 step:15K smpl:973K ep:8K epch:18.37 loss:0.035 grdn:0.237 lr:5.1e-05 updt_s:0.570 data_s:0.001
265
+ WARNING 2025-12-16 23:51:20 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
266
+ WARNING 2025-12-16 23:51:20 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
267
+ INFO 2025-12-16 23:53:14 ts/train.py:232 step:15K smpl:986K ep:8K epch:18.61 loss:0.034 grdn:0.233 lr:5.0e-05 updt_s:0.567 data_s:0.000
268
+ WARNING 2025-12-16 23:53:14 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
269
+ WARNING 2025-12-16 23:53:14 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
270
+ INFO 2025-12-16 23:55:08 ts/train.py:232 step:16K smpl:998K ep:8K epch:18.85 loss:0.034 grdn:0.232 lr:4.9e-05 updt_s:0.566 data_s:0.000
271
+ WARNING 2025-12-16 23:55:08 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
272
+ WARNING 2025-12-16 23:55:08 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
273
+ INFO 2025-12-16 23:57:04 ts/train.py:232 step:16K smpl:1M ep:8K epch:19.09 loss:0.034 grdn:0.233 lr:4.8e-05 updt_s:0.567 data_s:0.014
274
+ WARNING 2025-12-16 23:57:04 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
275
+ WARNING 2025-12-16 23:57:04 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
276
+ INFO 2025-12-16 23:58:58 ts/train.py:232 step:16K smpl:1M ep:8K epch:19.33 loss:0.033 grdn:0.227 lr:4.7e-05 updt_s:0.568 data_s:0.000
277
+ WARNING 2025-12-16 23:58:58 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
278
+ WARNING 2025-12-16 23:58:58 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
279
+ INFO 2025-12-17 00:00:53 ts/train.py:232 step:16K smpl:1M ep:8K epch:19.57 loss:0.033 grdn:0.230 lr:4.6e-05 updt_s:0.572 data_s:0.001
280
+ WARNING 2025-12-17 00:00:53 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
281
+ WARNING 2025-12-17 00:00:53 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
282
+ INFO 2025-12-17 00:02:48 ts/train.py:232 step:16K smpl:1M ep:9K epch:19.81 loss:0.032 grdn:0.224 lr:4.5e-05 updt_s:0.569 data_s:0.001
283
+ WARNING 2025-12-17 00:02:48 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
284
+ WARNING 2025-12-17 00:02:48 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
285
+ INFO 2025-12-17 00:04:45 ts/train.py:232 step:17K smpl:1M ep:9K epch:20.06 loss:0.032 grdn:0.226 lr:4.4e-05 updt_s:0.570 data_s:0.013
286
+ WARNING 2025-12-17 00:04:45 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
287
+ WARNING 2025-12-17 00:04:45 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
288
+ INFO 2025-12-17 00:06:39 ts/train.py:232 step:17K smpl:1M ep:9K epch:20.30 loss:0.032 grdn:0.230 lr:4.3e-05 updt_s:0.566 data_s:0.000
289
+ WARNING 2025-12-17 00:06:39 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
290
+ WARNING 2025-12-17 00:06:39 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
291
+ INFO 2025-12-17 00:08:32 ts/train.py:232 step:17K smpl:1M ep:9K epch:20.54 loss:0.031 grdn:0.221 lr:4.2e-05 updt_s:0.565 data_s:0.000
292
+ WARNING 2025-12-17 00:08:32 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
293
+ WARNING 2025-12-17 00:08:32 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
294
+ INFO 2025-12-17 00:10:24 ts/train.py:232 step:17K smpl:1M ep:9K epch:20.78 loss:0.031 grdn:0.223 lr:4.1e-05 updt_s:0.556 data_s:0.000
295
+ WARNING 2025-12-17 00:10:24 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
296
+ WARNING 2025-12-17 00:10:24 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
297
+ INFO 2025-12-17 00:12:19 ts/train.py:232 step:17K smpl:1M ep:9K epch:21.02 loss:0.031 grdn:0.225 lr:4.0e-05 updt_s:0.557 data_s:0.014
298
+ WARNING 2025-12-17 00:12:19 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
299
+ WARNING 2025-12-17 00:12:19 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
300
+ INFO 2025-12-17 00:14:12 ts/train.py:232 step:18K smpl:1M ep:9K epch:21.26 loss:0.031 grdn:0.219 lr:3.9e-05 updt_s:0.560 data_s:0.000
301
+ WARNING 2025-12-17 00:14:12 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
302
+ WARNING 2025-12-17 00:14:12 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
303
+ INFO 2025-12-17 00:16:06 ts/train.py:232 step:18K smpl:1M ep:9K epch:21.51 loss:0.031 grdn:0.226 lr:3.8e-05 updt_s:0.567 data_s:0.001
304
+ WARNING 2025-12-17 00:16:06 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
305
+ WARNING 2025-12-17 00:16:06 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
306
+ INFO 2025-12-17 00:18:00 ts/train.py:232 step:18K smpl:1M ep:9K epch:21.75 loss:0.030 grdn:0.217 lr:3.7e-05 updt_s:0.568 data_s:0.000
307
+ WARNING 2025-12-17 00:18:00 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
308
+ WARNING 2025-12-17 00:18:00 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
309
+ INFO 2025-12-17 00:19:54 ts/train.py:232 step:18K smpl:1M ep:9K epch:21.99 loss:0.030 grdn:0.221 lr:3.6e-05 updt_s:0.566 data_s:0.000
310
+ WARNING 2025-12-17 00:19:54 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
311
+ WARNING 2025-12-17 00:19:54 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
312
+ INFO 2025-12-17 00:21:51 ts/train.py:232 step:18K smpl:1M ep:10K epch:22.23 loss:0.030 grdn:0.216 lr:3.5e-05 updt_s:0.568 data_s:0.014
313
+ WARNING 2025-12-17 00:21:51 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
314
+ WARNING 2025-12-17 00:21:51 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
315
+ INFO 2025-12-17 00:23:46 ts/train.py:232 step:19K smpl:1M ep:10K epch:22.47 loss:0.030 grdn:0.220 lr:3.4e-05 updt_s:0.569 data_s:0.001
316
+ WARNING 2025-12-17 00:23:46 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
317
+ WARNING 2025-12-17 00:23:46 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
318
+ INFO 2025-12-17 00:25:40 ts/train.py:232 step:19K smpl:1M ep:10K epch:22.71 loss:0.030 grdn:0.216 lr:3.3e-05 updt_s:0.570 data_s:0.001
319
+ WARNING 2025-12-17 00:25:40 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
320
+ WARNING 2025-12-17 00:25:40 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
321
+ INFO 2025-12-17 00:27:36 ts/train.py:232 step:19K smpl:1M ep:10K epch:22.96 loss:0.029 grdn:0.216 lr:3.2e-05 updt_s:0.573 data_s:0.001
322
+ WARNING 2025-12-17 00:27:36 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
323
+ WARNING 2025-12-17 00:27:36 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
324
+ INFO 2025-12-17 00:29:34 ts/train.py:232 step:19K smpl:1M ep:10K epch:23.20 loss:0.029 grdn:0.214 lr:3.1e-05 updt_s:0.572 data_s:0.014
325
+ WARNING 2025-12-17 00:29:34 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
326
+ WARNING 2025-12-17 00:29:34 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
327
+ INFO 2025-12-17 00:31:28 ts/train.py:232 step:19K smpl:1M ep:10K epch:23.44 loss:0.029 grdn:0.212 lr:3.0e-05 updt_s:0.570 data_s:0.001
328
+ WARNING 2025-12-17 00:31:28 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
329
+ WARNING 2025-12-17 00:31:28 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
330
+ INFO 2025-12-17 00:33:23 ts/train.py:232 step:20K smpl:1M ep:10K epch:23.68 loss:0.028 grdn:0.207 lr:2.9e-05 updt_s:0.570 data_s:0.000
331
+ WARNING 2025-12-17 00:33:23 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
332
+ WARNING 2025-12-17 00:33:23 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
333
+ INFO 2025-12-17 00:35:16 ts/train.py:232 step:20K smpl:1M ep:10K epch:23.92 loss:0.028 grdn:0.208 lr:2.8e-05 updt_s:0.562 data_s:0.000
334
+ WARNING 2025-12-17 00:35:16 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
335
+ WARNING 2025-12-17 00:35:16 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
336
+ INFO 2025-12-17 00:37:12 ts/train.py:232 step:20K smpl:1M ep:10K epch:24.16 loss:0.028 grdn:0.208 lr:2.7e-05 updt_s:0.562 data_s:0.014
337
+ WARNING 2025-12-17 00:37:12 db_utils.py:141 WandB logging of key "losses_after_forward" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
338
+ WARNING 2025-12-17 00:37:12 db_utils.py:141 WandB logging of key "losses_after_rm_padding" was ignored as its type "<class 'torch.Tensor'>" is not handled by this wrapper.
339
+ INFO 2025-12-17 00:37:12 ts/train.py:241 Checkpoint policy after step 20000
340
+ Traceback (most recent call last):
341
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connection.py", line 204, in _new_conn
342
+ sock = connection.create_connection(
343
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/util/connection.py", line 60, in create_connection
344
+ for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
345
+ File "/usr/lib/python3.10/socket.py", line 955, in getaddrinfo
346
+ for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
347
+ socket.gaierror: [Errno -3] Temporary failure in name resolution
348
+
349
+ The above exception was the direct cause of the following exception:
350
+
351
+ Traceback (most recent call last):
352
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 787, in urlopen
353
+ response = self._make_request(
354
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 488, in _make_request
355
+ raise new_e
356
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 464, in _make_request
357
+ self._validate_conn(conn)
358
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn
359
+ conn.connect()
360
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connection.py", line 759, in connect
361
+ self.sock = sock = self._new_conn()
362
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connection.py", line 211, in _new_conn
363
+ raise NameResolutionError(self.host, self, e) from e
364
+ urllib3.exceptions.NameResolutionError: HTTPSConnection(host='api.wandb.ai', port=443): Failed to resolve 'api.wandb.ai' ([Errno -3] Temporary failure in name resolution)
365
+
366
+ The above exception was the direct cause of the following exception:
367
+
368
+ Traceback (most recent call last):
369
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/adapters.py", line 644, in send
370
+ resp = conn.urlopen(
371
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 841, in urlopen
372
+ retries = retries.increment(
373
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/util/retry.py", line 519, in increment
374
+ raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type]
375
+ urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='api.wandb.ai', port=443): Max retries exceeded with url: /graphql (Caused by NameResolutionError("HTTPSConnection(host='api.wandb.ai', port=443): Failed to resolve 'api.wandb.ai' ([Errno -3] Temporary failure in name resolution)"))
376
+
377
+ During handling of the above exception, another exception occurred:
378
+
379
+ Traceback (most recent call last):
380
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 387, in _verify_login
381
+ is_api_key_valid = api.validate_api_key()
382
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/apis/internal.py", line 66, in validate_api_key
383
+ return self.api.validate_api_key()
384
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/internal/internal_api.py", line 402, in validate_api_key
385
+ res = self.execute(gql("query { viewer { id } }"))
386
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/internal/internal_api.py", line 391, in execute
387
+ return self.client.execute(*args, **kwargs) # type: ignore
388
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/vendor/gql-0.2.0/wandb_gql/client.py", line 52, in execute
389
+ result = self._get_result(document, *args, **kwargs)
390
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/vendor/gql-0.2.0/wandb_gql/client.py", line 60, in _get_result
391
+ return self.transport.execute(document, *args, **kwargs)
392
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/lib/gql_request.py", line 70, in execute
393
+ request = self.session.post(self.url, **post_args)
394
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/sessions.py", line 637, in post
395
+ return self.request("POST", url, data=data, json=json, **kwargs)
396
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/sessions.py", line 589, in request
397
+ resp = self.send(prep, **send_kwargs)
398
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/sessions.py", line 703, in send
399
+ r = adapter.send(request, **kwargs)
400
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/adapters.py", line 677, in send
401
+ raise ConnectionError(e, request=request)
402
+ requests.exceptions.ConnectionError: HTTPSConnectionPool(host='api.wandb.ai', port=443): Max retries exceeded with url: /graphql (Caused by NameResolutionError("HTTPSConnection(host='api.wandb.ai', port=443): Failed to resolve 'api.wandb.ai' ([Errno -3] Temporary failure in name resolution)"))
403
+
404
+ The above exception was the direct cause of the following exception:
405
+
406
+ Traceback (most recent call last):
407
+ File "/code/lq/a2c2-libero/src/lerobot/scripts/train.py", line 291, in <module>
408
+ train()
409
+ File "/code/lq/a2c2-libero/src/lerobot/configs/parser.py", line 225, in wrapper_inner
410
+ response = fn(cfg, *args, **kwargs)
411
+ File "/code/lq/a2c2-libero/src/lerobot/scripts/train.py", line 246, in train
412
+ wandb_logger.log_policy(checkpoint_dir)
413
+ File "/code/lq/a2c2-libero/src/lerobot/utils/wandb_utils.py", line 116, in log_policy
414
+ self._wandb.log_artifact(artifact)
415
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 400, in wrapper
416
+ return func(self, *args, **kwargs)
417
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 458, in wrapper_fn
418
+ return func(self, *args, **kwargs)
419
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 445, in wrapper
420
+ return func(self, *args, **kwargs)
421
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3166, in log_artifact
422
+ return self._log_artifact(
423
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3311, in _log_artifact
424
+ self._assert_can_log_artifact(artifact)
425
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3371, in _assert_can_log_artifact
426
+ public_api = self._public_api()
427
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3359, in _public_api
428
+ self._cached_public_api = public.Api(overrides, api_key=self._settings.api_key)
429
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/apis/public/api.py", line 319, in __init__
430
+ wandb_login._verify_login(
431
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 389, in _verify_login
432
+ raise AuthenticationError(
433
+ wandb.errors.errors.AuthenticationError: Unable to connect to https://api.wandb.ai to verify API token.
434
+ Traceback (most recent call last):
435
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connection.py", line 204, in _new_conn
436
+ sock = connection.create_connection(
437
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/util/connection.py", line 60, in create_connection
438
+ for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
439
+ File "/usr/lib/python3.10/socket.py", line 955, in getaddrinfo
440
+ for res in _socket.getaddrinfo(host, port, family, type, proto, flags):
441
+ socket.gaierror: [Errno -3] Temporary failure in name resolution
442
+
443
+ The above exception was the direct cause of the following exception:
444
+
445
+ Traceback (most recent call last):
446
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 787, in urlopen
447
+ response = self._make_request(
448
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 488, in _make_request
449
+ raise new_e
450
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 464, in _make_request
451
+ self._validate_conn(conn)
452
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 1093, in _validate_conn
453
+ conn.connect()
454
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connection.py", line 759, in connect
455
+ self.sock = sock = self._new_conn()
456
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connection.py", line 211, in _new_conn
457
+ raise NameResolutionError(self.host, self, e) from e
458
+ urllib3.exceptions.NameResolutionError: HTTPSConnection(host='api.wandb.ai', port=443): Failed to resolve 'api.wandb.ai' ([Errno -3] Temporary failure in name resolution)
459
+
460
+ The above exception was the direct cause of the following exception:
461
+
462
+ Traceback (most recent call last):
463
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/adapters.py", line 644, in send
464
+ resp = conn.urlopen(
465
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/connectionpool.py", line 841, in urlopen
466
+ retries = retries.increment(
467
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/urllib3/util/retry.py", line 519, in increment
468
+ raise MaxRetryError(_pool, url, reason) from reason # type: ignore[arg-type]
469
+ urllib3.exceptions.MaxRetryError: HTTPSConnectionPool(host='api.wandb.ai', port=443): Max retries exceeded with url: /graphql (Caused by NameResolutionError("HTTPSConnection(host='api.wandb.ai', port=443): Failed to resolve 'api.wandb.ai' ([Errno -3] Temporary failure in name resolution)"))
470
+
471
+ During handling of the above exception, another exception occurred:
472
+
473
+ Traceback (most recent call last):
474
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 387, in _verify_login
475
+ is_api_key_valid = api.validate_api_key()
476
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/apis/internal.py", line 66, in validate_api_key
477
+ return self.api.validate_api_key()
478
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/internal/internal_api.py", line 402, in validate_api_key
479
+ res = self.execute(gql("query { viewer { id } }"))
480
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/internal/internal_api.py", line 391, in execute
481
+ return self.client.execute(*args, **kwargs) # type: ignore
482
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/vendor/gql-0.2.0/wandb_gql/client.py", line 52, in execute
483
+ result = self._get_result(document, *args, **kwargs)
484
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/vendor/gql-0.2.0/wandb_gql/client.py", line 60, in _get_result
485
+ return self.transport.execute(document, *args, **kwargs)
486
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/lib/gql_request.py", line 70, in execute
487
+ request = self.session.post(self.url, **post_args)
488
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/sessions.py", line 637, in post
489
+ return self.request("POST", url, data=data, json=json, **kwargs)
490
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/sessions.py", line 589, in request
491
+ resp = self.send(prep, **send_kwargs)
492
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/sessions.py", line 703, in send
493
+ r = adapter.send(request, **kwargs)
494
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/requests/adapters.py", line 677, in send
495
+ raise ConnectionError(e, request=request)
496
+ requests.exceptions.ConnectionError: HTTPSConnectionPool(host='api.wandb.ai', port=443): Max retries exceeded with url: /graphql (Caused by NameResolutionError("HTTPSConnection(host='api.wandb.ai', port=443): Failed to resolve 'api.wandb.ai' ([Errno -3] Temporary failure in name resolution)"))
497
+
498
+ The above exception was the direct cause of the following exception:
499
+
500
+ Traceback (most recent call last):
501
+ File "/code/lq/a2c2-libero/src/lerobot/scripts/train.py", line 291, in <module>
502
+ train()
503
+ File "/code/lq/a2c2-libero/src/lerobot/configs/parser.py", line 225, in wrapper_inner
504
+ response = fn(cfg, *args, **kwargs)
505
+ File "/code/lq/a2c2-libero/src/lerobot/scripts/train.py", line 246, in train
506
+ wandb_logger.log_policy(checkpoint_dir)
507
+ File "/code/lq/a2c2-libero/src/lerobot/utils/wandb_utils.py", line 116, in log_policy
508
+ self._wandb.log_artifact(artifact)
509
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 400, in wrapper
510
+ return func(self, *args, **kwargs)
511
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 458, in wrapper_fn
512
+ return func(self, *args, **kwargs)
513
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 445, in wrapper
514
+ return func(self, *args, **kwargs)
515
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3166, in log_artifact
516
+ return self._log_artifact(
517
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3311, in _log_artifact
518
+ self._assert_can_log_artifact(artifact)
519
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3371, in _assert_can_log_artifact
520
+ public_api = self._public_api()
521
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_run.py", line 3359, in _public_api
522
+ self._cached_public_api = public.Api(overrides, api_key=self._settings.api_key)
523
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/apis/public/api.py", line 319, in __init__
524
+ wandb_login._verify_login(
525
+ File "/code/lq/a2c2-libero/.venv/lib/python3.10/site-packages/wandb/sdk/wandb_login.py", line 389, in _verify_login
526
+ raise AuthenticationError(
527
+ wandb.errors.errors.AuthenticationError: Unable to connect to https://api.wandb.ai to verify API token.
wandb/run-20251216_174551-tckp70s8/files/requirements.txt ADDED
@@ -0,0 +1,166 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ async-timeout==5.0.1
2
+ tokenizers==0.21.4
3
+ nvidia-cusparse-cu12==12.5.8.93
4
+ protobuf==6.33.2
5
+ av==16.0.1
6
+ kiwisolver==1.4.9
7
+ nvidia-nvjitlink-cu12==12.8.93
8
+ cloudpickle==3.1.2
9
+ pynput==1.8.1
10
+ propcache==0.4.1
11
+ lerobot==0.2.0
12
+ six==1.17.0
13
+ smmap==5.0.2
14
+ Flask==3.1.2
15
+ imageio-ffmpeg==0.6.0
16
+ idna==3.11
17
+ Werkzeug==3.1.4
18
+ exceptiongroup==1.3.1
19
+ pyyaml-include==1.4.1
20
+ dill==0.3.8
21
+ huggingface-hub==0.36.0
22
+ ImageIO==2.37.2
23
+ anyio==4.12.0
24
+ referencing==0.37.0
25
+ torchvision==0.24.1
26
+ frozenlist==1.8.0
27
+ transformers==4.51.3
28
+ nvidia-cuda-runtime-cu12==12.8.90
29
+ nvidia-cusolver-cu12==11.7.3.90
30
+ pyparsing==3.2.5
31
+ opencv-python-headless==4.12.0.88
32
+ torchcodec==0.9.1
33
+ tqdm==4.67.1
34
+ tomli==2.3.0
35
+ scipy==1.15.3
36
+ easydict==1.13
37
+ safetensors==0.7.0
38
+ platformdirs==4.5.1
39
+ requests==2.32.5
40
+ gymnasium==0.29.1
41
+ hf_transfer==0.1.9
42
+ mergedeep==1.3.4
43
+ python-dateutil==2.9.0.post0
44
+ wcwidth==0.2.14
45
+ matplotlib==3.10.8
46
+ robosuite==1.5.1
47
+ nvidia-cudnn-cu12==9.10.2.21
48
+ diffusers==0.36.0
49
+ blinker==1.9.0
50
+ PyOpenGL==3.1.10
51
+ httpx==0.28.1
52
+ certifi==2025.11.12
53
+ jsonlines==4.0.0
54
+ networkx==3.4.2
55
+ Farama-Notifications==0.0.4
56
+ qpsolvers==4.8.2
57
+ httpcore==1.0.9
58
+ pyserial==3.5
59
+ filelock==3.20.0
60
+ draccus==0.10.0
61
+ mdurl==0.1.2
62
+ tzdata==2025.3
63
+ jupytext==1.18.1
64
+ GitPython==3.1.45
65
+ cmake==4.2.0
66
+ typing-inspect==0.9.0
67
+ wandb==0.23.1
68
+ PyYAML==6.0.3
69
+ Jinja2==3.1.6
70
+ mypy_extensions==1.1.0
71
+ annotated-types==0.7.0
72
+ inquirerpy==0.3.4
73
+ numpy==2.2.6
74
+ psutil==7.1.3
75
+ pydantic_core==2.41.5
76
+ nvidia-nvshmem-cu12==3.3.20
77
+ pillow==12.0.0
78
+ pluggy==1.6.0
79
+ pytest==9.0.2
80
+ toml==0.10.2
81
+ importlib_resources==6.5.2
82
+ fastjsonschema==2.21.2
83
+ num2words==0.5.14
84
+ accelerate==1.12.0
85
+ pyarrow==22.0.0
86
+ mpmath==1.3.0
87
+ yarl==1.22.0
88
+ jsonschema-specifications==2025.9.1
89
+ jupyter_core==5.9.1
90
+ hf-xet==1.2.0
91
+ nvidia-cuda-nvrtc-cu12==12.8.93
92
+ nvidia-nccl-cu12==2.27.5
93
+ python-xlib==0.33
94
+ contourpy==1.3.2
95
+ nvidia-cublas-cu12==12.8.4.1
96
+ pfzy==0.3.4
97
+ gym-notices==0.1.0
98
+ pandas==2.3.3
99
+ libero==0.1.0
100
+ pytz==2025.2
101
+ future==1.0.0
102
+ regex==2025.11.3
103
+ nvidia-curand-cu12==10.3.9.90
104
+ evdev==1.9.2
105
+ nvidia-cusparselt-cu12==0.7.1
106
+ docopt==0.6.2
107
+ fsspec==2025.3.0
108
+ aiohappyeyeballs==2.6.1
109
+ markdown-it-py==4.0.0
110
+ multidict==6.7.0
111
+ nvidia-cufile-cu12==1.13.1.3
112
+ orderly-set==5.5.0
113
+ daqp==0.7.2
114
+ sympy==1.14.0
115
+ cycler==0.12.1
116
+ rpds-py==0.30.0
117
+ Pygments==2.19.2
118
+ nvidia-cufft-cu12==11.3.3.83
119
+ traitlets==5.14.3
120
+ attrs==25.4.0
121
+ typing-inspection==0.4.2
122
+ etils==1.13.0
123
+ importlib_metadata==8.7.0
124
+ numba==0.63.1
125
+ mujoco==3.3.2
126
+ gitdb==4.0.12
127
+ gym==0.26.2
128
+ xxhash==3.6.0
129
+ rerun-sdk==0.22.1
130
+ nbformat==5.10.4
131
+ mink==0.0.13
132
+ itsdangerous==2.2.0
133
+ nvidia-nvtx-cu12==12.8.90
134
+ datasets==3.6.0
135
+ MarkupSafe==3.0.3
136
+ glfw==2.10.0
137
+ nvidia-cuda-cupti-cu12==12.8.90
138
+ jsonschema==4.25.1
139
+ packaging==25.0
140
+ triton==3.5.1
141
+ aiosignal==1.4.0
142
+ charset-normalizer==3.4.4
143
+ h11==0.16.0
144
+ deepdiff==8.6.1
145
+ llvmlite==0.46.0
146
+ aiohttp==3.13.2
147
+ iniconfig==2.3.0
148
+ zipp==3.23.0
149
+ typing_extensions==4.15.0
150
+ fonttools==4.61.1
151
+ pydantic==2.12.5
152
+ click==8.3.1
153
+ opencv-python==4.12.0.88
154
+ mdit-py-plugins==0.5.0
155
+ torch==2.9.1
156
+ bddl==3.6.0
157
+ urllib3==2.6.2
158
+ einops==0.8.1
159
+ joblib==1.5.3
160
+ prompt_toolkit==3.0.52
161
+ sentry-sdk==2.47.0
162
+ termcolor==3.2.0
163
+ nltk==3.9.2
164
+ multiprocess==0.70.16
165
+ absl-py==2.3.1
166
+ lerobot==0.2.0
wandb/run-20251216_174551-tckp70s8/files/wandb-metadata.json ADDED
@@ -0,0 +1,107 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-5.15.0-139-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.10.12",
4
+ "startedAt": "2025-12-16T09:45:51.384094Z",
5
+ "args": [
6
+ "--policy.type=smolvla",
7
+ "--policy.load_vlm_weights",
8
+ "True",
9
+ "--dataset.repo_id",
10
+ "Lakesenberg/libero_spatial_image",
11
+ "--dataset.root",
12
+ "/data/libero_spatial_image",
13
+ "--batch_size=64",
14
+ "--steps=100000",
15
+ "--policy.repo_id",
16
+ "username/libero_smolvla",
17
+ "--output_dir",
18
+ "libero_smolvla_a2c2",
19
+ "--job_name",
20
+ "libero_smolvla_a2c2",
21
+ "--wandb.enable",
22
+ "true"
23
+ ],
24
+ "program": "/code/lq/a2c2-libero/src/lerobot/scripts/train.py",
25
+ "codePath": "src/lerobot/scripts/train.py",
26
+ "codePathLocal": "src/lerobot/scripts/train.py",
27
+ "git": {
28
+ "remote": "https://github.com/k1000dai/a2c2-libero.git",
29
+ "commit": "54dd088302a0ef3f50c4add3ec927ab94d76a406"
30
+ },
31
+ "root": "libero_smolvla_a2c2",
32
+ "host": "GPU01",
33
+ "executable": "/code/lq/a2c2-libero/.venv/bin/python",
34
+ "cpu_count": 128,
35
+ "cpu_count_logical": 128,
36
+ "gpu": "NVIDIA A100-SXM4-80GB",
37
+ "gpu_count": 8,
38
+ "disk": {
39
+ "/": {
40
+ "total": "2362542452736",
41
+ "used": "2217778720768"
42
+ }
43
+ },
44
+ "memory": {
45
+ "total": "1082019954688"
46
+ },
47
+ "gpu_nvidia": [
48
+ {
49
+ "name": "NVIDIA A100-SXM4-80GB",
50
+ "memoryTotal": "85899345920",
51
+ "cudaCores": 6912,
52
+ "architecture": "Ampere",
53
+ "uuid": "GPU-d36e98cb-a1e9-190c-4f5b-4de233402cf5"
54
+ },
55
+ {
56
+ "name": "NVIDIA A100-SXM4-80GB",
57
+ "memoryTotal": "85899345920",
58
+ "cudaCores": 6912,
59
+ "architecture": "Ampere",
60
+ "uuid": "GPU-a5ea4ad0-8a8b-34af-3b05-58a0ac4c9fd0"
61
+ },
62
+ {
63
+ "name": "NVIDIA A100-SXM4-80GB",
64
+ "memoryTotal": "85899345920",
65
+ "cudaCores": 6912,
66
+ "architecture": "Ampere",
67
+ "uuid": "GPU-948a62a9-3401-457e-bba7-dc2781e19b1b"
68
+ },
69
+ {
70
+ "name": "NVIDIA A100-SXM4-80GB",
71
+ "memoryTotal": "85899345920",
72
+ "cudaCores": 6912,
73
+ "architecture": "Ampere",
74
+ "uuid": "GPU-5314fe46-34bf-c69c-dccc-976f70896178"
75
+ },
76
+ {
77
+ "name": "NVIDIA A100-SXM4-80GB",
78
+ "memoryTotal": "85899345920",
79
+ "cudaCores": 6912,
80
+ "architecture": "Ampere",
81
+ "uuid": "GPU-59048bb2-8115-237d-2ad3-a1f99b67c906"
82
+ },
83
+ {
84
+ "name": "NVIDIA A100-SXM4-80GB",
85
+ "memoryTotal": "85899345920",
86
+ "cudaCores": 6912,
87
+ "architecture": "Ampere",
88
+ "uuid": "GPU-a3f470ff-de9b-6067-d889-5e5d5e2a4860"
89
+ },
90
+ {
91
+ "name": "NVIDIA A100-SXM4-80GB",
92
+ "memoryTotal": "85899345920",
93
+ "cudaCores": 6912,
94
+ "architecture": "Ampere",
95
+ "uuid": "GPU-74a6ef3f-1b1c-5880-8fee-4441da7ac27a"
96
+ },
97
+ {
98
+ "name": "NVIDIA A100-SXM4-80GB",
99
+ "memoryTotal": "85899345920",
100
+ "cudaCores": 6912,
101
+ "architecture": "Ampere",
102
+ "uuid": "GPU-380c1f88-df1e-2984-c9fc-f424eb38064f"
103
+ }
104
+ ],
105
+ "cudaVersion": "12.8",
106
+ "writerId": "5kiukxb0n4qxfbottrvf5zv3vh1llpjw"
107
+ }
wandb/run-20251216_174551-tckp70s8/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_runtime":24706.033749086,"_step":20000,"train/lr":2.7316655657931005e-05,"train/episodes":10439.116481026997,"train/loss":0.023090576753020287,"train/update_s":0.5619091798737645,"train/dataloading_s":0.013679603883065283,"train/samples":1280000,"_timestamp":1.7659030324154131e+09,"train/epochs":24.164621483858788,"_wandb":{"runtime":24706},"train/steps":20000,"train/grad_norm":0.20806471958756448}
wandb/run-20251216_174551-tckp70s8/logs/debug-core.log ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-12-16T17:45:51.446999958+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpvambrt8z/port-2741015.txt","pid":2741015,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-12-16T17:45:51.449050783+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2741015}
3
+ {"time":"2025-12-16T17:45:51.448988553+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2741015-2741139-1250394604/socket","Net":"unix"}}
4
+ {"time":"2025-12-16T17:45:51.635551571+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-12-16T17:45:51.644601252+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"tckp70s8","id":"1(@)"}
6
+ {"time":"2025-12-16T17:45:52.426393607+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"tckp70s8","id":"1(@)"}
7
+ {"time":"2025-12-17T00:37:38.943715827+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
8
+ {"time":"2025-12-17T00:37:38.943961102+08:00","level":"INFO","msg":"server is shutting down"}
9
+ {"time":"2025-12-17T00:37:38.943950001+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2025-12-17T00:37:38.944082832+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
11
+ {"time":"2025-12-17T00:37:38.952146168+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2741015-2741139-1250394604/socket","Net":"unix"}}
12
+ {"time":"2025-12-17T01:14:04.137636326+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
13
+ {"time":"2025-12-17T01:14:04.137684938+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
14
+ {"time":"2025-12-17T01:14:04.137711589+08:00","level":"INFO","msg":"server is closed"}
wandb/run-20251216_174551-tckp70s8/logs/debug-internal.log ADDED
The diff for this file is too large to render. See raw diff
 
wandb/run-20251216_174551-tckp70s8/logs/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Current SDK version is 0.23.1
2
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Configure stats pid to 2741015
3
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Loading settings from /root/.config/wandb/settings
4
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Loading settings from /code/lq/a2c2-libero/wandb/settings
5
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_setup.py:_flush():80] Loading settings from environment variables
6
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_init.py:setup_run_log_directory():714] Logging user logs to libero_smolvla_a2c2/wandb/run-20251216_174551-tckp70s8/logs/debug.log
7
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_init.py:setup_run_log_directory():715] Logging internal logs to libero_smolvla_a2c2/wandb/run-20251216_174551-tckp70s8/logs/debug-internal.log
8
+ 2025-12-16 17:45:51,385 INFO MainThread:2741015 [wandb_init.py:init():841] calling init triggers
9
+ 2025-12-16 17:45:51,386 INFO MainThread:2741015 [wandb_init.py:init():846] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'Lakesenberg/libero_spatial_image', 'root': '/data/libero_spatial_image', 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec'}, 'env': None, 'policy': {'type': 'smolvla', 'n_obs_steps': 1, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': True, 'repo_id': 'username/libero_smolvla', 'private': None, 'tags': None, 'license': None, 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'resize_imgs_with_padding': [512, 512], 'empty_cameras': 0, 'adapt_to_pi_aloha': False, 'use_delta_joint_actions_aloha': False, 'tokenizer_max_length': 48, 'num_steps': 10, 'use_cache': True, 'freeze_vision_encoder': True, 'train_expert_only': True, 'train_state_proj': True, 'optimizer_lr': 0.0001, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 1e-10, 'optimizer_grad_clip_norm': 10, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'vlm_model_name': 'HuggingFaceTB/SmolVLM2-500M-Video-Instruct', 'load_vlm_weights': True, 'add_image_special_tokens': False, 'attention_mode': 'cross_attn', 'prefix_length': -1, 'pad_language_to': 'longest', 'num_expert_layers': -1, 'num_vlm_layers': 16, 'self_attn_every_n_layers': 2, 'expert_width_multiplier': 0.75, 'min_period': 0.004, 'max_period': 4.0}, 'output_dir': 'libero_smolvla_a2c2', 'job_name': 'libero_smolvla_a2c2', 'resume': False, 'seed': 1000, 'num_workers': 4, 'batch_size': 64, 'steps': 100000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 0.0001, 'weight_decay': 1e-10, 'grad_clip_norm': 10, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 0.0001, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, '_wandb': {}}
11
+ 2025-12-16 17:45:51,386 INFO MainThread:2741015 [wandb_init.py:init():889] starting backend
12
+ 2025-12-16 17:45:51,636 INFO MainThread:2741015 [wandb_init.py:init():892] sending inform_init request
13
+ 2025-12-16 17:45:51,642 INFO MainThread:2741015 [wandb_init.py:init():900] backend started and connected
14
+ 2025-12-16 17:45:51,646 INFO MainThread:2741015 [wandb_init.py:init():970] updated telemetry
15
+ 2025-12-16 17:45:51,654 INFO MainThread:2741015 [wandb_init.py:init():994] communicating run to backend with 90.0 second timeout
16
+ 2025-12-16 17:45:52,908 INFO MainThread:2741015 [wandb_init.py:init():1041] starting run threads in backend
17
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_console_start():2521] atexit reg
18
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_redirect():2369] redirect: wrap_raw
19
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_redirect():2438] Wrapping output streams.
20
+ 2025-12-16 17:45:53,074 INFO MainThread:2741015 [wandb_run.py:_redirect():2461] Redirects installed.
21
+ 2025-12-16 17:45:53,079 INFO MainThread:2741015 [wandb_init.py:init():1081] run started, returning control to user process
22
+ 2025-12-17 00:37:38,942 INFO wandb-AsyncioManager-main:2741015 [service_client.py:_forward_responses():80] Reached EOF.
23
+ 2025-12-17 00:37:38,943 INFO wandb-AsyncioManager-main:2741015 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles.
wandb/run-20251216_174551-tckp70s8/run-tckp70s8.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b637481ae5b34283752dc22073033b39cdfc54d76168bc81edb290a747816d61
3
+ size 6263908