hangVLA commited on
Commit
3125558
·
verified ·
1 Parent(s): bde20ec

Upload folder using huggingface_hub

Browse files
Files changed (24) hide show
  1. .gitattributes +1 -0
  2. checkpoints/010000/pretrained_model/config.json +91 -0
  3. checkpoints/010000/pretrained_model/model.safetensors +3 -0
  4. checkpoints/010000/pretrained_model/policy_postprocessor.json +32 -0
  5. checkpoints/010000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors +3 -0
  6. checkpoints/010000/pretrained_model/policy_preprocessor.json +91 -0
  7. checkpoints/010000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors +3 -0
  8. checkpoints/010000/pretrained_model/train_config.json +224 -0
  9. checkpoints/010000/training_state/optimizer_param_groups.json +798 -0
  10. checkpoints/010000/training_state/optimizer_state.safetensors +3 -0
  11. checkpoints/010000/training_state/rng_state.safetensors +3 -0
  12. checkpoints/010000/training_state/scheduler_state.json +14 -0
  13. checkpoints/010000/training_state/training_step.json +3 -0
  14. wandb/debug-internal.log +12 -0
  15. wandb/debug.log +23 -0
  16. wandb/run-20251130_205554-fadl65t9/files/config.yaml +255 -0
  17. wandb/run-20251130_205554-fadl65t9/files/output.log +145 -0
  18. wandb/run-20251130_205554-fadl65t9/files/requirements.txt +133 -0
  19. wandb/run-20251130_205554-fadl65t9/files/wandb-metadata.json +59 -0
  20. wandb/run-20251130_205554-fadl65t9/files/wandb-summary.json +1 -0
  21. wandb/run-20251130_205554-fadl65t9/logs/debug-core.log +15 -0
  22. wandb/run-20251130_205554-fadl65t9/logs/debug-internal.log +12 -0
  23. wandb/run-20251130_205554-fadl65t9/logs/debug.log +23 -0
  24. wandb/run-20251130_205554-fadl65t9/run-fadl65t9.wandb +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20251130_205554-fadl65t9/run-fadl65t9.wandb filter=lfs diff=lfs merge=lfs -text
checkpoints/010000/pretrained_model/config.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "pi0",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 10
9
+ ]
10
+ },
11
+ "observation.images.top": {
12
+ "type": "VISUAL",
13
+ "shape": [
14
+ 3,
15
+ 480,
16
+ 640
17
+ ]
18
+ },
19
+ "observation.images.wrist": {
20
+ "type": "VISUAL",
21
+ "shape": [
22
+ 3,
23
+ 480,
24
+ 640
25
+ ]
26
+ },
27
+ "observation.images.front": {
28
+ "type": "VISUAL",
29
+ "shape": [
30
+ 3,
31
+ 480,
32
+ 640
33
+ ]
34
+ }
35
+ },
36
+ "output_features": {
37
+ "action": {
38
+ "type": "ACTION",
39
+ "shape": [
40
+ 10
41
+ ]
42
+ }
43
+ },
44
+ "device": "cuda",
45
+ "use_amp": false,
46
+ "push_to_hub": false,
47
+ "repo_id": null,
48
+ "private": null,
49
+ "tags": null,
50
+ "license": null,
51
+ "pretrained_path": "lerobot/pi0_base",
52
+ "paligemma_variant": "gemma_2b",
53
+ "action_expert_variant": "gemma_300m",
54
+ "dtype": "bfloat16",
55
+ "chunk_size": 50,
56
+ "n_action_steps": 50,
57
+ "max_state_dim": 32,
58
+ "max_action_dim": 32,
59
+ "num_inference_steps": 10,
60
+ "time_sampling_beta_alpha": 1.5,
61
+ "time_sampling_beta_beta": 1.0,
62
+ "time_sampling_scale": 0.999,
63
+ "time_sampling_offset": 0.001,
64
+ "min_period": 0.004,
65
+ "max_period": 4.0,
66
+ "image_resolution": [
67
+ 224,
68
+ 224
69
+ ],
70
+ "empty_cameras": 0,
71
+ "normalization_mapping": {
72
+ "VISUAL": "IDENTITY",
73
+ "STATE": "MEAN_STD",
74
+ "ACTION": "MEAN_STD"
75
+ },
76
+ "gradient_checkpointing": true,
77
+ "compile_model": true,
78
+ "compile_mode": "max-autotune",
79
+ "optimizer_lr": 2.5e-05,
80
+ "optimizer_betas": [
81
+ 0.9,
82
+ 0.95
83
+ ],
84
+ "optimizer_eps": 1e-08,
85
+ "optimizer_weight_decay": 0.01,
86
+ "optimizer_grad_clip_norm": 1.0,
87
+ "scheduler_warmup_steps": 1000,
88
+ "scheduler_decay_steps": 30000,
89
+ "scheduler_decay_lr": 2.5e-06,
90
+ "tokenizer_max_length": 48
91
+ }
checkpoints/010000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0be115ceb5009188fcfa9ff8b7072809b9134533ee7b3e8ee2a583a507af90ba
3
+ size 7011548640
checkpoints/010000/pretrained_model/policy_postprocessor.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_postprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "unnormalizer_processor",
6
+ "config": {
7
+ "eps": 1e-08,
8
+ "features": {
9
+ "action": {
10
+ "type": "ACTION",
11
+ "shape": [
12
+ 10
13
+ ]
14
+ }
15
+ },
16
+ "norm_map": {
17
+ "VISUAL": "IDENTITY",
18
+ "STATE": "MEAN_STD",
19
+ "ACTION": "MEAN_STD"
20
+ }
21
+ },
22
+ "state_file": "policy_postprocessor_step_0_unnormalizer_processor.safetensors"
23
+ },
24
+ {
25
+ "registry_name": "device_processor",
26
+ "config": {
27
+ "device": "cpu",
28
+ "float_dtype": null
29
+ }
30
+ }
31
+ ]
32
+ }
checkpoints/010000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:006c815876cc8f364a8e1bcc96cbd1930cb6e50d543fdf3d73ba7387144fb62d
3
+ size 8880
checkpoints/010000/pretrained_model/policy_preprocessor.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_preprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "rename_observations_processor",
6
+ "config": {
7
+ "rename_map": {
8
+ "observation.images.camera1": "observation.images.wrist",
9
+ "observation.images.camera2": "observation.images.top",
10
+ "observation.images.camera3": "observation.images.front"
11
+ }
12
+ }
13
+ },
14
+ {
15
+ "registry_name": "to_batch_processor",
16
+ "config": {}
17
+ },
18
+ {
19
+ "registry_name": "pi0_new_line_processor",
20
+ "config": {}
21
+ },
22
+ {
23
+ "registry_name": "tokenizer_processor",
24
+ "config": {
25
+ "max_length": 48,
26
+ "task_key": "task",
27
+ "padding_side": "right",
28
+ "padding": "max_length",
29
+ "truncation": true,
30
+ "tokenizer_name": "google/paligemma-3b-pt-224"
31
+ }
32
+ },
33
+ {
34
+ "registry_name": "device_processor",
35
+ "config": {
36
+ "device": "cuda",
37
+ "float_dtype": null
38
+ }
39
+ },
40
+ {
41
+ "registry_name": "normalizer_processor",
42
+ "config": {
43
+ "eps": 1e-08,
44
+ "features": {
45
+ "observation.state": {
46
+ "type": "STATE",
47
+ "shape": [
48
+ 10
49
+ ]
50
+ },
51
+ "observation.images.top": {
52
+ "type": "VISUAL",
53
+ "shape": [
54
+ 3,
55
+ 480,
56
+ 640
57
+ ]
58
+ },
59
+ "observation.images.wrist": {
60
+ "type": "VISUAL",
61
+ "shape": [
62
+ 3,
63
+ 480,
64
+ 640
65
+ ]
66
+ },
67
+ "observation.images.front": {
68
+ "type": "VISUAL",
69
+ "shape": [
70
+ 3,
71
+ 480,
72
+ 640
73
+ ]
74
+ },
75
+ "action": {
76
+ "type": "ACTION",
77
+ "shape": [
78
+ 10
79
+ ]
80
+ }
81
+ },
82
+ "norm_map": {
83
+ "VISUAL": "IDENTITY",
84
+ "STATE": "MEAN_STD",
85
+ "ACTION": "MEAN_STD"
86
+ }
87
+ },
88
+ "state_file": "policy_preprocessor_step_5_normalizer_processor.safetensors"
89
+ }
90
+ ]
91
+ }
checkpoints/010000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:006c815876cc8f364a8e1bcc96cbd1930cb6e50d543fdf3d73ba7387144fb62d
3
+ size 8880
checkpoints/010000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,224 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "/home/vla/so101_1130_2tasks",
4
+ "root": null,
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ },
61
+ "affine": {
62
+ "weight": 1.0,
63
+ "type": "RandomAffine",
64
+ "kwargs": {
65
+ "degrees": [
66
+ -5.0,
67
+ 5.0
68
+ ],
69
+ "translate": [
70
+ 0.05,
71
+ 0.05
72
+ ]
73
+ }
74
+ }
75
+ }
76
+ },
77
+ "revision": null,
78
+ "use_imagenet_stats": true,
79
+ "video_backend": "torchcodec",
80
+ "streaming": false
81
+ },
82
+ "env": null,
83
+ "policy": {
84
+ "type": "pi0",
85
+ "n_obs_steps": 1,
86
+ "input_features": {
87
+ "observation.state": {
88
+ "type": "STATE",
89
+ "shape": [
90
+ 10
91
+ ]
92
+ },
93
+ "observation.images.top": {
94
+ "type": "VISUAL",
95
+ "shape": [
96
+ 3,
97
+ 480,
98
+ 640
99
+ ]
100
+ },
101
+ "observation.images.wrist": {
102
+ "type": "VISUAL",
103
+ "shape": [
104
+ 3,
105
+ 480,
106
+ 640
107
+ ]
108
+ },
109
+ "observation.images.front": {
110
+ "type": "VISUAL",
111
+ "shape": [
112
+ 3,
113
+ 480,
114
+ 640
115
+ ]
116
+ }
117
+ },
118
+ "output_features": {
119
+ "action": {
120
+ "type": "ACTION",
121
+ "shape": [
122
+ 10
123
+ ]
124
+ }
125
+ },
126
+ "device": "cuda",
127
+ "use_amp": false,
128
+ "push_to_hub": false,
129
+ "repo_id": null,
130
+ "private": null,
131
+ "tags": null,
132
+ "license": null,
133
+ "pretrained_path": "lerobot/pi0_base",
134
+ "paligemma_variant": "gemma_2b",
135
+ "action_expert_variant": "gemma_300m",
136
+ "dtype": "bfloat16",
137
+ "chunk_size": 50,
138
+ "n_action_steps": 50,
139
+ "max_state_dim": 32,
140
+ "max_action_dim": 32,
141
+ "num_inference_steps": 10,
142
+ "time_sampling_beta_alpha": 1.5,
143
+ "time_sampling_beta_beta": 1.0,
144
+ "time_sampling_scale": 0.999,
145
+ "time_sampling_offset": 0.001,
146
+ "min_period": 0.004,
147
+ "max_period": 4.0,
148
+ "image_resolution": [
149
+ 224,
150
+ 224
151
+ ],
152
+ "empty_cameras": 0,
153
+ "normalization_mapping": {
154
+ "VISUAL": "IDENTITY",
155
+ "STATE": "MEAN_STD",
156
+ "ACTION": "MEAN_STD"
157
+ },
158
+ "gradient_checkpointing": true,
159
+ "compile_model": true,
160
+ "compile_mode": "max-autotune",
161
+ "optimizer_lr": 2.5e-05,
162
+ "optimizer_betas": [
163
+ 0.9,
164
+ 0.95
165
+ ],
166
+ "optimizer_eps": 1e-08,
167
+ "optimizer_weight_decay": 0.01,
168
+ "optimizer_grad_clip_norm": 1.0,
169
+ "scheduler_warmup_steps": 1000,
170
+ "scheduler_decay_steps": 30000,
171
+ "scheduler_decay_lr": 2.5e-06,
172
+ "tokenizer_max_length": 48
173
+ },
174
+ "output_dir": "/home/vla/so101_1130_2tasks_pi0train",
175
+ "job_name": "so101_1130_2tasks_pi0train",
176
+ "resume": false,
177
+ "seed": 1000,
178
+ "num_workers": 16,
179
+ "batch_size": 24,
180
+ "steps": 10000,
181
+ "eval_freq": 20000,
182
+ "log_freq": 200,
183
+ "save_checkpoint": true,
184
+ "save_freq": 20000,
185
+ "use_policy_training_preset": true,
186
+ "optimizer": {
187
+ "type": "adamw",
188
+ "lr": 2.5e-05,
189
+ "weight_decay": 0.01,
190
+ "grad_clip_norm": 1.0,
191
+ "betas": [
192
+ 0.9,
193
+ 0.95
194
+ ],
195
+ "eps": 1e-08
196
+ },
197
+ "scheduler": {
198
+ "type": "cosine_decay_with_warmup",
199
+ "num_warmup_steps": 1000,
200
+ "num_decay_steps": 30000,
201
+ "peak_lr": 2.5e-05,
202
+ "decay_lr": 2.5e-06
203
+ },
204
+ "eval": {
205
+ "n_episodes": 50,
206
+ "batch_size": 50,
207
+ "use_async_envs": false
208
+ },
209
+ "wandb": {
210
+ "enable": true,
211
+ "disable_artifact": false,
212
+ "project": "lerobot",
213
+ "entity": null,
214
+ "notes": null,
215
+ "run_id": "fadl65t9",
216
+ "mode": null
217
+ },
218
+ "checkpoint_path": null,
219
+ "rename_map": {
220
+ "observation.images.camera1": "observation.images.wrist",
221
+ "observation.images.camera2": "observation.images.top",
222
+ "observation.images.camera3": "observation.images.front"
223
+ }
224
+ }
checkpoints/010000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,798 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.5e-06,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 0.01,
10
+ "amsgrad": false,
11
+ "maximize": false,
12
+ "foreach": null,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "decoupled_weight_decay": true,
17
+ "initial_lr": 2.5e-05,
18
+ "params": [
19
+ 0,
20
+ 1,
21
+ 2,
22
+ 3,
23
+ 4,
24
+ 5,
25
+ 6,
26
+ 7,
27
+ 8,
28
+ 9,
29
+ 10,
30
+ 11,
31
+ 12,
32
+ 13,
33
+ 14,
34
+ 15,
35
+ 16,
36
+ 17,
37
+ 18,
38
+ 19,
39
+ 20,
40
+ 21,
41
+ 22,
42
+ 23,
43
+ 24,
44
+ 25,
45
+ 26,
46
+ 27,
47
+ 28,
48
+ 29,
49
+ 30,
50
+ 31,
51
+ 32,
52
+ 33,
53
+ 34,
54
+ 35,
55
+ 36,
56
+ 37,
57
+ 38,
58
+ 39,
59
+ 40,
60
+ 41,
61
+ 42,
62
+ 43,
63
+ 44,
64
+ 45,
65
+ 46,
66
+ 47,
67
+ 48,
68
+ 49,
69
+ 50,
70
+ 51,
71
+ 52,
72
+ 53,
73
+ 54,
74
+ 55,
75
+ 56,
76
+ 57,
77
+ 58,
78
+ 59,
79
+ 60,
80
+ 61,
81
+ 62,
82
+ 63,
83
+ 64,
84
+ 65,
85
+ 66,
86
+ 67,
87
+ 68,
88
+ 69,
89
+ 70,
90
+ 71,
91
+ 72,
92
+ 73,
93
+ 74,
94
+ 75,
95
+ 76,
96
+ 77,
97
+ 78,
98
+ 79,
99
+ 80,
100
+ 81,
101
+ 82,
102
+ 83,
103
+ 84,
104
+ 85,
105
+ 86,
106
+ 87,
107
+ 88,
108
+ 89,
109
+ 90,
110
+ 91,
111
+ 92,
112
+ 93,
113
+ 94,
114
+ 95,
115
+ 96,
116
+ 97,
117
+ 98,
118
+ 99,
119
+ 100,
120
+ 101,
121
+ 102,
122
+ 103,
123
+ 104,
124
+ 105,
125
+ 106,
126
+ 107,
127
+ 108,
128
+ 109,
129
+ 110,
130
+ 111,
131
+ 112,
132
+ 113,
133
+ 114,
134
+ 115,
135
+ 116,
136
+ 117,
137
+ 118,
138
+ 119,
139
+ 120,
140
+ 121,
141
+ 122,
142
+ 123,
143
+ 124,
144
+ 125,
145
+ 126,
146
+ 127,
147
+ 128,
148
+ 129,
149
+ 130,
150
+ 131,
151
+ 132,
152
+ 133,
153
+ 134,
154
+ 135,
155
+ 136,
156
+ 137,
157
+ 138,
158
+ 139,
159
+ 140,
160
+ 141,
161
+ 142,
162
+ 143,
163
+ 144,
164
+ 145,
165
+ 146,
166
+ 147,
167
+ 148,
168
+ 149,
169
+ 150,
170
+ 151,
171
+ 152,
172
+ 153,
173
+ 154,
174
+ 155,
175
+ 156,
176
+ 157,
177
+ 158,
178
+ 159,
179
+ 160,
180
+ 161,
181
+ 162,
182
+ 163,
183
+ 164,
184
+ 165,
185
+ 166,
186
+ 167,
187
+ 168,
188
+ 169,
189
+ 170,
190
+ 171,
191
+ 172,
192
+ 173,
193
+ 174,
194
+ 175,
195
+ 176,
196
+ 177,
197
+ 178,
198
+ 179,
199
+ 180,
200
+ 181,
201
+ 182,
202
+ 183,
203
+ 184,
204
+ 185,
205
+ 186,
206
+ 187,
207
+ 188,
208
+ 189,
209
+ 190,
210
+ 191,
211
+ 192,
212
+ 193,
213
+ 194,
214
+ 195,
215
+ 196,
216
+ 197,
217
+ 198,
218
+ 199,
219
+ 200,
220
+ 201,
221
+ 202,
222
+ 203,
223
+ 204,
224
+ 205,
225
+ 206,
226
+ 207,
227
+ 208,
228
+ 209,
229
+ 210,
230
+ 211,
231
+ 212,
232
+ 213,
233
+ 214,
234
+ 215,
235
+ 216,
236
+ 217,
237
+ 218,
238
+ 219,
239
+ 220,
240
+ 221,
241
+ 222,
242
+ 223,
243
+ 224,
244
+ 225,
245
+ 226,
246
+ 227,
247
+ 228,
248
+ 229,
249
+ 230,
250
+ 231,
251
+ 232,
252
+ 233,
253
+ 234,
254
+ 235,
255
+ 236,
256
+ 237,
257
+ 238,
258
+ 239,
259
+ 240,
260
+ 241,
261
+ 242,
262
+ 243,
263
+ 244,
264
+ 245,
265
+ 246,
266
+ 247,
267
+ 248,
268
+ 249,
269
+ 250,
270
+ 251,
271
+ 252,
272
+ 253,
273
+ 254,
274
+ 255,
275
+ 256,
276
+ 257,
277
+ 258,
278
+ 259,
279
+ 260,
280
+ 261,
281
+ 262,
282
+ 263,
283
+ 264,
284
+ 265,
285
+ 266,
286
+ 267,
287
+ 268,
288
+ 269,
289
+ 270,
290
+ 271,
291
+ 272,
292
+ 273,
293
+ 274,
294
+ 275,
295
+ 276,
296
+ 277,
297
+ 278,
298
+ 279,
299
+ 280,
300
+ 281,
301
+ 282,
302
+ 283,
303
+ 284,
304
+ 285,
305
+ 286,
306
+ 287,
307
+ 288,
308
+ 289,
309
+ 290,
310
+ 291,
311
+ 292,
312
+ 293,
313
+ 294,
314
+ 295,
315
+ 296,
316
+ 297,
317
+ 298,
318
+ 299,
319
+ 300,
320
+ 301,
321
+ 302,
322
+ 303,
323
+ 304,
324
+ 305,
325
+ 306,
326
+ 307,
327
+ 308,
328
+ 309,
329
+ 310,
330
+ 311,
331
+ 312,
332
+ 313,
333
+ 314,
334
+ 315,
335
+ 316,
336
+ 317,
337
+ 318,
338
+ 319,
339
+ 320,
340
+ 321,
341
+ 322,
342
+ 323,
343
+ 324,
344
+ 325,
345
+ 326,
346
+ 327,
347
+ 328,
348
+ 329,
349
+ 330,
350
+ 331,
351
+ 332,
352
+ 333,
353
+ 334,
354
+ 335,
355
+ 336,
356
+ 337,
357
+ 338,
358
+ 339,
359
+ 340,
360
+ 341,
361
+ 342,
362
+ 343,
363
+ 344,
364
+ 345,
365
+ 346,
366
+ 347,
367
+ 348,
368
+ 349,
369
+ 350,
370
+ 351,
371
+ 352,
372
+ 353,
373
+ 354,
374
+ 355,
375
+ 356,
376
+ 357,
377
+ 358,
378
+ 359,
379
+ 360,
380
+ 361,
381
+ 362,
382
+ 363,
383
+ 364,
384
+ 365,
385
+ 366,
386
+ 367,
387
+ 368,
388
+ 369,
389
+ 370,
390
+ 371,
391
+ 372,
392
+ 373,
393
+ 374,
394
+ 375,
395
+ 376,
396
+ 377,
397
+ 378,
398
+ 379,
399
+ 380,
400
+ 381,
401
+ 382,
402
+ 383,
403
+ 384,
404
+ 385,
405
+ 386,
406
+ 387,
407
+ 388,
408
+ 389,
409
+ 390,
410
+ 391,
411
+ 392,
412
+ 393,
413
+ 394,
414
+ 395,
415
+ 396,
416
+ 397,
417
+ 398,
418
+ 399,
419
+ 400,
420
+ 401,
421
+ 402,
422
+ 403,
423
+ 404,
424
+ 405,
425
+ 406,
426
+ 407,
427
+ 408,
428
+ 409,
429
+ 410,
430
+ 411,
431
+ 412,
432
+ 413,
433
+ 414,
434
+ 415,
435
+ 416,
436
+ 417,
437
+ 418,
438
+ 419,
439
+ 420,
440
+ 421,
441
+ 422,
442
+ 423,
443
+ 424,
444
+ 425,
445
+ 426,
446
+ 427,
447
+ 428,
448
+ 429,
449
+ 430,
450
+ 431,
451
+ 432,
452
+ 433,
453
+ 434,
454
+ 435,
455
+ 436,
456
+ 437,
457
+ 438,
458
+ 439,
459
+ 440,
460
+ 441,
461
+ 442,
462
+ 443,
463
+ 444,
464
+ 445,
465
+ 446,
466
+ 447,
467
+ 448,
468
+ 449,
469
+ 450,
470
+ 451,
471
+ 452,
472
+ 453,
473
+ 454,
474
+ 455,
475
+ 456,
476
+ 457,
477
+ 458,
478
+ 459,
479
+ 460,
480
+ 461,
481
+ 462,
482
+ 463,
483
+ 464,
484
+ 465,
485
+ 466,
486
+ 467,
487
+ 468,
488
+ 469,
489
+ 470,
490
+ 471,
491
+ 472,
492
+ 473,
493
+ 474,
494
+ 475,
495
+ 476,
496
+ 477,
497
+ 478,
498
+ 479,
499
+ 480,
500
+ 481,
501
+ 482,
502
+ 483,
503
+ 484,
504
+ 485,
505
+ 486,
506
+ 487,
507
+ 488,
508
+ 489,
509
+ 490,
510
+ 491,
511
+ 492,
512
+ 493,
513
+ 494,
514
+ 495,
515
+ 496,
516
+ 497,
517
+ 498,
518
+ 499,
519
+ 500,
520
+ 501,
521
+ 502,
522
+ 503,
523
+ 504,
524
+ 505,
525
+ 506,
526
+ 507,
527
+ 508,
528
+ 509,
529
+ 510,
530
+ 511,
531
+ 512,
532
+ 513,
533
+ 514,
534
+ 515,
535
+ 516,
536
+ 517,
537
+ 518,
538
+ 519,
539
+ 520,
540
+ 521,
541
+ 522,
542
+ 523,
543
+ 524,
544
+ 525,
545
+ 526,
546
+ 527,
547
+ 528,
548
+ 529,
549
+ 530,
550
+ 531,
551
+ 532,
552
+ 533,
553
+ 534,
554
+ 535,
555
+ 536,
556
+ 537,
557
+ 538,
558
+ 539,
559
+ 540,
560
+ 541,
561
+ 542,
562
+ 543,
563
+ 544,
564
+ 545,
565
+ 546,
566
+ 547,
567
+ 548,
568
+ 549,
569
+ 550,
570
+ 551,
571
+ 552,
572
+ 553,
573
+ 554,
574
+ 555,
575
+ 556,
576
+ 557,
577
+ 558,
578
+ 559,
579
+ 560,
580
+ 561,
581
+ 562,
582
+ 563,
583
+ 564,
584
+ 565,
585
+ 566,
586
+ 567,
587
+ 568,
588
+ 569,
589
+ 570,
590
+ 571,
591
+ 572,
592
+ 573,
593
+ 574,
594
+ 575,
595
+ 576,
596
+ 577,
597
+ 578,
598
+ 579,
599
+ 580,
600
+ 581,
601
+ 582,
602
+ 583,
603
+ 584,
604
+ 585,
605
+ 586,
606
+ 587,
607
+ 588,
608
+ 589,
609
+ 590,
610
+ 591,
611
+ 592,
612
+ 593,
613
+ 594,
614
+ 595,
615
+ 596,
616
+ 597,
617
+ 598,
618
+ 599,
619
+ 600,
620
+ 601,
621
+ 602,
622
+ 603,
623
+ 604,
624
+ 605,
625
+ 606,
626
+ 607,
627
+ 608,
628
+ 609,
629
+ 610,
630
+ 611,
631
+ 612,
632
+ 613,
633
+ 614,
634
+ 615,
635
+ 616,
636
+ 617,
637
+ 618,
638
+ 619,
639
+ 620,
640
+ 621,
641
+ 622,
642
+ 623,
643
+ 624,
644
+ 625,
645
+ 626,
646
+ 627,
647
+ 628,
648
+ 629,
649
+ 630,
650
+ 631,
651
+ 632,
652
+ 633,
653
+ 634,
654
+ 635,
655
+ 636,
656
+ 637,
657
+ 638,
658
+ 639,
659
+ 640,
660
+ 641,
661
+ 642,
662
+ 643,
663
+ 644,
664
+ 645,
665
+ 646,
666
+ 647,
667
+ 648,
668
+ 649,
669
+ 650,
670
+ 651,
671
+ 652,
672
+ 653,
673
+ 654,
674
+ 655,
675
+ 656,
676
+ 657,
677
+ 658,
678
+ 659,
679
+ 660,
680
+ 661,
681
+ 662,
682
+ 663,
683
+ 664,
684
+ 665,
685
+ 666,
686
+ 667,
687
+ 668,
688
+ 669,
689
+ 670,
690
+ 671,
691
+ 672,
692
+ 673,
693
+ 674,
694
+ 675,
695
+ 676,
696
+ 677,
697
+ 678,
698
+ 679,
699
+ 680,
700
+ 681,
701
+ 682,
702
+ 683,
703
+ 684,
704
+ 685,
705
+ 686,
706
+ 687,
707
+ 688,
708
+ 689,
709
+ 690,
710
+ 691,
711
+ 692,
712
+ 693,
713
+ 694,
714
+ 695,
715
+ 696,
716
+ 697,
717
+ 698,
718
+ 699,
719
+ 700,
720
+ 701,
721
+ 702,
722
+ 703,
723
+ 704,
724
+ 705,
725
+ 706,
726
+ 707,
727
+ 708,
728
+ 709,
729
+ 710,
730
+ 711,
731
+ 712,
732
+ 713,
733
+ 714,
734
+ 715,
735
+ 716,
736
+ 717,
737
+ 718,
738
+ 719,
739
+ 720,
740
+ 721,
741
+ 722,
742
+ 723,
743
+ 724,
744
+ 725,
745
+ 726,
746
+ 727,
747
+ 728,
748
+ 729,
749
+ 730,
750
+ 731,
751
+ 732,
752
+ 733,
753
+ 734,
754
+ 735,
755
+ 736,
756
+ 737,
757
+ 738,
758
+ 739,
759
+ 740,
760
+ 741,
761
+ 742,
762
+ 743,
763
+ 744,
764
+ 745,
765
+ 746,
766
+ 747,
767
+ 748,
768
+ 749,
769
+ 750,
770
+ 751,
771
+ 752,
772
+ 753,
773
+ 754,
774
+ 755,
775
+ 756,
776
+ 757,
777
+ 758,
778
+ 759,
779
+ 760,
780
+ 761,
781
+ 762,
782
+ 763,
783
+ 764,
784
+ 765,
785
+ 766,
786
+ 767,
787
+ 768,
788
+ 769,
789
+ 770,
790
+ 771,
791
+ 772,
792
+ 773,
793
+ 774,
794
+ 775,
795
+ 776
796
+ ]
797
+ }
798
+ ]
checkpoints/010000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:620cad9091343baae7e6a9ed072fc3169a69d2f7b0ae87e71e68047417dfd122
3
+ size 12550279200
checkpoints/010000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02c78e77f9c867ce1a8ff0b831e665b289ded5d815a1eb47de2e475660b22e56
3
+ size 15708
checkpoints/010000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 2.5e-05
4
+ ],
5
+ "last_epoch": 10000,
6
+ "_step_count": 10001,
7
+ "_get_lr_called_within_step": false,
8
+ "_last_lr": [
9
+ 2.5e-06
10
+ ],
11
+ "lr_lambdas": [
12
+ null
13
+ ]
14
+ }
checkpoints/010000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 10000
3
+ }
wandb/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-30T20:55:54.770291237+09:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
2
+ {"time":"2025-11-30T20:55:55.192165852+09:00","level":"INFO","msg":"stream: created new stream","id":"fadl65t9"}
3
+ {"time":"2025-11-30T20:55:55.192248512+09:00","level":"INFO","msg":"stream: started","id":"fadl65t9"}
4
+ {"time":"2025-11-30T20:55:55.19229908+09:00","level":"INFO","msg":"writer: started","stream_id":"fadl65t9"}
5
+ {"time":"2025-11-30T20:55:55.192297075+09:00","level":"INFO","msg":"handler: started","stream_id":"fadl65t9"}
6
+ {"time":"2025-11-30T20:55:55.192405884+09:00","level":"INFO","msg":"sender: started","stream_id":"fadl65t9"}
7
+ {"time":"2025-11-30T23:08:40.702759164+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/wjdfufgusrn1234-konkuk-university/lerobot/fadl65t9/file_stream\": dial tcp 35.186.228.49:443: connect: connection timed out"}
8
+ {"time":"2025-12-01T09:14:24.3819952+09:00","level":"INFO","msg":"stream: closing","id":"fadl65t9"}
9
+ {"time":"2025-12-01T09:15:33.217839615+09:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
10
+ {"time":"2025-12-01T09:15:33.622503049+09:00","level":"INFO","msg":"handler: closed","stream_id":"fadl65t9"}
11
+ {"time":"2025-12-01T09:15:33.622701638+09:00","level":"INFO","msg":"sender: closed","stream_id":"fadl65t9"}
12
+ {"time":"2025-12-01T09:15:33.622736417+09:00","level":"INFO","msg":"stream: closed","id":"fadl65t9"}
wandb/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
2
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Configure stats pid to 2873154
3
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Loading settings from /home/vla/.config/wandb/settings
4
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Loading settings from /home/vla/wandb/settings
5
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /home/vla/so101_1130_2tasks_pi0train/wandb/run-20251130_205554-fadl65t9/logs/debug.log
7
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /home/vla/so101_1130_2tasks_pi0train/wandb/run-20251130_205554-fadl65t9/logs/debug-internal.log
8
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:init():813] calling init triggers
9
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': '/home/vla/so101_1130_2tasks', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi0', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': None, 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi0_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'image_resolution': [224, 224], 'empty_cameras': 0, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'gradient_checkpointing': True, 'compile_model': True, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'tokenizer_max_length': 48}, 'output_dir': '/home/vla/so101_1130_2tasks_pi0train', 'job_name': 'so101_1130_2tasks_pi0train', 'resume': False, 'seed': 1000, 'num_workers': 16, 'batch_size': 24, 'steps': 10000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {'observation.images.camera1': 'observation.images.wrist', 'observation.images.camera2': 'observation.images.top', 'observation.images.camera3': 'observation.images.front'}, '_wandb': {}}
11
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:init():854] starting backend
12
+ 2025-11-30 20:55:54,752 INFO MainThread:2873154 [wandb_init.py:init():857] sending inform_init request
13
+ 2025-11-30 20:55:54,759 INFO MainThread:2873154 [wandb_init.py:init():865] backend started and connected
14
+ 2025-11-30 20:55:54,765 INFO MainThread:2873154 [wandb_init.py:init():936] updated telemetry
15
+ 2025-11-30 20:55:54,766 INFO MainThread:2873154 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
16
+ 2025-11-30 20:55:55,602 INFO MainThread:2873154 [wandb_init.py:init():1011] starting run threads in backend
17
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-11-30 20:55:55,717 INFO MainThread:2873154 [wandb_init.py:init():1049] run started, returning control to user process
22
+ 2025-12-01 09:14:24,382 INFO wandb-AsyncioManager-main:2873154 [service_client.py:_forward_responses():84] Reached EOF.
23
+ 2025-12-01 09:14:24,382 INFO wandb-AsyncioManager-main:2873154 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles.
wandb/run-20251130_205554-fadl65t9/files/config.yaml ADDED
@@ -0,0 +1,255 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.21.4
4
+ e:
5
+ ujjhup6u1if9spwto41rqzxwewcfoopf:
6
+ args:
7
+ - --dataset.repo_id=/home/vla/so101_1130_2tasks
8
+ - --policy.type=pi0
9
+ - --policy.pretrained_path=lerobot/pi0_base
10
+ - --output_dir=/home/vla/so101_1130_2tasks_pi0train
11
+ - --job_name=so101_1130_2tasks_pi0train
12
+ - --steps=10000
13
+ - --batch_size=24
14
+ - --num_workers=16
15
+ - --policy.device=cuda
16
+ - --wandb.enable=true
17
+ - --policy.compile_model=true
18
+ - --policy.gradient_checkpointing=true
19
+ - --policy.dtype=bfloat16
20
+ - --policy.push_to_hub=false
21
+ - '--rename_map={"observation.images.camera1": "observation.images.wrist", "observation.images.camera2": "observation.images.top", "observation.images.camera3": "observation.images.front"}'
22
+ codePath: miniforge3/envs/lerobot/bin/lerobot-train
23
+ codePathLocal: miniforge3/envs/lerobot/bin/lerobot-train
24
+ cpu_count: 56
25
+ cpu_count_logical: 112
26
+ cudaVersion: "12.8"
27
+ disk:
28
+ /:
29
+ total: "943412031488"
30
+ used: "781631213568"
31
+ executable: /home/vla/miniforge3/envs/lerobot/bin/python3.10
32
+ gpu: NVIDIA RTX A6000
33
+ gpu_count: 2
34
+ gpu_nvidia:
35
+ - architecture: Ampere
36
+ cudaCores: 10752
37
+ memoryTotal: "51527024640"
38
+ name: NVIDIA RTX A6000
39
+ uuid: GPU-c895b718-8738-3fff-b42c-92cad284b785
40
+ - architecture: Ampere
41
+ cudaCores: 10752
42
+ memoryTotal: "51527024640"
43
+ name: NVIDIA RTX A6000
44
+ uuid: GPU-34849e7e-d621-dccc-ca70-b463192b3472
45
+ host: edu02
46
+ memory:
47
+ total: "540628525056"
48
+ os: Linux-6.8.0-88-generic-x86_64-with-glibc2.39
49
+ program: /home/vla/miniforge3/envs/lerobot/bin/lerobot-train
50
+ python: CPython 3.10.19
51
+ root: /home/vla/so101_1130_2tasks_pi0train
52
+ startedAt: "2025-11-30T11:55:54.544647Z"
53
+ writerId: ujjhup6u1if9spwto41rqzxwewcfoopf
54
+ m: []
55
+ python_version: 3.10.19
56
+ t:
57
+ "1":
58
+ - 1
59
+ - 11
60
+ - 41
61
+ - 49
62
+ - 51
63
+ - 71
64
+ - 83
65
+ "2":
66
+ - 1
67
+ - 11
68
+ - 41
69
+ - 49
70
+ - 51
71
+ - 71
72
+ - 83
73
+ "3":
74
+ - 13
75
+ - 15
76
+ - 16
77
+ - 61
78
+ "4": 3.10.19
79
+ "5": 0.21.4
80
+ "6": 4.53.3
81
+ "10":
82
+ - 21
83
+ "12": 0.21.4
84
+ "13": linux-x86_64
85
+ batch_size:
86
+ value: 24
87
+ checkpoint_path:
88
+ value: null
89
+ dataset:
90
+ value:
91
+ episodes: null
92
+ image_transforms:
93
+ enable: false
94
+ max_num_transforms: 3
95
+ random_order: false
96
+ tfs:
97
+ affine:
98
+ kwargs:
99
+ degrees:
100
+ - -5
101
+ - 5
102
+ translate:
103
+ - 0.05
104
+ - 0.05
105
+ type: RandomAffine
106
+ weight: 1
107
+ brightness:
108
+ kwargs:
109
+ brightness:
110
+ - 0.8
111
+ - 1.2
112
+ type: ColorJitter
113
+ weight: 1
114
+ contrast:
115
+ kwargs:
116
+ contrast:
117
+ - 0.8
118
+ - 1.2
119
+ type: ColorJitter
120
+ weight: 1
121
+ hue:
122
+ kwargs:
123
+ hue:
124
+ - -0.05
125
+ - 0.05
126
+ type: ColorJitter
127
+ weight: 1
128
+ saturation:
129
+ kwargs:
130
+ saturation:
131
+ - 0.5
132
+ - 1.5
133
+ type: ColorJitter
134
+ weight: 1
135
+ sharpness:
136
+ kwargs:
137
+ sharpness:
138
+ - 0.5
139
+ - 1.5
140
+ type: SharpnessJitter
141
+ weight: 1
142
+ repo_id: /home/vla/so101_1130_2tasks
143
+ revision: null
144
+ root: null
145
+ streaming: false
146
+ use_imagenet_stats: true
147
+ video_backend: torchcodec
148
+ env:
149
+ value: null
150
+ eval:
151
+ value:
152
+ batch_size: 50
153
+ n_episodes: 50
154
+ use_async_envs: false
155
+ eval_freq:
156
+ value: 20000
157
+ job_name:
158
+ value: so101_1130_2tasks_pi0train
159
+ log_freq:
160
+ value: 200
161
+ num_workers:
162
+ value: 16
163
+ optimizer:
164
+ value:
165
+ betas:
166
+ - 0.9
167
+ - 0.95
168
+ eps: 1e-08
169
+ grad_clip_norm: 1
170
+ lr: 2.5e-05
171
+ type: adamw
172
+ weight_decay: 0.01
173
+ output_dir:
174
+ value: /home/vla/so101_1130_2tasks_pi0train
175
+ policy:
176
+ value:
177
+ action_expert_variant: gemma_300m
178
+ chunk_size: 50
179
+ compile_mode: max-autotune
180
+ compile_model: true
181
+ device: cuda
182
+ dtype: bfloat16
183
+ empty_cameras: 0
184
+ gradient_checkpointing: true
185
+ image_resolution:
186
+ - 224
187
+ - 224
188
+ license: null
189
+ max_action_dim: 32
190
+ max_period: 4
191
+ max_state_dim: 32
192
+ min_period: 0.004
193
+ n_action_steps: 50
194
+ n_obs_steps: 1
195
+ normalization_mapping:
196
+ ACTION: MEAN_STD
197
+ STATE: MEAN_STD
198
+ VISUAL: IDENTITY
199
+ num_inference_steps: 10
200
+ optimizer_betas:
201
+ - 0.9
202
+ - 0.95
203
+ optimizer_eps: 1e-08
204
+ optimizer_grad_clip_norm: 1
205
+ optimizer_lr: 2.5e-05
206
+ optimizer_weight_decay: 0.01
207
+ paligemma_variant: gemma_2b
208
+ pretrained_path: lerobot/pi0_base
209
+ private: null
210
+ push_to_hub: false
211
+ repo_id: null
212
+ scheduler_decay_lr: 2.5e-06
213
+ scheduler_decay_steps: 30000
214
+ scheduler_warmup_steps: 1000
215
+ tags: null
216
+ time_sampling_beta_alpha: 1.5
217
+ time_sampling_beta_beta: 1
218
+ time_sampling_offset: 0.001
219
+ time_sampling_scale: 0.999
220
+ tokenizer_max_length: 48
221
+ type: pi0
222
+ use_amp: false
223
+ rename_map:
224
+ value:
225
+ observation.images.camera1: observation.images.wrist
226
+ observation.images.camera2: observation.images.top
227
+ observation.images.camera3: observation.images.front
228
+ resume:
229
+ value: false
230
+ save_checkpoint:
231
+ value: true
232
+ save_freq:
233
+ value: 20000
234
+ scheduler:
235
+ value:
236
+ decay_lr: 2.5e-06
237
+ num_decay_steps: 30000
238
+ num_warmup_steps: 1000
239
+ peak_lr: 2.5e-05
240
+ type: cosine_decay_with_warmup
241
+ seed:
242
+ value: 1000
243
+ steps:
244
+ value: 10000
245
+ use_policy_training_preset:
246
+ value: true
247
+ wandb:
248
+ value:
249
+ disable_artifact: false
250
+ enable: true
251
+ entity: null
252
+ mode: null
253
+ notes: null
254
+ project: lerobot
255
+ run_id: null
wandb/run-20251130_205554-fadl65t9/files/output.log ADDED
@@ -0,0 +1,145 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ INFO 2025-11-30 20:55:55 db_utils.py:102 Logs will be synced with wandb.
2
+ INFO 2025-11-30 20:55:55 db_utils.py:103 Track this run --> https://wandb.ai/wjdfufgusrn1234-konkuk-university/lerobot/runs/fadl65t9
3
+ INFO 2025-11-30 20:55:55 ot_train.py:183 Creating dataset
4
+ INFO 2025-11-30 20:55:56 ot_train.py:202 Creating policy
5
+ The PI0 model is a direct port of the OpenPI implementation.
6
+ This implementation follows the original OpenPI structure for compatibility.
7
+ Original implementation: https://github.com/Physical-Intelligence/openpi
8
+ INFO 2025-11-30 20:56:53 ling_pi0.py:553 Enabled gradient checkpointing for PI0Pytorch model
9
+ Loading model from: lerobot/pi0_base
10
+ ✓ Loaded state dict from model.safetensors
11
+ WARNING 2025-11-30 20:56:55 ing_pi0.py:1046 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.bias
12
+ WARNING 2025-11-30 20:56:55 ing_pi0.py:1046 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.weight
13
+ Remapped 777 state dict keys
14
+ Warning: Could not remap state dict keys: Error(s) in loading state_dict for PI0Policy:
15
+ Missing key(s) in state_dict: "model.paligemma_with_expert.paligemma.model.language_model.embed_tokens.weight".
16
+ INFO 2025-11-30 20:57:01 ot_train.py:247 Creating optimizer and scheduler
17
+ INFO 2025-11-30 20:57:01 hedulers.py:105 Auto-scaling LR scheduler: num_training_steps (10000) < num_decay_steps (30000). Scaling warmup: 1000 → 333, decay: 30000 → 10000 (scale factor: 0.333)
18
+ INFO 2025-11-30 20:57:01 ot_train.py:259 Output dir: /home/vla/so101_1130_2tasks_pi0train
19
+ INFO 2025-11-30 20:57:01 ot_train.py:262 cfg.steps=10000 (10K)
20
+ INFO 2025-11-30 20:57:01 ot_train.py:263 dataset.num_frames=68067 (68K)
21
+ INFO 2025-11-30 20:57:01 ot_train.py:264 dataset.num_episodes=41
22
+ INFO 2025-11-30 20:57:01 ot_train.py:267 Effective batch size: 24 x 2 = 48
23
+ INFO 2025-11-30 20:57:01 ot_train.py:268 num_learnable_params=3501372176 (4B)
24
+ INFO 2025-11-30 20:57:01 ot_train.py:269 num_total_params=3501372176 (4B)
25
+ INFO 2025-11-30 20:57:01 ot_train.py:324 Start offline training on a fixed dataset
26
+ [rank0]:W1130 20:57:08.754000 2873154 site-packages/torch/_logging/_internal.py:1130] [0/0] Profiler function <class 'torch.autograd.profiler.record_function'> will be ignored
27
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] Graph break from `Tensor.item()`, consider setting:
28
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] torch._dynamo.config.capture_scalar_outputs = True
29
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] or:
30
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] env TORCHDYNAMO_CAPTURE_SCALAR_OUTPUTS=1
31
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] to include these operations in the captured graph.
32
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
33
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] Graph break: from user code at:
34
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/accelerate/utils/operations.py", line 819, in forward
35
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return model_forward(*args, **kwargs)
36
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/accelerate/utils/operations.py", line 807, in __call__
37
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return convert_to_fp32(self.model_forward(*args, **kwargs))
38
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/miniforge3/envs/lerobot/lib/python3.10/site-packages/torch/amp/autocast_mode.py", line 44, in decorate_autocast
39
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] return func(*args, **kwargs)
40
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] File "/home/vla/lerobot/src/lerobot/policies/pi0/modeling_pi0.py", line 1188, in forward
41
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0] "loss": loss.item(),
42
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
43
+ [rank0]:W1130 20:57:19.655000 2873154 site-packages/torch/_dynamo/variables/tensor.py:913] [4/0]
44
+ INFO 2025-11-30 21:15:07 ot_train.py:351 step:200 smpl:19K ep:12 epch:0.28 loss:0.416 grdn:5.439 lr:7.6e-06 updt_s:5.372 data_s:0.056
45
+ WARNING 2025-11-30 21:15:07 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
46
+ INFO 2025-11-30 21:29:45 ot_train.py:351 step:400 smpl:38K ep:23 epch:0.56 loss:0.184 grdn:2.688 lr:2.2e-05 updt_s:4.369 data_s:0.022
47
+ WARNING 2025-11-30 21:29:45 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
48
+ INFO 2025-11-30 21:44:24 ot_train.py:351 step:600 smpl:58K ep:35 epch:0.85 loss:0.154 grdn:1.554 lr:2.5e-05 updt_s:4.370 data_s:0.022
49
+ WARNING 2025-11-30 21:44:24 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
50
+ INFO 2025-11-30 21:59:04 ot_train.py:351 step:800 smpl:77K ep:46 epch:1.13 loss:0.139 grdn:1.217 lr:2.5e-05 updt_s:4.378 data_s:0.022
51
+ WARNING 2025-11-30 21:59:04 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
52
+ INFO 2025-11-30 22:13:42 ot_train.py:351 step:1K smpl:96K ep:58 epch:1.41 loss:0.135 grdn:1.141 lr:2.5e-05 updt_s:4.368 data_s:0.021
53
+ WARNING 2025-11-30 22:13:42 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
54
+ INFO 2025-11-30 22:28:22 ot_train.py:351 step:1K smpl:115K ep:69 epch:1.69 loss:0.129 grdn:1.079 lr:2.4e-05 updt_s:4.377 data_s:0.021
55
+ WARNING 2025-11-30 22:28:22 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
56
+ INFO 2025-11-30 22:43:01 ot_train.py:351 step:1K smpl:134K ep:81 epch:1.97 loss:0.121 grdn:1.039 lr:2.4e-05 updt_s:4.375 data_s:0.021
57
+ WARNING 2025-11-30 22:43:01 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
58
+ INFO 2025-11-30 22:57:46 ot_train.py:351 step:2K smpl:154K ep:93 epch:2.26 loss:0.119 grdn:0.997 lr:2.4e-05 updt_s:4.376 data_s:0.051
59
+ WARNING 2025-11-30 22:57:46 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
60
+ INFO 2025-11-30 23:12:24 ot_train.py:351 step:2K smpl:173K ep:104 epch:2.54 loss:0.119 grdn:0.958 lr:2.3e-05 updt_s:4.369 data_s:0.022
61
+ WARNING 2025-11-30 23:12:24 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
62
+ INFO 2025-11-30 23:27:03 ot_train.py:351 step:2K smpl:192K ep:116 epch:2.82 loss:0.116 grdn:0.949 lr:2.3e-05 updt_s:4.369 data_s:0.022
63
+ WARNING 2025-11-30 23:27:03 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
64
+ INFO 2025-11-30 23:41:42 ot_train.py:351 step:2K smpl:211K ep:127 epch:3.10 loss:0.114 grdn:0.919 lr:2.3e-05 updt_s:4.374 data_s:0.022
65
+ WARNING 2025-11-30 23:41:42 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
66
+ INFO 2025-11-30 23:56:20 ot_train.py:351 step:2K smpl:230K ep:139 epch:3.38 loss:0.107 grdn:0.917 lr:2.2e-05 updt_s:4.367 data_s:0.021
67
+ WARNING 2025-11-30 23:56:20 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
68
+ INFO 2025-12-01 00:10:57 ot_train.py:351 step:3K smpl:250K ep:150 epch:3.67 loss:0.111 grdn:0.933 lr:2.2e-05 updt_s:4.366 data_s:0.021
69
+ WARNING 2025-12-01 00:10:57 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
70
+ INFO 2025-12-01 00:25:36 ot_train.py:351 step:3K smpl:269K ep:162 epch:3.95 loss:0.107 grdn:0.908 lr:2.1e-05 updt_s:4.369 data_s:0.021
71
+ WARNING 2025-12-01 00:25:36 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
72
+ INFO 2025-12-01 00:40:20 ot_train.py:351 step:3K smpl:288K ep:173 epch:4.23 loss:0.103 grdn:0.889 lr:2.1e-05 updt_s:4.373 data_s:0.049
73
+ WARNING 2025-12-01 00:40:20 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
74
+ INFO 2025-12-01 00:54:58 ot_train.py:351 step:3K smpl:307K ep:185 epch:4.51 loss:0.099 grdn:0.903 lr:2.0e-05 updt_s:4.367 data_s:0.021
75
+ WARNING 2025-12-01 00:54:58 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
76
+ INFO 2025-12-01 01:09:36 ot_train.py:351 step:3K smpl:326K ep:197 epch:4.80 loss:0.099 grdn:0.892 lr:1.9e-05 updt_s:4.367 data_s:0.021
77
+ WARNING 2025-12-01 01:09:36 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
78
+ INFO 2025-12-01 01:24:14 ot_train.py:351 step:4K smpl:346K ep:208 epch:5.08 loss:0.093 grdn:0.882 lr:1.9e-05 updt_s:4.372 data_s:0.022
79
+ WARNING 2025-12-01 01:24:14 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
80
+ INFO 2025-12-01 01:38:54 ot_train.py:351 step:4K smpl:365K ep:220 epch:5.36 loss:0.094 grdn:0.909 lr:1.8e-05 updt_s:4.377 data_s:0.021
81
+ WARNING 2025-12-01 01:38:54 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
82
+ INFO 2025-12-01 01:53:32 ot_train.py:351 step:4K smpl:384K ep:231 epch:5.64 loss:0.094 grdn:0.911 lr:1.8e-05 updt_s:4.367 data_s:0.021
83
+ WARNING 2025-12-01 01:53:32 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
84
+ INFO 2025-12-01 02:08:09 ot_train.py:351 step:4K smpl:403K ep:243 epch:5.92 loss:0.095 grdn:0.915 lr:1.7e-05 updt_s:4.366 data_s:0.021
85
+ WARNING 2025-12-01 02:08:09 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
86
+ INFO 2025-12-01 02:22:58 ot_train.py:351 step:4K smpl:422K ep:254 epch:6.21 loss:0.094 grdn:0.933 lr:1.6e-05 updt_s:4.391 data_s:0.050
87
+ WARNING 2025-12-01 02:22:58 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
88
+ INFO 2025-12-01 02:37:35 ot_train.py:351 step:5K smpl:442K ep:266 epch:6.49 loss:0.097 grdn:0.905 lr:1.6e-05 updt_s:4.365 data_s:0.021
89
+ WARNING 2025-12-01 02:37:35 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
90
+ INFO 2025-12-01 02:52:13 ot_train.py:351 step:5K smpl:461K ep:278 epch:6.77 loss:0.092 grdn:0.935 lr:1.5e-05 updt_s:4.367 data_s:0.021
91
+ WARNING 2025-12-01 02:52:13 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
92
+ INFO 2025-12-01 03:06:51 ot_train.py:351 step:5K smpl:480K ep:289 epch:7.05 loss:0.091 grdn:0.929 lr:1.4e-05 updt_s:4.370 data_s:0.022
93
+ WARNING 2025-12-01 03:06:51 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
94
+ INFO 2025-12-01 03:21:29 ot_train.py:351 step:5K smpl:499K ep:301 epch:7.33 loss:0.088 grdn:0.929 lr:1.3e-05 updt_s:4.365 data_s:0.021
95
+ WARNING 2025-12-01 03:21:29 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
96
+ INFO 2025-12-01 03:36:06 ot_train.py:351 step:5K smpl:518K ep:312 epch:7.62 loss:0.088 grdn:0.986 lr:1.3e-05 updt_s:4.365 data_s:0.021
97
+ WARNING 2025-12-01 03:36:06 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
98
+ INFO 2025-12-01 03:50:47 ot_train.py:351 step:6K smpl:538K ep:324 epch:7.90 loss:0.087 grdn:0.996 lr:1.2e-05 updt_s:4.384 data_s:0.022
99
+ WARNING 2025-12-01 03:50:47 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
100
+ INFO 2025-12-01 04:05:30 ot_train.py:351 step:6K smpl:557K ep:335 epch:8.18 loss:0.089 grdn:0.965 lr:1.1e-05 updt_s:4.365 data_s:0.049
101
+ WARNING 2025-12-01 04:05:30 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
102
+ INFO 2025-12-01 04:20:08 ot_train.py:351 step:6K smpl:576K ep:347 epch:8.46 loss:0.087 grdn:1.031 lr:1.1e-05 updt_s:4.365 data_s:0.021
103
+ WARNING 2025-12-01 04:20:08 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
104
+ INFO 2025-12-01 04:34:45 ot_train.py:351 step:6K smpl:595K ep:359 epch:8.74 loss:0.080 grdn:0.967 lr:9.9e-06 updt_s:4.366 data_s:0.022
105
+ WARNING 2025-12-01 04:34:45 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
106
+ INFO 2025-12-01 04:49:24 ot_train.py:351 step:6K smpl:614K ep:370 epch:9.03 loss:0.085 grdn:0.993 lr:9.3e-06 updt_s:4.374 data_s:0.021
107
+ WARNING 2025-12-01 04:49:24 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
108
+ INFO 2025-12-01 05:04:04 ot_train.py:351 step:7K smpl:634K ep:382 epch:9.31 loss:0.091 grdn:1.040 lr:8.6e-06 updt_s:4.378 data_s:0.022
109
+ WARNING 2025-12-01 05:04:04 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
110
+ INFO 2025-12-01 05:18:42 ot_train.py:351 step:7K smpl:653K ep:393 epch:9.59 loss:0.083 grdn:1.055 lr:8.0e-06 updt_s:4.366 data_s:0.021
111
+ WARNING 2025-12-01 05:18:42 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
112
+ INFO 2025-12-01 05:33:21 ot_train.py:351 step:7K smpl:672K ep:405 epch:9.87 loss:0.084 grdn:1.071 lr:7.4e-06 updt_s:4.371 data_s:0.022
113
+ WARNING 2025-12-01 05:33:21 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
114
+ INFO 2025-12-01 05:48:06 ot_train.py:351 step:7K smpl:691K ep:416 epch:10.15 loss:0.083 grdn:1.082 lr:6.9e-06 updt_s:4.375 data_s:0.049
115
+ WARNING 2025-12-01 05:48:06 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
116
+ INFO 2025-12-01 06:02:43 ot_train.py:351 step:7K smpl:710K ep:428 epch:10.44 loss:0.082 grdn:1.067 lr:6.3e-06 updt_s:4.366 data_s:0.022
117
+ WARNING 2025-12-01 06:02:43 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
118
+ INFO 2025-12-01 06:17:22 ot_train.py:351 step:8K smpl:730K ep:439 epch:10.72 loss:0.082 grdn:1.072 lr:5.8e-06 updt_s:4.373 data_s:0.022
119
+ WARNING 2025-12-01 06:17:22 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
120
+ INFO 2025-12-01 06:32:00 ot_train.py:351 step:8K smpl:749K ep:451 epch:11.00 loss:0.077 grdn:1.102 lr:5.3e-06 updt_s:4.366 data_s:0.022
121
+ WARNING 2025-12-01 06:32:00 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
122
+ INFO 2025-12-01 06:46:38 ot_train.py:351 step:8K smpl:768K ep:463 epch:11.28 loss:0.080 grdn:1.120 lr:4.9e-06 updt_s:4.367 data_s:0.021
123
+ WARNING 2025-12-01 06:46:38 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
124
+ INFO 2025-12-01 07:01:17 ot_train.py:351 step:8K smpl:787K ep:474 epch:11.57 loss:0.081 grdn:1.095 lr:4.4e-06 updt_s:4.376 data_s:0.022
125
+ WARNING 2025-12-01 07:01:17 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
126
+ INFO 2025-12-01 07:15:56 ot_train.py:351 step:8K smpl:806K ep:486 epch:11.85 loss:0.085 grdn:1.109 lr:4.1e-06 updt_s:4.370 data_s:0.022
127
+ WARNING 2025-12-01 07:15:56 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
128
+ INFO 2025-12-01 07:30:39 ot_train.py:351 step:9K smpl:826K ep:497 epch:12.13 loss:0.083 grdn:1.097 lr:3.7e-06 updt_s:4.366 data_s:0.050
129
+ WARNING 2025-12-01 07:30:39 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
130
+ INFO 2025-12-01 07:45:23 ot_train.py:351 step:9K smpl:845K ep:509 epch:12.41 loss:0.087 grdn:1.147 lr:3.4e-06 updt_s:4.396 data_s:0.025
131
+ WARNING 2025-12-01 07:45:23 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
132
+ INFO 2025-12-01 08:00:08 ot_train.py:351 step:9K smpl:864K ep:520 epch:12.69 loss:0.079 grdn:1.187 lr:3.2e-06 updt_s:4.400 data_s:0.023
133
+ WARNING 2025-12-01 08:00:08 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
134
+ INFO 2025-12-01 08:14:46 ot_train.py:351 step:9K smpl:883K ep:532 epch:12.98 loss:0.078 grdn:1.163 lr:2.9e-06 updt_s:4.367 data_s:0.021
135
+ WARNING 2025-12-01 08:14:46 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
136
+ INFO 2025-12-01 08:29:23 ot_train.py:351 step:9K smpl:902K ep:544 epch:13.26 loss:0.078 grdn:1.109 lr:2.8e-06 updt_s:4.366 data_s:0.020
137
+ WARNING 2025-12-01 08:29:23 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
138
+ INFO 2025-12-01 08:44:00 ot_train.py:351 step:10K smpl:922K ep:555 epch:13.54 loss:0.088 grdn:1.169 lr:2.6e-06 updt_s:4.364 data_s:0.022
139
+ WARNING 2025-12-01 08:44:00 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
140
+ INFO 2025-12-01 08:58:41 ot_train.py:351 step:10K smpl:941K ep:567 epch:13.82 loss:0.081 grdn:1.151 lr:2.6e-06 updt_s:4.385 data_s:0.021
141
+ WARNING 2025-12-01 08:58:41 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
142
+ INFO 2025-12-01 09:13:24 ot_train.py:351 step:10K smpl:960K ep:578 epch:14.10 loss:0.075 grdn:1.127 lr:2.5e-06 updt_s:4.364 data_s:0.047
143
+ WARNING 2025-12-01 09:13:24 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
144
+ INFO 2025-12-01 09:13:24 ot_train.py:361 Checkpoint policy after step 10000
145
+ INFO 2025-12-01 09:14:22 ot_train.py:430 End of training
wandb/run-20251130_205554-fadl65t9/files/requirements.txt ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ h11==0.16.0
2
+ wcwidth==0.2.14
3
+ nvidia-cuda-cupti-cu12==12.6.80
4
+ async-timeout==5.0.1
5
+ dill==0.4.0
6
+ typing_extensions==4.15.0
7
+ pytz==2025.2
8
+ cloudpickle==3.1.2
9
+ pydantic==2.12.4
10
+ python-xlib==0.33
11
+ Jinja2==3.1.6
12
+ gmpy2==2.2.1
13
+ aiohappyeyeballs==2.6.1
14
+ einops==0.8.1
15
+ mpmath==1.3.0
16
+ hyperframe==6.1.0
17
+ smmap==5.0.2
18
+ mergedeep==1.3.4
19
+ Farama-Notifications==0.0.4
20
+ shellingham==1.5.4
21
+ yarl==1.22.0
22
+ peft==0.18.0
23
+ regex==2025.11.3
24
+ deepdiff==8.6.1
25
+ nvidia-cufile-cu12==1.11.1.6
26
+ certifi==2025.11.12
27
+ multiprocess==0.70.16
28
+ cmake==4.1.2
29
+ MarkupSafe==3.0.3
30
+ aiohttp==3.13.2
31
+ psutil==7.1.3
32
+ python-dateutil==2.9.0.post0
33
+ h2==4.3.0
34
+ httpcore==1.0.9
35
+ nvidia-cusparse-cu12==12.5.4.2
36
+ click==8.3.0
37
+ frozenlist==1.8.0
38
+ av==15.1.0
39
+ urllib3==2.5.0
40
+ sentry-sdk==2.44.0
41
+ pynput==1.8.1
42
+ pandas==2.3.3
43
+ triton==3.3.1
44
+ aiosignal==1.4.0
45
+ multidict==6.7.0
46
+ mypy_extensions==1.1.0
47
+ transformers==4.53.3
48
+ anyio==4.11.0
49
+ lerobot==0.4.1
50
+ zipp==3.23.0
51
+ platformdirs==4.5.0
52
+ pydantic_core==2.41.5
53
+ PyYAML==6.0.3
54
+ orderly-set==5.5.0
55
+ protobuf==6.33.0
56
+ typing-inspection==0.4.2
57
+ sympy==1.14.0
58
+ flash_attn==2.8.3
59
+ wheel==0.45.1
60
+ attrs==25.4.0
61
+ tqdm==4.67.1
62
+ nvidia-cudnn-cu12==9.5.1.17
63
+ Brotli==1.2.0
64
+ cffi==2.0.0
65
+ fsspec==2025.9.0
66
+ gitdb==4.0.12
67
+ importlib_metadata==8.7.0
68
+ nvidia-cusparselt-cu12==0.6.3
69
+ nvidia-cublas-cu12==12.6.4.1
70
+ opencv-python-headless==4.12.0.88
71
+ exceptiongroup==1.3.0
72
+ filelock==3.20.0
73
+ safetensors==0.6.2
74
+ pyserial==3.5
75
+ annotated-types==0.7.0
76
+ nvidia-cufft-cu12==11.3.0.4
77
+ pyyaml-include==1.4.1
78
+ PySocks==1.7.1
79
+ ImageIO==2.37.2
80
+ nvidia-cusolver-cu12==11.7.1.2
81
+ datasets==4.1.1
82
+ hf_transfer==0.1.9
83
+ networkx==3.4.2
84
+ docopt==0.6.2
85
+ tzdata==2025.2
86
+ num2words==0.5.14
87
+ hf-xet==1.2.0
88
+ nvidia-nccl-cu12==2.26.2
89
+ sniffio==1.3.1
90
+ huggingface-hub==0.35.3
91
+ xxhash==3.6.0
92
+ requests==2.32.5
93
+ idna==3.11
94
+ httpx==0.28.1
95
+ torch==2.7.1
96
+ typing-inspect==0.9.0
97
+ hpack==4.1.0
98
+ tokenizers==0.21.4
99
+ draccus==0.10.0
100
+ charset-normalizer==3.4.4
101
+ evdev==1.9.2
102
+ zstandard==0.25.0
103
+ pfzy==0.3.4
104
+ pillow==12.0.0
105
+ pycparser==2.22
106
+ torchvision==0.22.1
107
+ termcolor==3.2.0
108
+ wandb==0.21.4
109
+ nvidia-cuda-runtime-cu12==12.6.77
110
+ typer-slim==0.20.0
111
+ rerun-sdk==0.26.2
112
+ gymnasium==1.2.2
113
+ pyarrow==22.0.0
114
+ six==1.17.0
115
+ nvidia-nvjitlink-cu12==12.6.85
116
+ GitPython==3.1.45
117
+ prompt_toolkit==3.0.52
118
+ imageio-ffmpeg==0.6.0
119
+ nvidia-nvtx-cu12==12.6.77
120
+ torchcodec==0.5
121
+ nvidia-curand-cu12==10.3.7.77
122
+ diffusers==0.35.2
123
+ propcache==0.4.1
124
+ packaging==25.0
125
+ setuptools==80.9.0
126
+ jsonlines==4.0.0
127
+ numpy==2.2.6
128
+ pip==25.3
129
+ nvidia-cuda-nvrtc-cu12==12.6.77
130
+ accelerate==1.11.0
131
+ inquirerpy==0.3.4
132
+ toml==0.10.2
133
+ lerobot==0.4.1
wandb/run-20251130_205554-fadl65t9/files/wandb-metadata.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-88-generic-x86_64-with-glibc2.39",
3
+ "python": "CPython 3.10.19",
4
+ "startedAt": "2025-11-30T11:55:54.544647Z",
5
+ "args": [
6
+ "--dataset.repo_id=/home/vla/so101_1130_2tasks",
7
+ "--policy.type=pi0",
8
+ "--policy.pretrained_path=lerobot/pi0_base",
9
+ "--output_dir=/home/vla/so101_1130_2tasks_pi0train",
10
+ "--job_name=so101_1130_2tasks_pi0train",
11
+ "--steps=10000",
12
+ "--batch_size=24",
13
+ "--num_workers=16",
14
+ "--policy.device=cuda",
15
+ "--wandb.enable=true",
16
+ "--policy.compile_model=true",
17
+ "--policy.gradient_checkpointing=true",
18
+ "--policy.dtype=bfloat16",
19
+ "--policy.push_to_hub=false",
20
+ "--rename_map={\"observation.images.camera1\": \"observation.images.wrist\", \"observation.images.camera2\": \"observation.images.top\", \"observation.images.camera3\": \"observation.images.front\"}"
21
+ ],
22
+ "program": "/home/vla/miniforge3/envs/lerobot/bin/lerobot-train",
23
+ "codePath": "miniforge3/envs/lerobot/bin/lerobot-train",
24
+ "codePathLocal": "miniforge3/envs/lerobot/bin/lerobot-train",
25
+ "root": "/home/vla/so101_1130_2tasks_pi0train",
26
+ "host": "edu02",
27
+ "executable": "/home/vla/miniforge3/envs/lerobot/bin/python3.10",
28
+ "cpu_count": 56,
29
+ "cpu_count_logical": 112,
30
+ "gpu": "NVIDIA RTX A6000",
31
+ "gpu_count": 2,
32
+ "disk": {
33
+ "/": {
34
+ "total": "943412031488",
35
+ "used": "781631213568"
36
+ }
37
+ },
38
+ "memory": {
39
+ "total": "540628525056"
40
+ },
41
+ "gpu_nvidia": [
42
+ {
43
+ "name": "NVIDIA RTX A6000",
44
+ "memoryTotal": "51527024640",
45
+ "cudaCores": 10752,
46
+ "architecture": "Ampere",
47
+ "uuid": "GPU-c895b718-8738-3fff-b42c-92cad284b785"
48
+ },
49
+ {
50
+ "name": "NVIDIA RTX A6000",
51
+ "memoryTotal": "51527024640",
52
+ "cudaCores": 10752,
53
+ "architecture": "Ampere",
54
+ "uuid": "GPU-34849e7e-d621-dccc-ca70-b463192b3472"
55
+ }
56
+ ],
57
+ "cudaVersion": "12.8",
58
+ "writerId": "ujjhup6u1if9spwto41rqzxwewcfoopf"
59
+ }
wandb/run-20251130_205554-fadl65t9/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_runtime":44308.776981773,"train/lr":2.507345336505503e-06,"train/episodes":578.2537793644497,"_wandb":{"runtime":44308},"train/grad_norm":1.126771520972252,"train/dataloading_s":0.046947422818047924,"train/samples":960000,"train/loss":0.014316217973828316,"train/update_s":4.364254662292661,"_step":10000,"train/steps":10000,"_timestamp":1.764548004236067e+09,"train/epochs":14.103750716206092}
wandb/run-20251130_205554-fadl65t9/logs/debug-core.log ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-30T20:55:54.583678337+09:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpupgd6xqw/port-2873154.txt","pid":2873154,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false}
2
+ {"time":"2025-11-30T20:55:54.584770568+09:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2873154}
3
+ {"time":"2025-11-30T20:55:54.584640749+09:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2873154-2873179-3748399540/socket","Net":"unix"}}
4
+ {"time":"2025-11-30T20:55:54.75178464+09:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"}
5
+ {"time":"2025-11-30T20:55:54.770103443+09:00","level":"INFO","msg":"handleInformInit: received","streamId":"fadl65t9","id":"1(@)"}
6
+ {"time":"2025-11-30T20:55:55.192280145+09:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"fadl65t9","id":"1(@)"}
7
+ {"time":"2025-12-01T09:14:24.381745908+09:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"}
8
+ {"time":"2025-12-01T09:14:24.382025501+09:00","level":"INFO","msg":"server is shutting down"}
9
+ {"time":"2025-12-01T09:14:24.382016638+09:00","level":"INFO","msg":"connection: closing","id":"1(@)"}
10
+ {"time":"2025-12-01T09:14:24.382211502+09:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"}
11
+ {"time":"2025-12-01T09:14:24.382186003+09:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2873154-2873179-3748399540/socket","Net":"unix"}}
12
+ {"time":"2025-12-01T09:15:33.167908876+09:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write unix /tmp/wandb-2873154-2873179-3748399540/socket->@: use of closed network connection","id":"1(@)"}
13
+ {"time":"2025-12-01T09:15:33.62398672+09:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"}
14
+ {"time":"2025-12-01T09:15:33.624071897+09:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"}
15
+ {"time":"2025-12-01T09:15:33.624112637+09:00","level":"INFO","msg":"server is closed"}
wandb/run-20251130_205554-fadl65t9/logs/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-11-30T20:55:54.770291237+09:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
2
+ {"time":"2025-11-30T20:55:55.192165852+09:00","level":"INFO","msg":"stream: created new stream","id":"fadl65t9"}
3
+ {"time":"2025-11-30T20:55:55.192248512+09:00","level":"INFO","msg":"stream: started","id":"fadl65t9"}
4
+ {"time":"2025-11-30T20:55:55.19229908+09:00","level":"INFO","msg":"writer: started","stream_id":"fadl65t9"}
5
+ {"time":"2025-11-30T20:55:55.192297075+09:00","level":"INFO","msg":"handler: started","stream_id":"fadl65t9"}
6
+ {"time":"2025-11-30T20:55:55.192405884+09:00","level":"INFO","msg":"sender: started","stream_id":"fadl65t9"}
7
+ {"time":"2025-11-30T23:08:40.702759164+09:00","level":"INFO","msg":"api: retrying error","error":"Post \"https://api.wandb.ai/files/wjdfufgusrn1234-konkuk-university/lerobot/fadl65t9/file_stream\": dial tcp 35.186.228.49:443: connect: connection timed out"}
8
+ {"time":"2025-12-01T09:14:24.3819952+09:00","level":"INFO","msg":"stream: closing","id":"fadl65t9"}
9
+ {"time":"2025-12-01T09:15:33.217839615+09:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
10
+ {"time":"2025-12-01T09:15:33.622503049+09:00","level":"INFO","msg":"handler: closed","stream_id":"fadl65t9"}
11
+ {"time":"2025-12-01T09:15:33.622701638+09:00","level":"INFO","msg":"sender: closed","stream_id":"fadl65t9"}
12
+ {"time":"2025-12-01T09:15:33.622736417+09:00","level":"INFO","msg":"stream: closed","id":"fadl65t9"}
wandb/run-20251130_205554-fadl65t9/logs/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
2
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Configure stats pid to 2873154
3
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Loading settings from /home/vla/.config/wandb/settings
4
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Loading settings from /home/vla/wandb/settings
5
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /home/vla/so101_1130_2tasks_pi0train/wandb/run-20251130_205554-fadl65t9/logs/debug.log
7
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /home/vla/so101_1130_2tasks_pi0train/wandb/run-20251130_205554-fadl65t9/logs/debug-internal.log
8
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:init():813] calling init triggers
9
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': '/home/vla/so101_1130_2tasks', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi0', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': None, 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi0_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'image_resolution': [224, 224], 'empty_cameras': 0, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.MEAN_STD: 'MEAN_STD'>, 'ACTION': <NormalizationMode.MEAN_STD: 'MEAN_STD'>}, 'gradient_checkpointing': True, 'compile_model': True, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06, 'tokenizer_max_length': 48}, 'output_dir': '/home/vla/so101_1130_2tasks_pi0train', 'job_name': 'so101_1130_2tasks_pi0train', 'resume': False, 'seed': 1000, 'num_workers': 16, 'batch_size': 24, 'steps': 10000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {'observation.images.camera1': 'observation.images.wrist', 'observation.images.camera2': 'observation.images.top', 'observation.images.camera3': 'observation.images.front'}, '_wandb': {}}
11
+ 2025-11-30 20:55:54,545 INFO MainThread:2873154 [wandb_init.py:init():854] starting backend
12
+ 2025-11-30 20:55:54,752 INFO MainThread:2873154 [wandb_init.py:init():857] sending inform_init request
13
+ 2025-11-30 20:55:54,759 INFO MainThread:2873154 [wandb_init.py:init():865] backend started and connected
14
+ 2025-11-30 20:55:54,765 INFO MainThread:2873154 [wandb_init.py:init():936] updated telemetry
15
+ 2025-11-30 20:55:54,766 INFO MainThread:2873154 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
16
+ 2025-11-30 20:55:55,602 INFO MainThread:2873154 [wandb_init.py:init():1011] starting run threads in backend
17
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-11-30 20:55:55,714 INFO MainThread:2873154 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-11-30 20:55:55,717 INFO MainThread:2873154 [wandb_init.py:init():1049] run started, returning control to user process
22
+ 2025-12-01 09:14:24,382 INFO wandb-AsyncioManager-main:2873154 [service_client.py:_forward_responses():84] Reached EOF.
23
+ 2025-12-01 09:14:24,382 INFO wandb-AsyncioManager-main:2873154 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles.
wandb/run-20251130_205554-fadl65t9/run-fadl65t9.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:123d4b0a3ef7213230544a4400811ab9d1c651e8ab4f7d4ff3c5d351a0bf0161
3
+ size 3493520