Zasha01 commited on
Commit
b1b4794
·
verified ·
1 Parent(s): 7ff272f

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ wandb/run-20251213_112024-hxte4s1l/run-hxte4s1l.wandb filter=lfs diff=lfs merge=lfs -text
checkpoints/003000/pretrained_model/config.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "pi05",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 14
9
+ ]
10
+ },
11
+ "observation.images.context": {
12
+ "type": "VISUAL",
13
+ "shape": [
14
+ 3,
15
+ 360,
16
+ 640
17
+ ]
18
+ },
19
+ "observation.images.right_wrist": {
20
+ "type": "VISUAL",
21
+ "shape": [
22
+ 3,
23
+ 360,
24
+ 640
25
+ ]
26
+ },
27
+ "observation.images.left_wrist": {
28
+ "type": "VISUAL",
29
+ "shape": [
30
+ 3,
31
+ 360,
32
+ 640
33
+ ]
34
+ }
35
+ },
36
+ "output_features": {
37
+ "action": {
38
+ "type": "ACTION",
39
+ "shape": [
40
+ 14
41
+ ]
42
+ }
43
+ },
44
+ "device": "cuda",
45
+ "use_amp": false,
46
+ "push_to_hub": false,
47
+ "repo_id": "local/pi05",
48
+ "private": null,
49
+ "tags": null,
50
+ "license": null,
51
+ "pretrained_path": "lerobot/pi05_base",
52
+ "paligemma_variant": "gemma_2b",
53
+ "action_expert_variant": "gemma_300m",
54
+ "dtype": "bfloat16",
55
+ "chunk_size": 50,
56
+ "n_action_steps": 50,
57
+ "max_state_dim": 32,
58
+ "max_action_dim": 32,
59
+ "num_inference_steps": 10,
60
+ "time_sampling_beta_alpha": 1.5,
61
+ "time_sampling_beta_beta": 1.0,
62
+ "time_sampling_scale": 0.999,
63
+ "time_sampling_offset": 0.001,
64
+ "min_period": 0.004,
65
+ "max_period": 4.0,
66
+ "rtc_config": null,
67
+ "image_resolution": [
68
+ 224,
69
+ 224
70
+ ],
71
+ "empty_cameras": 0,
72
+ "tokenizer_max_length": 200,
73
+ "normalization_mapping": {
74
+ "VISUAL": "IDENTITY",
75
+ "STATE": "QUANTILES",
76
+ "ACTION": "QUANTILES"
77
+ },
78
+ "gradient_checkpointing": true,
79
+ "compile_model": false,
80
+ "compile_mode": "max-autotune",
81
+ "optimizer_lr": 2.5e-05,
82
+ "optimizer_betas": [
83
+ 0.9,
84
+ 0.95
85
+ ],
86
+ "optimizer_eps": 1e-08,
87
+ "optimizer_weight_decay": 0.01,
88
+ "optimizer_grad_clip_norm": 1.0,
89
+ "scheduler_warmup_steps": 1000,
90
+ "scheduler_decay_steps": 30000,
91
+ "scheduler_decay_lr": 2.5e-06
92
+ }
checkpoints/003000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ac4ff5a16df193bc4d79cdc12b2712e6fdb5ccd7ad351e4daf4a3172aa324d5
3
+ size 7473096344
checkpoints/003000/pretrained_model/policy_postprocessor.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_postprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "unnormalizer_processor",
6
+ "config": {
7
+ "eps": 1e-08,
8
+ "features": {
9
+ "action": {
10
+ "type": "ACTION",
11
+ "shape": [
12
+ 14
13
+ ]
14
+ }
15
+ },
16
+ "norm_map": {
17
+ "VISUAL": "IDENTITY",
18
+ "STATE": "QUANTILES",
19
+ "ACTION": "QUANTILES"
20
+ }
21
+ },
22
+ "state_file": "policy_postprocessor_step_0_unnormalizer_processor.safetensors"
23
+ },
24
+ {
25
+ "registry_name": "device_processor",
26
+ "config": {
27
+ "device": "cpu",
28
+ "float_dtype": null
29
+ }
30
+ }
31
+ ]
32
+ }
checkpoints/003000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f99aec40164823a2f22a0ec5032e767eaa10338e84e19744d066a2daef88c7d0
3
+ size 9328
checkpoints/003000/pretrained_model/policy_preprocessor.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_preprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "rename_observations_processor",
6
+ "config": {
7
+ "rename_map": {}
8
+ }
9
+ },
10
+ {
11
+ "registry_name": "to_batch_processor",
12
+ "config": {}
13
+ },
14
+ {
15
+ "registry_name": "normalizer_processor",
16
+ "config": {
17
+ "eps": 1e-08,
18
+ "features": {
19
+ "observation.state": {
20
+ "type": "STATE",
21
+ "shape": [
22
+ 14
23
+ ]
24
+ },
25
+ "observation.images.context": {
26
+ "type": "VISUAL",
27
+ "shape": [
28
+ 3,
29
+ 360,
30
+ 640
31
+ ]
32
+ },
33
+ "observation.images.right_wrist": {
34
+ "type": "VISUAL",
35
+ "shape": [
36
+ 3,
37
+ 360,
38
+ 640
39
+ ]
40
+ },
41
+ "observation.images.left_wrist": {
42
+ "type": "VISUAL",
43
+ "shape": [
44
+ 3,
45
+ 360,
46
+ 640
47
+ ]
48
+ },
49
+ "action": {
50
+ "type": "ACTION",
51
+ "shape": [
52
+ 14
53
+ ]
54
+ }
55
+ },
56
+ "norm_map": {
57
+ "VISUAL": "IDENTITY",
58
+ "STATE": "QUANTILES",
59
+ "ACTION": "QUANTILES"
60
+ }
61
+ },
62
+ "state_file": "policy_preprocessor_step_2_normalizer_processor.safetensors"
63
+ },
64
+ {
65
+ "registry_name": "pi05_prepare_state_tokenizer_processor_step",
66
+ "config": {}
67
+ },
68
+ {
69
+ "registry_name": "tokenizer_processor",
70
+ "config": {
71
+ "max_length": 200,
72
+ "task_key": "task",
73
+ "padding_side": "right",
74
+ "padding": "max_length",
75
+ "truncation": true,
76
+ "tokenizer_name": "google/paligemma-3b-pt-224"
77
+ }
78
+ },
79
+ {
80
+ "registry_name": "device_processor",
81
+ "config": {
82
+ "device": "cuda",
83
+ "float_dtype": null
84
+ }
85
+ }
86
+ ]
87
+ }
checkpoints/003000/pretrained_model/policy_preprocessor_step_2_normalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f99aec40164823a2f22a0ec5032e767eaa10338e84e19744d066a2daef88c7d0
3
+ size 9328
checkpoints/003000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,221 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "Zasha01/cube_transfer_final",
4
+ "root": null,
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ },
61
+ "affine": {
62
+ "weight": 1.0,
63
+ "type": "RandomAffine",
64
+ "kwargs": {
65
+ "degrees": [
66
+ -5.0,
67
+ 5.0
68
+ ],
69
+ "translate": [
70
+ 0.05,
71
+ 0.05
72
+ ]
73
+ }
74
+ }
75
+ }
76
+ },
77
+ "revision": null,
78
+ "use_imagenet_stats": true,
79
+ "video_backend": "torchcodec",
80
+ "streaming": false
81
+ },
82
+ "env": null,
83
+ "policy": {
84
+ "type": "pi05",
85
+ "n_obs_steps": 1,
86
+ "input_features": {
87
+ "observation.state": {
88
+ "type": "STATE",
89
+ "shape": [
90
+ 14
91
+ ]
92
+ },
93
+ "observation.images.context": {
94
+ "type": "VISUAL",
95
+ "shape": [
96
+ 3,
97
+ 360,
98
+ 640
99
+ ]
100
+ },
101
+ "observation.images.right_wrist": {
102
+ "type": "VISUAL",
103
+ "shape": [
104
+ 3,
105
+ 360,
106
+ 640
107
+ ]
108
+ },
109
+ "observation.images.left_wrist": {
110
+ "type": "VISUAL",
111
+ "shape": [
112
+ 3,
113
+ 360,
114
+ 640
115
+ ]
116
+ }
117
+ },
118
+ "output_features": {
119
+ "action": {
120
+ "type": "ACTION",
121
+ "shape": [
122
+ 14
123
+ ]
124
+ }
125
+ },
126
+ "device": "cuda",
127
+ "use_amp": false,
128
+ "push_to_hub": false,
129
+ "repo_id": "local/pi05",
130
+ "private": null,
131
+ "tags": null,
132
+ "license": null,
133
+ "pretrained_path": "lerobot/pi05_base",
134
+ "paligemma_variant": "gemma_2b",
135
+ "action_expert_variant": "gemma_300m",
136
+ "dtype": "bfloat16",
137
+ "chunk_size": 50,
138
+ "n_action_steps": 50,
139
+ "max_state_dim": 32,
140
+ "max_action_dim": 32,
141
+ "num_inference_steps": 10,
142
+ "time_sampling_beta_alpha": 1.5,
143
+ "time_sampling_beta_beta": 1.0,
144
+ "time_sampling_scale": 0.999,
145
+ "time_sampling_offset": 0.001,
146
+ "min_period": 0.004,
147
+ "max_period": 4.0,
148
+ "rtc_config": null,
149
+ "image_resolution": [
150
+ 224,
151
+ 224
152
+ ],
153
+ "empty_cameras": 0,
154
+ "tokenizer_max_length": 200,
155
+ "normalization_mapping": {
156
+ "VISUAL": "IDENTITY",
157
+ "STATE": "QUANTILES",
158
+ "ACTION": "QUANTILES"
159
+ },
160
+ "gradient_checkpointing": true,
161
+ "compile_model": false,
162
+ "compile_mode": "max-autotune",
163
+ "optimizer_lr": 2.5e-05,
164
+ "optimizer_betas": [
165
+ 0.9,
166
+ 0.95
167
+ ],
168
+ "optimizer_eps": 1e-08,
169
+ "optimizer_weight_decay": 0.01,
170
+ "optimizer_grad_clip_norm": 1.0,
171
+ "scheduler_warmup_steps": 1000,
172
+ "scheduler_decay_steps": 30000,
173
+ "scheduler_decay_lr": 2.5e-06
174
+ },
175
+ "output_dir": "/workspace/outputs/pi05",
176
+ "job_name": "pi05_cube_transfer_final",
177
+ "resume": false,
178
+ "seed": 1000,
179
+ "num_workers": 4,
180
+ "batch_size": 16,
181
+ "steps": 3000,
182
+ "eval_freq": 20000,
183
+ "log_freq": 200,
184
+ "save_checkpoint": true,
185
+ "save_freq": 20000,
186
+ "use_policy_training_preset": true,
187
+ "optimizer": {
188
+ "type": "adamw",
189
+ "lr": 2.5e-05,
190
+ "weight_decay": 0.01,
191
+ "grad_clip_norm": 1.0,
192
+ "betas": [
193
+ 0.9,
194
+ 0.95
195
+ ],
196
+ "eps": 1e-08
197
+ },
198
+ "scheduler": {
199
+ "type": "cosine_decay_with_warmup",
200
+ "num_warmup_steps": 1000,
201
+ "num_decay_steps": 30000,
202
+ "peak_lr": 2.5e-05,
203
+ "decay_lr": 2.5e-06
204
+ },
205
+ "eval": {
206
+ "n_episodes": 50,
207
+ "batch_size": 50,
208
+ "use_async_envs": false
209
+ },
210
+ "wandb": {
211
+ "enable": true,
212
+ "disable_artifact": false,
213
+ "project": "lerobot",
214
+ "entity": null,
215
+ "notes": null,
216
+ "run_id": "hxte4s1l",
217
+ "mode": null
218
+ },
219
+ "checkpoint_path": null,
220
+ "rename_map": {}
221
+ }
checkpoints/003000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,833 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.5e-06,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 0.01,
10
+ "amsgrad": false,
11
+ "maximize": false,
12
+ "foreach": null,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "decoupled_weight_decay": true,
17
+ "initial_lr": 2.5e-05,
18
+ "params": [
19
+ 0,
20
+ 1,
21
+ 2,
22
+ 3,
23
+ 4,
24
+ 5,
25
+ 6,
26
+ 7,
27
+ 8,
28
+ 9,
29
+ 10,
30
+ 11,
31
+ 12,
32
+ 13,
33
+ 14,
34
+ 15,
35
+ 16,
36
+ 17,
37
+ 18,
38
+ 19,
39
+ 20,
40
+ 21,
41
+ 22,
42
+ 23,
43
+ 24,
44
+ 25,
45
+ 26,
46
+ 27,
47
+ 28,
48
+ 29,
49
+ 30,
50
+ 31,
51
+ 32,
52
+ 33,
53
+ 34,
54
+ 35,
55
+ 36,
56
+ 37,
57
+ 38,
58
+ 39,
59
+ 40,
60
+ 41,
61
+ 42,
62
+ 43,
63
+ 44,
64
+ 45,
65
+ 46,
66
+ 47,
67
+ 48,
68
+ 49,
69
+ 50,
70
+ 51,
71
+ 52,
72
+ 53,
73
+ 54,
74
+ 55,
75
+ 56,
76
+ 57,
77
+ 58,
78
+ 59,
79
+ 60,
80
+ 61,
81
+ 62,
82
+ 63,
83
+ 64,
84
+ 65,
85
+ 66,
86
+ 67,
87
+ 68,
88
+ 69,
89
+ 70,
90
+ 71,
91
+ 72,
92
+ 73,
93
+ 74,
94
+ 75,
95
+ 76,
96
+ 77,
97
+ 78,
98
+ 79,
99
+ 80,
100
+ 81,
101
+ 82,
102
+ 83,
103
+ 84,
104
+ 85,
105
+ 86,
106
+ 87,
107
+ 88,
108
+ 89,
109
+ 90,
110
+ 91,
111
+ 92,
112
+ 93,
113
+ 94,
114
+ 95,
115
+ 96,
116
+ 97,
117
+ 98,
118
+ 99,
119
+ 100,
120
+ 101,
121
+ 102,
122
+ 103,
123
+ 104,
124
+ 105,
125
+ 106,
126
+ 107,
127
+ 108,
128
+ 109,
129
+ 110,
130
+ 111,
131
+ 112,
132
+ 113,
133
+ 114,
134
+ 115,
135
+ 116,
136
+ 117,
137
+ 118,
138
+ 119,
139
+ 120,
140
+ 121,
141
+ 122,
142
+ 123,
143
+ 124,
144
+ 125,
145
+ 126,
146
+ 127,
147
+ 128,
148
+ 129,
149
+ 130,
150
+ 131,
151
+ 132,
152
+ 133,
153
+ 134,
154
+ 135,
155
+ 136,
156
+ 137,
157
+ 138,
158
+ 139,
159
+ 140,
160
+ 141,
161
+ 142,
162
+ 143,
163
+ 144,
164
+ 145,
165
+ 146,
166
+ 147,
167
+ 148,
168
+ 149,
169
+ 150,
170
+ 151,
171
+ 152,
172
+ 153,
173
+ 154,
174
+ 155,
175
+ 156,
176
+ 157,
177
+ 158,
178
+ 159,
179
+ 160,
180
+ 161,
181
+ 162,
182
+ 163,
183
+ 164,
184
+ 165,
185
+ 166,
186
+ 167,
187
+ 168,
188
+ 169,
189
+ 170,
190
+ 171,
191
+ 172,
192
+ 173,
193
+ 174,
194
+ 175,
195
+ 176,
196
+ 177,
197
+ 178,
198
+ 179,
199
+ 180,
200
+ 181,
201
+ 182,
202
+ 183,
203
+ 184,
204
+ 185,
205
+ 186,
206
+ 187,
207
+ 188,
208
+ 189,
209
+ 190,
210
+ 191,
211
+ 192,
212
+ 193,
213
+ 194,
214
+ 195,
215
+ 196,
216
+ 197,
217
+ 198,
218
+ 199,
219
+ 200,
220
+ 201,
221
+ 202,
222
+ 203,
223
+ 204,
224
+ 205,
225
+ 206,
226
+ 207,
227
+ 208,
228
+ 209,
229
+ 210,
230
+ 211,
231
+ 212,
232
+ 213,
233
+ 214,
234
+ 215,
235
+ 216,
236
+ 217,
237
+ 218,
238
+ 219,
239
+ 220,
240
+ 221,
241
+ 222,
242
+ 223,
243
+ 224,
244
+ 225,
245
+ 226,
246
+ 227,
247
+ 228,
248
+ 229,
249
+ 230,
250
+ 231,
251
+ 232,
252
+ 233,
253
+ 234,
254
+ 235,
255
+ 236,
256
+ 237,
257
+ 238,
258
+ 239,
259
+ 240,
260
+ 241,
261
+ 242,
262
+ 243,
263
+ 244,
264
+ 245,
265
+ 246,
266
+ 247,
267
+ 248,
268
+ 249,
269
+ 250,
270
+ 251,
271
+ 252,
272
+ 253,
273
+ 254,
274
+ 255,
275
+ 256,
276
+ 257,
277
+ 258,
278
+ 259,
279
+ 260,
280
+ 261,
281
+ 262,
282
+ 263,
283
+ 264,
284
+ 265,
285
+ 266,
286
+ 267,
287
+ 268,
288
+ 269,
289
+ 270,
290
+ 271,
291
+ 272,
292
+ 273,
293
+ 274,
294
+ 275,
295
+ 276,
296
+ 277,
297
+ 278,
298
+ 279,
299
+ 280,
300
+ 281,
301
+ 282,
302
+ 283,
303
+ 284,
304
+ 285,
305
+ 286,
306
+ 287,
307
+ 288,
308
+ 289,
309
+ 290,
310
+ 291,
311
+ 292,
312
+ 293,
313
+ 294,
314
+ 295,
315
+ 296,
316
+ 297,
317
+ 298,
318
+ 299,
319
+ 300,
320
+ 301,
321
+ 302,
322
+ 303,
323
+ 304,
324
+ 305,
325
+ 306,
326
+ 307,
327
+ 308,
328
+ 309,
329
+ 310,
330
+ 311,
331
+ 312,
332
+ 313,
333
+ 314,
334
+ 315,
335
+ 316,
336
+ 317,
337
+ 318,
338
+ 319,
339
+ 320,
340
+ 321,
341
+ 322,
342
+ 323,
343
+ 324,
344
+ 325,
345
+ 326,
346
+ 327,
347
+ 328,
348
+ 329,
349
+ 330,
350
+ 331,
351
+ 332,
352
+ 333,
353
+ 334,
354
+ 335,
355
+ 336,
356
+ 337,
357
+ 338,
358
+ 339,
359
+ 340,
360
+ 341,
361
+ 342,
362
+ 343,
363
+ 344,
364
+ 345,
365
+ 346,
366
+ 347,
367
+ 348,
368
+ 349,
369
+ 350,
370
+ 351,
371
+ 352,
372
+ 353,
373
+ 354,
374
+ 355,
375
+ 356,
376
+ 357,
377
+ 358,
378
+ 359,
379
+ 360,
380
+ 361,
381
+ 362,
382
+ 363,
383
+ 364,
384
+ 365,
385
+ 366,
386
+ 367,
387
+ 368,
388
+ 369,
389
+ 370,
390
+ 371,
391
+ 372,
392
+ 373,
393
+ 374,
394
+ 375,
395
+ 376,
396
+ 377,
397
+ 378,
398
+ 379,
399
+ 380,
400
+ 381,
401
+ 382,
402
+ 383,
403
+ 384,
404
+ 385,
405
+ 386,
406
+ 387,
407
+ 388,
408
+ 389,
409
+ 390,
410
+ 391,
411
+ 392,
412
+ 393,
413
+ 394,
414
+ 395,
415
+ 396,
416
+ 397,
417
+ 398,
418
+ 399,
419
+ 400,
420
+ 401,
421
+ 402,
422
+ 403,
423
+ 404,
424
+ 405,
425
+ 406,
426
+ 407,
427
+ 408,
428
+ 409,
429
+ 410,
430
+ 411,
431
+ 412,
432
+ 413,
433
+ 414,
434
+ 415,
435
+ 416,
436
+ 417,
437
+ 418,
438
+ 419,
439
+ 420,
440
+ 421,
441
+ 422,
442
+ 423,
443
+ 424,
444
+ 425,
445
+ 426,
446
+ 427,
447
+ 428,
448
+ 429,
449
+ 430,
450
+ 431,
451
+ 432,
452
+ 433,
453
+ 434,
454
+ 435,
455
+ 436,
456
+ 437,
457
+ 438,
458
+ 439,
459
+ 440,
460
+ 441,
461
+ 442,
462
+ 443,
463
+ 444,
464
+ 445,
465
+ 446,
466
+ 447,
467
+ 448,
468
+ 449,
469
+ 450,
470
+ 451,
471
+ 452,
472
+ 453,
473
+ 454,
474
+ 455,
475
+ 456,
476
+ 457,
477
+ 458,
478
+ 459,
479
+ 460,
480
+ 461,
481
+ 462,
482
+ 463,
483
+ 464,
484
+ 465,
485
+ 466,
486
+ 467,
487
+ 468,
488
+ 469,
489
+ 470,
490
+ 471,
491
+ 472,
492
+ 473,
493
+ 474,
494
+ 475,
495
+ 476,
496
+ 477,
497
+ 478,
498
+ 479,
499
+ 480,
500
+ 481,
501
+ 482,
502
+ 483,
503
+ 484,
504
+ 485,
505
+ 486,
506
+ 487,
507
+ 488,
508
+ 489,
509
+ 490,
510
+ 491,
511
+ 492,
512
+ 493,
513
+ 494,
514
+ 495,
515
+ 496,
516
+ 497,
517
+ 498,
518
+ 499,
519
+ 500,
520
+ 501,
521
+ 502,
522
+ 503,
523
+ 504,
524
+ 505,
525
+ 506,
526
+ 507,
527
+ 508,
528
+ 509,
529
+ 510,
530
+ 511,
531
+ 512,
532
+ 513,
533
+ 514,
534
+ 515,
535
+ 516,
536
+ 517,
537
+ 518,
538
+ 519,
539
+ 520,
540
+ 521,
541
+ 522,
542
+ 523,
543
+ 524,
544
+ 525,
545
+ 526,
546
+ 527,
547
+ 528,
548
+ 529,
549
+ 530,
550
+ 531,
551
+ 532,
552
+ 533,
553
+ 534,
554
+ 535,
555
+ 536,
556
+ 537,
557
+ 538,
558
+ 539,
559
+ 540,
560
+ 541,
561
+ 542,
562
+ 543,
563
+ 544,
564
+ 545,
565
+ 546,
566
+ 547,
567
+ 548,
568
+ 549,
569
+ 550,
570
+ 551,
571
+ 552,
572
+ 553,
573
+ 554,
574
+ 555,
575
+ 556,
576
+ 557,
577
+ 558,
578
+ 559,
579
+ 560,
580
+ 561,
581
+ 562,
582
+ 563,
583
+ 564,
584
+ 565,
585
+ 566,
586
+ 567,
587
+ 568,
588
+ 569,
589
+ 570,
590
+ 571,
591
+ 572,
592
+ 573,
593
+ 574,
594
+ 575,
595
+ 576,
596
+ 577,
597
+ 578,
598
+ 579,
599
+ 580,
600
+ 581,
601
+ 582,
602
+ 583,
603
+ 584,
604
+ 585,
605
+ 586,
606
+ 587,
607
+ 588,
608
+ 589,
609
+ 590,
610
+ 591,
611
+ 592,
612
+ 593,
613
+ 594,
614
+ 595,
615
+ 596,
616
+ 597,
617
+ 598,
618
+ 599,
619
+ 600,
620
+ 601,
621
+ 602,
622
+ 603,
623
+ 604,
624
+ 605,
625
+ 606,
626
+ 607,
627
+ 608,
628
+ 609,
629
+ 610,
630
+ 611,
631
+ 612,
632
+ 613,
633
+ 614,
634
+ 615,
635
+ 616,
636
+ 617,
637
+ 618,
638
+ 619,
639
+ 620,
640
+ 621,
641
+ 622,
642
+ 623,
643
+ 624,
644
+ 625,
645
+ 626,
646
+ 627,
647
+ 628,
648
+ 629,
649
+ 630,
650
+ 631,
651
+ 632,
652
+ 633,
653
+ 634,
654
+ 635,
655
+ 636,
656
+ 637,
657
+ 638,
658
+ 639,
659
+ 640,
660
+ 641,
661
+ 642,
662
+ 643,
663
+ 644,
664
+ 645,
665
+ 646,
666
+ 647,
667
+ 648,
668
+ 649,
669
+ 650,
670
+ 651,
671
+ 652,
672
+ 653,
673
+ 654,
674
+ 655,
675
+ 656,
676
+ 657,
677
+ 658,
678
+ 659,
679
+ 660,
680
+ 661,
681
+ 662,
682
+ 663,
683
+ 664,
684
+ 665,
685
+ 666,
686
+ 667,
687
+ 668,
688
+ 669,
689
+ 670,
690
+ 671,
691
+ 672,
692
+ 673,
693
+ 674,
694
+ 675,
695
+ 676,
696
+ 677,
697
+ 678,
698
+ 679,
699
+ 680,
700
+ 681,
701
+ 682,
702
+ 683,
703
+ 684,
704
+ 685,
705
+ 686,
706
+ 687,
707
+ 688,
708
+ 689,
709
+ 690,
710
+ 691,
711
+ 692,
712
+ 693,
713
+ 694,
714
+ 695,
715
+ 696,
716
+ 697,
717
+ 698,
718
+ 699,
719
+ 700,
720
+ 701,
721
+ 702,
722
+ 703,
723
+ 704,
724
+ 705,
725
+ 706,
726
+ 707,
727
+ 708,
728
+ 709,
729
+ 710,
730
+ 711,
731
+ 712,
732
+ 713,
733
+ 714,
734
+ 715,
735
+ 716,
736
+ 717,
737
+ 718,
738
+ 719,
739
+ 720,
740
+ 721,
741
+ 722,
742
+ 723,
743
+ 724,
744
+ 725,
745
+ 726,
746
+ 727,
747
+ 728,
748
+ 729,
749
+ 730,
750
+ 731,
751
+ 732,
752
+ 733,
753
+ 734,
754
+ 735,
755
+ 736,
756
+ 737,
757
+ 738,
758
+ 739,
759
+ 740,
760
+ 741,
761
+ 742,
762
+ 743,
763
+ 744,
764
+ 745,
765
+ 746,
766
+ 747,
767
+ 748,
768
+ 749,
769
+ 750,
770
+ 751,
771
+ 752,
772
+ 753,
773
+ 754,
774
+ 755,
775
+ 756,
776
+ 757,
777
+ 758,
778
+ 759,
779
+ 760,
780
+ 761,
781
+ 762,
782
+ 763,
783
+ 764,
784
+ 765,
785
+ 766,
786
+ 767,
787
+ 768,
788
+ 769,
789
+ 770,
790
+ 771,
791
+ 772,
792
+ 773,
793
+ 774,
794
+ 775,
795
+ 776,
796
+ 777,
797
+ 778,
798
+ 779,
799
+ 780,
800
+ 781,
801
+ 782,
802
+ 783,
803
+ 784,
804
+ 785,
805
+ 786,
806
+ 787,
807
+ 788,
808
+ 789,
809
+ 790,
810
+ 791,
811
+ 792,
812
+ 793,
813
+ 794,
814
+ 795,
815
+ 796,
816
+ 797,
817
+ 798,
818
+ 799,
819
+ 800,
820
+ 801,
821
+ 802,
822
+ 803,
823
+ 804,
824
+ 805,
825
+ 806,
826
+ 807,
827
+ 808,
828
+ 809,
829
+ 810,
830
+ 811
831
+ ]
832
+ }
833
+ ]
checkpoints/003000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84e56969a0090cae0cf3323e17e839ac6c995aea77da3bb3339b835deab094d6
3
+ size 13473373724
checkpoints/003000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:214b4688be9687b08db35cdc95addd3d85ac9286d6bfdb079fc5e07f171a7709
3
+ size 15708
checkpoints/003000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 2.5e-05
4
+ ],
5
+ "last_epoch": 3000,
6
+ "_step_count": 3001,
7
+ "_get_lr_called_within_step": false,
8
+ "_last_lr": [
9
+ 2.5e-06
10
+ ],
11
+ "lr_lambdas": [
12
+ null
13
+ ]
14
+ }
checkpoints/003000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 3000
3
+ }
wandb/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-12-13T11:20:24.459784868Z","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
2
+ {"time":"2025-12-13T11:20:24.698837605Z","level":"INFO","msg":"stream: created new stream","id":"hxte4s1l"}
3
+ {"time":"2025-12-13T11:20:24.698877505Z","level":"INFO","msg":"stream: started","id":"hxte4s1l"}
4
+ {"time":"2025-12-13T11:20:24.698902182Z","level":"INFO","msg":"writer: started","stream_id":"hxte4s1l"}
5
+ {"time":"2025-12-13T11:20:24.698971295Z","level":"INFO","msg":"sender: started","stream_id":"hxte4s1l"}
6
+ {"time":"2025-12-13T11:20:24.699023033Z","level":"INFO","msg":"handler: started","stream_id":"hxte4s1l"}
7
+ {"time":"2025-12-13T12:50:49.694367446Z","level":"INFO","msg":"stream: closing","id":"hxte4s1l"}
8
+ {"time":"2025-12-13T12:51:00.022925337Z","level":"ERROR","msg":"error adding file to cache","err":"write /root/.cache/wandb/artifacts/tmp/4043578555: no space left on device"}
9
+ {"time":"2025-12-13T12:54:15.718798587Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
10
+ {"time":"2025-12-13T12:54:15.849607597Z","level":"INFO","msg":"handler: closed","stream_id":"hxte4s1l"}
11
+ {"time":"2025-12-13T12:54:15.849760457Z","level":"INFO","msg":"sender: closed","stream_id":"hxte4s1l"}
12
+ {"time":"2025-12-13T12:54:15.849774167Z","level":"INFO","msg":"stream: closed","id":"hxte4s1l"}
wandb/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
2
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Configure stats pid to 139
3
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Loading settings from /root/.config/wandb/settings
4
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Loading settings from /workspace/wandb/settings
5
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /workspace/outputs/pi05/wandb/run-20251213_112024-hxte4s1l/logs/debug.log
7
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /workspace/outputs/pi05/wandb/run-20251213_112024-hxte4s1l/logs/debug-internal.log
8
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:init():813] calling init triggers
9
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'Zasha01/cube_transfer_final', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi05', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': 'local/pi05', 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi05_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'rtc_config': None, 'image_resolution': [224, 224], 'empty_cameras': 0, 'tokenizer_max_length': 200, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.QUANTILES: 'QUANTILES'>, 'ACTION': <NormalizationMode.QUANTILES: 'QUANTILES'>}, 'gradient_checkpointing': True, 'compile_model': False, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06}, 'output_dir': '/workspace/outputs/pi05', 'job_name': 'pi05_cube_transfer_final', 'resume': False, 'seed': 1000, 'num_workers': 4, 'batch_size': 16, 'steps': 3000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {}, '_wandb': {}}
11
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:init():854] starting backend
12
+ 2025-12-13 11:20:24,453 INFO MainThread:139 [wandb_init.py:init():857] sending inform_init request
13
+ 2025-12-13 11:20:24,456 INFO MainThread:139 [wandb_init.py:init():865] backend started and connected
14
+ 2025-12-13 11:20:24,457 INFO MainThread:139 [wandb_init.py:init():936] updated telemetry
15
+ 2025-12-13 11:20:24,461 INFO MainThread:139 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
16
+ 2025-12-13 11:20:24,949 INFO MainThread:139 [wandb_init.py:init():1011] starting run threads in backend
17
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-12-13 11:20:25,018 INFO MainThread:139 [wandb_init.py:init():1049] run started, returning control to user process
22
+ 2025-12-13 12:50:49,695 INFO wandb-AsyncioManager-main:139 [service_client.py:_forward_responses():84] Reached EOF.
23
+ 2025-12-13 12:50:49,695 INFO wandb-AsyncioManager-main:139 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles.
wandb/run-20251213_112024-hxte4s1l/files/config.yaml ADDED
@@ -0,0 +1,245 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _wandb:
2
+ value:
3
+ cli_version: 0.21.4
4
+ e:
5
+ kudd1msmjkuf2t5u1uuhaqdaw70qdzvr:
6
+ args:
7
+ - --dataset.repo_id=Zasha01/cube_transfer_final
8
+ - --policy.type=pi05
9
+ - --output_dir=/workspace/outputs/pi05
10
+ - --job_name=pi05_cube_transfer_final
11
+ - --policy.repo_id=local/pi05
12
+ - --policy.push_to_hub=false
13
+ - --policy.pretrained_path=lerobot/pi05_base
14
+ - --policy.compile_model=false
15
+ - --policy.gradient_checkpointing=true
16
+ - --wandb.enable=true
17
+ - --policy.dtype=bfloat16
18
+ - --steps=3000
19
+ - --policy.device=cuda
20
+ - --batch_size=16
21
+ cpu_count: 28
22
+ cpu_count_logical: 28
23
+ cudaVersion: "12.8"
24
+ disk:
25
+ /:
26
+ total: "103865303040"
27
+ used: "52151582720"
28
+ executable: /usr/bin/python3
29
+ git:
30
+ remote: https://github.com/Manufex/vla-launchables
31
+ gpu: NVIDIA H100 PCIe
32
+ gpu_count: 1
33
+ gpu_nvidia:
34
+ - architecture: Hopper
35
+ cudaCores: 14592
36
+ memoryTotal: "85520809984"
37
+ name: NVIDIA H100 PCIe
38
+ uuid: GPU-610ede60-6f89-0fa7-06b2-87bb3ed58374
39
+ host: brev-e9tf75hch
40
+ memory:
41
+ total: "190128013312"
42
+ os: Linux-6.8.0-64-generic-x86_64-with-glibc2.35
43
+ program: /opt/lerobot/src/lerobot/scripts/lerobot_train.py
44
+ python: CPython 3.10.12
45
+ root: /workspace/outputs/pi05
46
+ startedAt: "2025-12-13T11:20:24.247192Z"
47
+ writerId: kudd1msmjkuf2t5u1uuhaqdaw70qdzvr
48
+ m: []
49
+ python_version: 3.10.12
50
+ t:
51
+ "1":
52
+ - 1
53
+ - 11
54
+ - 41
55
+ - 49
56
+ - 51
57
+ - 71
58
+ - 83
59
+ "2":
60
+ - 1
61
+ - 11
62
+ - 41
63
+ - 49
64
+ - 51
65
+ - 71
66
+ - 83
67
+ "3":
68
+ - 13
69
+ - 15
70
+ - 16
71
+ - 61
72
+ "4": 3.10.12
73
+ "5": 0.21.4
74
+ "6": 4.53.3
75
+ "10":
76
+ - 21
77
+ "12": 0.21.4
78
+ "13": linux-x86_64
79
+ batch_size:
80
+ value: 16
81
+ checkpoint_path:
82
+ value: null
83
+ dataset:
84
+ value:
85
+ episodes: null
86
+ image_transforms:
87
+ enable: false
88
+ max_num_transforms: 3
89
+ random_order: false
90
+ tfs:
91
+ affine:
92
+ kwargs:
93
+ degrees:
94
+ - -5
95
+ - 5
96
+ translate:
97
+ - 0.05
98
+ - 0.05
99
+ type: RandomAffine
100
+ weight: 1
101
+ brightness:
102
+ kwargs:
103
+ brightness:
104
+ - 0.8
105
+ - 1.2
106
+ type: ColorJitter
107
+ weight: 1
108
+ contrast:
109
+ kwargs:
110
+ contrast:
111
+ - 0.8
112
+ - 1.2
113
+ type: ColorJitter
114
+ weight: 1
115
+ hue:
116
+ kwargs:
117
+ hue:
118
+ - -0.05
119
+ - 0.05
120
+ type: ColorJitter
121
+ weight: 1
122
+ saturation:
123
+ kwargs:
124
+ saturation:
125
+ - 0.5
126
+ - 1.5
127
+ type: ColorJitter
128
+ weight: 1
129
+ sharpness:
130
+ kwargs:
131
+ sharpness:
132
+ - 0.5
133
+ - 1.5
134
+ type: SharpnessJitter
135
+ weight: 1
136
+ repo_id: Zasha01/cube_transfer_final
137
+ revision: null
138
+ root: null
139
+ streaming: false
140
+ use_imagenet_stats: true
141
+ video_backend: torchcodec
142
+ env:
143
+ value: null
144
+ eval:
145
+ value:
146
+ batch_size: 50
147
+ n_episodes: 50
148
+ use_async_envs: false
149
+ eval_freq:
150
+ value: 20000
151
+ job_name:
152
+ value: pi05_cube_transfer_final
153
+ log_freq:
154
+ value: 200
155
+ num_workers:
156
+ value: 4
157
+ optimizer:
158
+ value:
159
+ betas:
160
+ - 0.9
161
+ - 0.95
162
+ eps: 1e-08
163
+ grad_clip_norm: 1
164
+ lr: 2.5e-05
165
+ type: adamw
166
+ weight_decay: 0.01
167
+ output_dir:
168
+ value: /workspace/outputs/pi05
169
+ policy:
170
+ value:
171
+ action_expert_variant: gemma_300m
172
+ chunk_size: 50
173
+ compile_mode: max-autotune
174
+ compile_model: false
175
+ device: cuda
176
+ dtype: bfloat16
177
+ empty_cameras: 0
178
+ gradient_checkpointing: true
179
+ image_resolution:
180
+ - 224
181
+ - 224
182
+ license: null
183
+ max_action_dim: 32
184
+ max_period: 4
185
+ max_state_dim: 32
186
+ min_period: 0.004
187
+ n_action_steps: 50
188
+ n_obs_steps: 1
189
+ normalization_mapping:
190
+ ACTION: QUANTILES
191
+ STATE: QUANTILES
192
+ VISUAL: IDENTITY
193
+ num_inference_steps: 10
194
+ optimizer_betas:
195
+ - 0.9
196
+ - 0.95
197
+ optimizer_eps: 1e-08
198
+ optimizer_grad_clip_norm: 1
199
+ optimizer_lr: 2.5e-05
200
+ optimizer_weight_decay: 0.01
201
+ paligemma_variant: gemma_2b
202
+ pretrained_path: lerobot/pi05_base
203
+ private: null
204
+ push_to_hub: false
205
+ repo_id: local/pi05
206
+ rtc_config: null
207
+ scheduler_decay_lr: 2.5e-06
208
+ scheduler_decay_steps: 30000
209
+ scheduler_warmup_steps: 1000
210
+ tags: null
211
+ time_sampling_beta_alpha: 1.5
212
+ time_sampling_beta_beta: 1
213
+ time_sampling_offset: 0.001
214
+ time_sampling_scale: 0.999
215
+ tokenizer_max_length: 200
216
+ type: pi05
217
+ use_amp: false
218
+ resume:
219
+ value: false
220
+ save_checkpoint:
221
+ value: true
222
+ save_freq:
223
+ value: 20000
224
+ scheduler:
225
+ value:
226
+ decay_lr: 2.5e-06
227
+ num_decay_steps: 30000
228
+ num_warmup_steps: 1000
229
+ peak_lr: 2.5e-05
230
+ type: cosine_decay_with_warmup
231
+ seed:
232
+ value: 1000
233
+ steps:
234
+ value: 3000
235
+ use_policy_training_preset:
236
+ value: true
237
+ wandb:
238
+ value:
239
+ disable_artifact: false
240
+ enable: true
241
+ entity: null
242
+ mode: null
243
+ notes: null
244
+ project: lerobot
245
+ run_id: null
wandb/run-20251213_112024-hxte4s1l/files/output.log ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ INFO 2025-12-13 11:20:25 db_utils.py:102 Logs will be synced with wandb.
2
+ INFO 2025-12-13 11:20:25 db_utils.py:103 Track this run --> https://wandb.ai/z-sharfeddine-technical-university-of-munich/lerobot/runs/hxte4s1l
3
+ INFO 2025-12-13 11:20:25 ot_train.py:184 Creating dataset
4
+ meta/episodes/chunk-000/file-000.parquet: 100%|█| 307k/307k [00:00<00:00, 1.07MB
5
+ info.json: 4.49kB [00:00, 7.97MB/s]
6
+ stats.json: 20.9kB [00:00, 37.7MB/s]
7
+ meta/tasks.parquet: 100%|██████████████████| 2.29k/2.29k [00:00<00:00, 6.37kB/s]
8
+ .gitattributes: 2.46kB [00:00, 6.96MB/s]
9
+ README.md: 4.98kB [00:00, 9.28MB/s]
10
+ data/chunk-000/file-000.parquet: 100%|█████| 2.66M/2.66M [00:00<00:00, 16.7MB/s]
11
+ videos/observation.images.context/chunk-(…): 100%|█| 511M/511M [00:01<00:00, 274
12
+ videos/observation.images.context/chunk-(…): 100%|█| 202M/202M [00:01<00:00, 107
13
+ videos/observation.images.left_wrist/chu(…): 100%|█| 514M/514M [00:01<00:00, 307
14
+ videos/observation.images.left_wrist/chu(…): 100%|█| 115M/115M [00:01<00:00, 81.
15
+ videos/observation.images.right_wrist/ch(…): 100%|█| 512M/512M [00:01<00:00, 288
16
+ videos/observation.images.right_wrist/ch(…): 100%|█| 27.0M/27.0M [00:00<00:00, 9
17
+ INFO 2025-12-13 11:20:37 ot_train.py:203 Creating policy
18
+ The PI05 model is a direct port of the OpenPI implementation.
19
+ This implementation follows the original OpenPI structure for compatibility.
20
+ Original implementation: https://github.com/Physical-Intelligence/openpi
21
+ INFO 2025-12-13 11:21:22 ing_pi05.py:568 Enabled gradient checkpointing for PI05Pytorch model
22
+ Loading model from: lerobot/pi05_base
23
+ model.safetensors: 100%|███████████████████| 14.5G/14.5G [00:12<00:00, 1.17GB/s]
24
+ ✓ Loaded state dict from model.safetensors
25
+ WARNING 2025-12-13 11:21:38 ng_pi05.py:1074 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.bias
26
+ WARNING 2025-12-13 11:21:38 ng_pi05.py:1074 Vision embedding key might need handling: paligemma_with_expert.paligemma.model.vision_tower.vision_model.embeddings.patch_embedding.weight
27
+ Remapped: action_in_proj.bias -> model.action_in_proj.bias
28
+ Remapped: action_in_proj.weight -> model.action_in_proj.weight
29
+ Remapped: action_out_proj.bias -> model.action_out_proj.bias
30
+ Remapped: action_out_proj.weight -> model.action_out_proj.weight
31
+ Remapped: paligemma_with_expert.gemma_expert.lm_head.weight -> model.paligemma_with_expert.gemma_expert.lm_head.weight
32
+ Remapped: paligemma_with_expert.gemma_expert.model.layers.0.input_layernorm.dense.bias -> model.paligemma_with_expert.gemma_expert.model.layers.0.input_layernorm.dense.bias
33
+ Remapped: paligemma_with_expert.gemma_expert.model.layers.0.input_layernorm.dense.weight -> model.paligemma_with_expert.gemma_expert.model.layers.0.input_layernorm.dense.weight
34
+ Remapped: paligemma_with_expert.gemma_expert.model.layers.0.mlp.down_proj.weight -> model.paligemma_with_expert.gemma_expert.model.layers.0.mlp.down_proj.weight
35
+ Remapped: paligemma_with_expert.gemma_expert.model.layers.0.mlp.gate_proj.weight -> model.paligemma_with_expert.gemma_expert.model.layers.0.mlp.gate_proj.weight
36
+ Remapped: paligemma_with_expert.gemma_expert.model.layers.0.mlp.up_proj.weight -> model.paligemma_with_expert.gemma_expert.model.layers.0.mlp.up_proj.weight
37
+ Remapped 812 state dict keys
38
+ Warning: Could not remap state dict keys: Error(s) in loading state_dict for PI05Policy:
39
+ Missing key(s) in state_dict: "model.paligemma_with_expert.paligemma.model.language_model.embed_tokens.weight".
40
+ policy_preprocessor.json: 1.03kB [00:00, 2.39MB/s]
41
+ tokenizer_config.json: 100%|███████████████| 40.0k/40.0k [00:00<00:00, 67.6MB/s]
42
+ tokenizer.model: 100%|█████████████████████| 4.26M/4.26M [00:00<00:00, 10.3MB/s]
43
+ tokenizer.json: 100%|██████████████████████| 17.5M/17.5M [00:00<00:00, 44.9MB/s]
44
+ added_tokens.json: 100%|██████████████████████| 24.0/24.0 [00:00<00:00, 187kB/s]
45
+ special_tokens_map.json: 100%|█████████████████| 607/607 [00:00<00:00, 4.37MB/s]
46
+ policy_postprocessor.json: 100%|███████████████| 450/450 [00:00<00:00, 3.50MB/s]
47
+ INFO 2025-12-13 11:21:46 ot_train.py:248 Creating optimizer and scheduler
48
+ INFO 2025-12-13 11:21:46 hedulers.py:105 Auto-scaling LR scheduler: num_training_steps (3000) < num_decay_steps (30000). Scaling warmup: 1000 → 100, decay: 30000 → 3000 (scale factor: 0.100)
49
+ INFO 2025-12-13 11:21:46 ot_train.py:260 Output dir: /workspace/outputs/pi05
50
+ INFO 2025-12-13 11:21:46 ot_train.py:267 cfg.steps=3000 (3K)
51
+ INFO 2025-12-13 11:21:46 ot_train.py:268 dataset.num_frames=27217 (27K)
52
+ INFO 2025-12-13 11:21:46 ot_train.py:269 dataset.num_episodes=47
53
+ INFO 2025-12-13 11:21:46 ot_train.py:272 Effective batch size: 16 x 1 = 16
54
+ INFO 2025-12-13 11:21:46 ot_train.py:273 num_learnable_params=3616757520 (4B)
55
+ INFO 2025-12-13 11:21:46 ot_train.py:274 num_total_params=3616757520 (4B)
56
+ INFO 2025-12-13 11:21:46 ot_train.py:330 Start offline training on a fixed dataset
57
+ INFO 2025-12-13 11:27:43 ot_train.py:357 step:200 smpl:3K ep:6 epch:0.12 loss:0.127 grdn:1.447 lr:1.9e-05 updt_s:1.732 data_s:0.050
58
+ WARNING 2025-12-13 11:27:43 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
59
+ INFO 2025-12-13 11:33:39 ot_train.py:357 step:400 smpl:6K ep:11 epch:0.24 loss:0.079 grdn:0.940 lr:2.4e-05 updt_s:1.733 data_s:0.041
60
+ WARNING 2025-12-13 11:33:39 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
61
+ INFO 2025-12-13 11:39:34 ot_train.py:357 step:600 smpl:10K ep:17 epch:0.35 loss:0.067 grdn:0.855 lr:2.3e-05 updt_s:1.732 data_s:0.041
62
+ WARNING 2025-12-13 11:39:34 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
63
+ INFO 2025-12-13 11:45:28 ot_train.py:357 step:800 smpl:13K ep:22 epch:0.47 loss:0.061 grdn:0.829 lr:2.2e-05 updt_s:1.725 data_s:0.041
64
+ WARNING 2025-12-13 11:45:28 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
65
+ INFO 2025-12-13 11:51:20 ot_train.py:357 step:1K smpl:16K ep:28 epch:0.59 loss:0.057 grdn:0.805 lr:2.0e-05 updt_s:1.718 data_s:0.041
66
+ WARNING 2025-12-13 11:51:20 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
67
+ INFO 2025-12-13 11:57:11 ot_train.py:357 step:1K smpl:19K ep:33 epch:0.71 loss:0.051 grdn:0.813 lr:1.8e-05 updt_s:1.710 data_s:0.041
68
+ WARNING 2025-12-13 11:57:11 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
69
+ INFO 2025-12-13 12:03:02 ot_train.py:357 step:1K smpl:22K ep:39 epch:0.82 loss:0.046 grdn:0.758 lr:1.6e-05 updt_s:1.710 data_s:0.041
70
+ WARNING 2025-12-13 12:03:02 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
71
+ INFO 2025-12-13 12:08:54 ot_train.py:357 step:2K smpl:26K ep:44 epch:0.94 loss:0.044 grdn:0.787 lr:1.4e-05 updt_s:1.710 data_s:0.046
72
+ WARNING 2025-12-13 12:08:54 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
73
+ INFO 2025-12-13 12:14:45 ot_train.py:357 step:2K smpl:29K ep:50 epch:1.06 loss:0.041 grdn:0.771 lr:1.1e-05 updt_s:1.707 data_s:0.048
74
+ WARNING 2025-12-13 12:14:45 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
75
+ INFO 2025-12-13 12:20:37 ot_train.py:357 step:2K smpl:32K ep:55 epch:1.18 loss:0.036 grdn:0.693 lr:9.2e-06 updt_s:1.713 data_s:0.041
76
+ WARNING 2025-12-13 12:20:37 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
77
+ INFO 2025-12-13 12:26:29 ot_train.py:357 step:2K smpl:35K ep:61 epch:1.29 loss:0.035 grdn:0.691 lr:7.1e-06 updt_s:1.715 data_s:0.041
78
+ WARNING 2025-12-13 12:26:29 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
79
+ INFO 2025-12-13 12:32:20 ot_train.py:357 step:2K smpl:38K ep:66 epch:1.41 loss:0.036 grdn:0.748 lr:5.4e-06 updt_s:1.713 data_s:0.041
80
+ WARNING 2025-12-13 12:32:20 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
81
+ INFO 2025-12-13 12:38:11 ot_train.py:357 step:3K smpl:42K ep:72 epch:1.53 loss:0.035 grdn:0.767 lr:4.0e-06 updt_s:1.708 data_s:0.041
82
+ WARNING 2025-12-13 12:38:11 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
83
+ INFO 2025-12-13 12:44:02 ot_train.py:357 step:3K smpl:45K ep:77 epch:1.65 loss:0.033 grdn:0.726 lr:3.1e-06 updt_s:1.712 data_s:0.041
84
+ WARNING 2025-12-13 12:44:02 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
85
+ INFO 2025-12-13 12:49:53 ot_train.py:357 step:3K smpl:48K ep:83 epch:1.76 loss:0.034 grdn:0.740 lr:2.6e-06 updt_s:1.710 data_s:0.041
86
+ WARNING 2025-12-13 12:49:53 db_utils.py:141 WandB logging of key "loss_per_dim" was ignored as its type "<class 'list'>" is not handled by this wrapper.
87
+ INFO 2025-12-13 12:49:53 ot_train.py:367 Checkpoint policy after step 3000
88
+ INFO 2025-12-13 12:50:49 ot_train.py:438 End of training
wandb/run-20251213_112024-hxte4s1l/files/requirements.txt ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ inquirerpy==0.3.4
2
+ deepdiff==8.6.1
3
+ xxhash==3.6.0
4
+ dill==0.4.0
5
+ lerobot==0.4.3
6
+ attrs==25.4.0
7
+ tokenizers==0.21.4
8
+ accelerate==1.12.0
9
+ ImageIO==2.37.2
10
+ nvidia-cusparselt-cu12==0.6.3
11
+ safetensors==0.7.0
12
+ imageio-ffmpeg==0.6.0
13
+ Farama-Notifications==0.0.4
14
+ wcwidth==0.2.14
15
+ pyyaml-include==1.4.1
16
+ python-xlib==0.33
17
+ yarl==1.22.0
18
+ nvidia-cusolver-cu12==11.7.1.2
19
+ aiosignal==1.4.0
20
+ nvidia-cusparse-cu12==12.5.4.2
21
+ wandb==0.21.4
22
+ nvidia-cufile-cu12==1.11.1.6
23
+ nvidia-cuda-runtime-cu12==12.6.77
24
+ nvidia-nvjitlink-cu12==12.6.85
25
+ six==1.17.0
26
+ triton==3.3.1
27
+ nvidia-nvtx-cu12==12.6.77
28
+ nvidia-curand-cu12==10.3.7.77
29
+ nvidia-cuda-nvrtc-cu12==12.6.77
30
+ nvidia-cufft-cu12==11.3.0.4
31
+ rerun-sdk==0.26.2
32
+ evdev==1.9.2
33
+ torch==2.7.1
34
+ opencv-python-headless==4.12.0.88
35
+ huggingface-hub==0.35.3
36
+ importlib_metadata==8.7.0
37
+ psutil==7.1.3
38
+ av==15.1.0
39
+ pfzy==0.3.4
40
+ cmake==4.1.3
41
+ tzdata==2025.2
42
+ typing-inspect==0.9.0
43
+ datasets==4.1.1
44
+ pandas==2.3.3
45
+ torchcodec==0.5
46
+ sympy==1.14.0
47
+ nvidia-cudnn-cu12==9.5.1.17
48
+ draccus==0.10.0
49
+ hf_transfer==0.1.9
50
+ pyarrow==22.0.0
51
+ pytz==2025.2
52
+ multiprocess==0.70.16
53
+ python-dateutil==2.9.0.post0
54
+ nvidia-cuda-cupti-cu12==12.6.80
55
+ aiohappyeyeballs==2.6.1
56
+ pyserial==3.5
57
+ cloudpickle==3.1.2
58
+ mypy_extensions==1.1.0
59
+ einops==0.8.1
60
+ torchvision==0.22.1
61
+ aiohttp==3.13.2
62
+ frozenlist==1.8.0
63
+ mergedeep==1.3.4
64
+ regex==2025.11.3
65
+ jsonlines==4.0.0
66
+ prompt_toolkit==3.0.52
67
+ transformers==4.53.3
68
+ toml==0.10.2
69
+ zipp==3.23.0
70
+ nvidia-cublas-cu12==12.6.4.1
71
+ propcache==0.4.1
72
+ diffusers==0.35.2
73
+ nvidia-nccl-cu12==2.26.2
74
+ termcolor==3.2.0
75
+ async-timeout==5.0.1
76
+ gymnasium==1.2.2
77
+ pynput==1.8.1
78
+ multidict==6.7.0
79
+ orderly-set==5.5.0
80
+ GitPython==3.1.45
81
+ pydantic_core==2.41.5
82
+ shellingham==1.5.4
83
+ requests==2.32.5
84
+ httpcore==1.0.9
85
+ gitdb==4.0.12
86
+ typer-slim==0.20.0
87
+ smmap==5.0.2
88
+ anyio==4.12.0
89
+ packaging==25.0
90
+ tqdm==4.67.1
91
+ PyYAML==6.0.3
92
+ pydantic==2.12.5
93
+ h11==0.16.0
94
+ certifi==2025.11.12
95
+ urllib3==2.6.2
96
+ click==8.3.1
97
+ idna==3.11
98
+ typing-inspection==0.4.2
99
+ httpx==0.28.1
100
+ protobuf==6.33.2
101
+ annotated-types==0.7.0
102
+ charset-normalizer==3.4.4
103
+ sentry-sdk==2.47.0
104
+ platformdirs==4.5.1
105
+ exceptiongroup==1.3.1
106
+ hf-xet==1.2.0
107
+ networkx==3.3
108
+ mpmath==1.3.0
109
+ numpy==2.1.2
110
+ pip==25.3
111
+ Jinja2==3.1.6
112
+ pillow==11.3.0
113
+ MarkupSafe==2.1.5
114
+ setuptools==80.9.0
115
+ filelock==3.19.1
116
+ torchaudio==2.5.1+cu121
117
+ fsspec==2025.9.0
118
+ typing_extensions==4.15.0
119
+ wheel==0.45.1
120
+ lerobot==0.4.3
121
+ setuptools==59.6.0
122
+ pip==22.0.2
123
+ wheel==0.37.1
wandb/run-20251213_112024-hxte4s1l/files/wandb-metadata.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "os": "Linux-6.8.0-64-generic-x86_64-with-glibc2.35",
3
+ "python": "CPython 3.10.12",
4
+ "startedAt": "2025-12-13T11:20:24.247192Z",
5
+ "args": [
6
+ "--dataset.repo_id=Zasha01/cube_transfer_final",
7
+ "--policy.type=pi05",
8
+ "--output_dir=/workspace/outputs/pi05",
9
+ "--job_name=pi05_cube_transfer_final",
10
+ "--policy.repo_id=local/pi05",
11
+ "--policy.push_to_hub=false",
12
+ "--policy.pretrained_path=lerobot/pi05_base",
13
+ "--policy.compile_model=false",
14
+ "--policy.gradient_checkpointing=true",
15
+ "--wandb.enable=true",
16
+ "--policy.dtype=bfloat16",
17
+ "--steps=3000",
18
+ "--policy.device=cuda",
19
+ "--batch_size=16"
20
+ ],
21
+ "program": "/opt/lerobot/src/lerobot/scripts/lerobot_train.py",
22
+ "git": {
23
+ "remote": "https://github.com/Manufex/vla-launchables"
24
+ },
25
+ "root": "/workspace/outputs/pi05",
26
+ "host": "brev-e9tf75hch",
27
+ "executable": "/usr/bin/python3",
28
+ "cpu_count": 28,
29
+ "cpu_count_logical": 28,
30
+ "gpu": "NVIDIA H100 PCIe",
31
+ "gpu_count": 1,
32
+ "disk": {
33
+ "/": {
34
+ "total": "103865303040",
35
+ "used": "52151582720"
36
+ }
37
+ },
38
+ "memory": {
39
+ "total": "190128013312"
40
+ },
41
+ "gpu_nvidia": [
42
+ {
43
+ "name": "NVIDIA H100 PCIe",
44
+ "memoryTotal": "85520809984",
45
+ "cudaCores": 14592,
46
+ "architecture": "Hopper",
47
+ "uuid": "GPU-610ede60-6f89-0fa7-06b2-87bb3ed58374"
48
+ }
49
+ ],
50
+ "cudaVersion": "12.8",
51
+ "writerId": "kudd1msmjkuf2t5u1uuhaqdaw70qdzvr"
52
+ }
wandb/run-20251213_112024-hxte4s1l/files/wandb-summary.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"_timestamp":1.7656301934634764e+09,"train/update_s":1.709934723770034,"train/dataloading_s":0.041204123439974864,"train/steps":3000,"_wandb":{"runtime":5424},"train/loss":0.02786409668624401,"train/samples":48000,"train/epochs":1.7636036300841387,"train/episodes":82.88937061395451,"_runtime":5424.744288297,"train/grad_norm":0.7404379284381867,"train/lr":2.5814529262737084e-06,"_step":3000}
wandb/run-20251213_112024-hxte4s1l/logs/debug-internal.log ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"time":"2025-12-13T11:20:24.459784868Z","level":"INFO","msg":"stream: starting","core version":"0.21.4"}
2
+ {"time":"2025-12-13T11:20:24.698837605Z","level":"INFO","msg":"stream: created new stream","id":"hxte4s1l"}
3
+ {"time":"2025-12-13T11:20:24.698877505Z","level":"INFO","msg":"stream: started","id":"hxte4s1l"}
4
+ {"time":"2025-12-13T11:20:24.698902182Z","level":"INFO","msg":"writer: started","stream_id":"hxte4s1l"}
5
+ {"time":"2025-12-13T11:20:24.698971295Z","level":"INFO","msg":"sender: started","stream_id":"hxte4s1l"}
6
+ {"time":"2025-12-13T11:20:24.699023033Z","level":"INFO","msg":"handler: started","stream_id":"hxte4s1l"}
7
+ {"time":"2025-12-13T12:50:49.694367446Z","level":"INFO","msg":"stream: closing","id":"hxte4s1l"}
8
+ {"time":"2025-12-13T12:51:00.022925337Z","level":"ERROR","msg":"error adding file to cache","err":"write /root/.cache/wandb/artifacts/tmp/4043578555: no space left on device"}
9
+ {"time":"2025-12-13T12:54:15.718798587Z","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
10
+ {"time":"2025-12-13T12:54:15.849607597Z","level":"INFO","msg":"handler: closed","stream_id":"hxte4s1l"}
11
+ {"time":"2025-12-13T12:54:15.849760457Z","level":"INFO","msg":"sender: closed","stream_id":"hxte4s1l"}
12
+ {"time":"2025-12-13T12:54:15.849774167Z","level":"INFO","msg":"stream: closed","id":"hxte4s1l"}
wandb/run-20251213_112024-hxte4s1l/logs/debug.log ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4
2
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Configure stats pid to 139
3
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Loading settings from /root/.config/wandb/settings
4
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Loading settings from /workspace/wandb/settings
5
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_setup.py:_flush():81] Loading settings from environment variables
6
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:setup_run_log_directory():686] Logging user logs to /workspace/outputs/pi05/wandb/run-20251213_112024-hxte4s1l/logs/debug.log
7
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to /workspace/outputs/pi05/wandb/run-20251213_112024-hxte4s1l/logs/debug-internal.log
8
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:init():813] calling init triggers
9
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:init():818] wandb.init called with sweep_config: {}
10
+ config: {'dataset': {'repo_id': 'Zasha01/cube_transfer_final', 'root': None, 'episodes': None, 'image_transforms': {'enable': False, 'max_num_transforms': 3, 'random_order': False, 'tfs': {'brightness': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'brightness': [0.8, 1.2]}}, 'contrast': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'contrast': [0.8, 1.2]}}, 'saturation': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'saturation': [0.5, 1.5]}}, 'hue': {'weight': 1.0, 'type': 'ColorJitter', 'kwargs': {'hue': [-0.05, 0.05]}}, 'sharpness': {'weight': 1.0, 'type': 'SharpnessJitter', 'kwargs': {'sharpness': [0.5, 1.5]}}, 'affine': {'weight': 1.0, 'type': 'RandomAffine', 'kwargs': {'degrees': [-5.0, 5.0], 'translate': [0.05, 0.05]}}}}, 'revision': None, 'use_imagenet_stats': True, 'video_backend': 'torchcodec', 'streaming': False}, 'env': None, 'policy': {'type': 'pi05', 'n_obs_steps': 1, 'input_features': {}, 'output_features': {}, 'device': 'cuda', 'use_amp': False, 'push_to_hub': False, 'repo_id': 'local/pi05', 'private': None, 'tags': None, 'license': None, 'pretrained_path': 'lerobot/pi05_base', 'paligemma_variant': 'gemma_2b', 'action_expert_variant': 'gemma_300m', 'dtype': 'bfloat16', 'chunk_size': 50, 'n_action_steps': 50, 'max_state_dim': 32, 'max_action_dim': 32, 'num_inference_steps': 10, 'time_sampling_beta_alpha': 1.5, 'time_sampling_beta_beta': 1.0, 'time_sampling_scale': 0.999, 'time_sampling_offset': 0.001, 'min_period': 0.004, 'max_period': 4.0, 'rtc_config': None, 'image_resolution': [224, 224], 'empty_cameras': 0, 'tokenizer_max_length': 200, 'normalization_mapping': {'VISUAL': <NormalizationMode.IDENTITY: 'IDENTITY'>, 'STATE': <NormalizationMode.QUANTILES: 'QUANTILES'>, 'ACTION': <NormalizationMode.QUANTILES: 'QUANTILES'>}, 'gradient_checkpointing': True, 'compile_model': False, 'compile_mode': 'max-autotune', 'optimizer_lr': 2.5e-05, 'optimizer_betas': [0.9, 0.95], 'optimizer_eps': 1e-08, 'optimizer_weight_decay': 0.01, 'optimizer_grad_clip_norm': 1.0, 'scheduler_warmup_steps': 1000, 'scheduler_decay_steps': 30000, 'scheduler_decay_lr': 2.5e-06}, 'output_dir': '/workspace/outputs/pi05', 'job_name': 'pi05_cube_transfer_final', 'resume': False, 'seed': 1000, 'num_workers': 4, 'batch_size': 16, 'steps': 3000, 'eval_freq': 20000, 'log_freq': 200, 'save_checkpoint': True, 'save_freq': 20000, 'use_policy_training_preset': True, 'optimizer': {'type': 'adamw', 'lr': 2.5e-05, 'weight_decay': 0.01, 'grad_clip_norm': 1.0, 'betas': [0.9, 0.95], 'eps': 1e-08}, 'scheduler': {'type': 'cosine_decay_with_warmup', 'num_warmup_steps': 1000, 'num_decay_steps': 30000, 'peak_lr': 2.5e-05, 'decay_lr': 2.5e-06}, 'eval': {'n_episodes': 50, 'batch_size': 50, 'use_async_envs': False}, 'wandb': {'enable': True, 'disable_artifact': False, 'project': 'lerobot', 'entity': None, 'notes': None, 'run_id': None, 'mode': None}, 'checkpoint_path': None, 'rename_map': {}, '_wandb': {}}
11
+ 2025-12-13 11:20:24,248 INFO MainThread:139 [wandb_init.py:init():854] starting backend
12
+ 2025-12-13 11:20:24,453 INFO MainThread:139 [wandb_init.py:init():857] sending inform_init request
13
+ 2025-12-13 11:20:24,456 INFO MainThread:139 [wandb_init.py:init():865] backend started and connected
14
+ 2025-12-13 11:20:24,457 INFO MainThread:139 [wandb_init.py:init():936] updated telemetry
15
+ 2025-12-13 11:20:24,461 INFO MainThread:139 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout
16
+ 2025-12-13 11:20:24,949 INFO MainThread:139 [wandb_init.py:init():1011] starting run threads in backend
17
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_console_start():2506] atexit reg
18
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_redirect():2354] redirect: wrap_raw
19
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_redirect():2423] Wrapping output streams.
20
+ 2025-12-13 11:20:25,016 INFO MainThread:139 [wandb_run.py:_redirect():2446] Redirects installed.
21
+ 2025-12-13 11:20:25,018 INFO MainThread:139 [wandb_init.py:init():1049] run started, returning control to user process
22
+ 2025-12-13 12:50:49,695 INFO wandb-AsyncioManager-main:139 [service_client.py:_forward_responses():84] Reached EOF.
23
+ 2025-12-13 12:50:49,695 INFO wandb-AsyncioManager-main:139 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles.
wandb/run-20251213_112024-hxte4s1l/run-hxte4s1l.wandb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dac77113b716bd466d8565f9dff2953d2a3d15295ec9e005ee61e70fecf357e
3
+ size 367178