5usu2806 commited on
Commit
25d5f23
·
1 Parent(s): b896a19

Add training checkpoints (020000–last)

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. checkpoints/020000/pretrained_model/config.json +81 -0
  2. checkpoints/020000/pretrained_model/model.safetensors +3 -0
  3. checkpoints/020000/pretrained_model/policy_postprocessor.json +32 -0
  4. checkpoints/020000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors +3 -0
  5. checkpoints/020000/pretrained_model/policy_preprocessor.json +71 -0
  6. checkpoints/020000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors +3 -0
  7. checkpoints/020000/pretrained_model/train_config.json +210 -0
  8. checkpoints/020000/training_state/optimizer_param_groups.json +521 -0
  9. checkpoints/020000/training_state/optimizer_state.safetensors +3 -0
  10. checkpoints/020000/training_state/rng_state.safetensors +3 -0
  11. checkpoints/020000/training_state/scheduler_state.json +14 -0
  12. checkpoints/020000/training_state/training_step.json +3 -0
  13. checkpoints/040000/pretrained_model/config.json +81 -0
  14. checkpoints/040000/pretrained_model/model.safetensors +3 -0
  15. checkpoints/040000/pretrained_model/policy_postprocessor.json +32 -0
  16. checkpoints/040000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors +3 -0
  17. checkpoints/040000/pretrained_model/policy_preprocessor.json +71 -0
  18. checkpoints/040000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors +3 -0
  19. checkpoints/040000/pretrained_model/train_config.json +210 -0
  20. checkpoints/040000/training_state/optimizer_param_groups.json +521 -0
  21. checkpoints/040000/training_state/optimizer_state.safetensors +3 -0
  22. checkpoints/040000/training_state/rng_state.safetensors +3 -0
  23. checkpoints/040000/training_state/scheduler_state.json +14 -0
  24. checkpoints/040000/training_state/training_step.json +3 -0
  25. checkpoints/060000/pretrained_model/config.json +81 -0
  26. checkpoints/060000/pretrained_model/model.safetensors +3 -0
  27. checkpoints/060000/pretrained_model/policy_postprocessor.json +32 -0
  28. checkpoints/060000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors +3 -0
  29. checkpoints/060000/pretrained_model/policy_preprocessor.json +71 -0
  30. checkpoints/060000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors +3 -0
  31. checkpoints/060000/pretrained_model/train_config.json +210 -0
  32. checkpoints/060000/training_state/optimizer_param_groups.json +521 -0
  33. checkpoints/060000/training_state/optimizer_state.safetensors +3 -0
  34. checkpoints/060000/training_state/rng_state.safetensors +3 -0
  35. checkpoints/060000/training_state/scheduler_state.json +14 -0
  36. checkpoints/060000/training_state/training_step.json +3 -0
  37. checkpoints/080000/pretrained_model/config.json +81 -0
  38. checkpoints/080000/pretrained_model/model.safetensors +3 -0
  39. checkpoints/080000/pretrained_model/policy_postprocessor.json +32 -0
  40. checkpoints/080000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors +3 -0
  41. checkpoints/080000/pretrained_model/policy_preprocessor.json +71 -0
  42. checkpoints/080000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors +3 -0
  43. checkpoints/080000/pretrained_model/train_config.json +210 -0
  44. checkpoints/080000/training_state/optimizer_param_groups.json +521 -0
  45. checkpoints/080000/training_state/optimizer_state.safetensors +3 -0
  46. checkpoints/080000/training_state/rng_state.safetensors +3 -0
  47. checkpoints/080000/training_state/scheduler_state.json +14 -0
  48. checkpoints/080000/training_state/training_step.json +3 -0
  49. checkpoints/100000/pretrained_model/config.json +81 -0
  50. checkpoints/100000/pretrained_model/model.safetensors +3 -0
checkpoints/020000/pretrained_model/config.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 6
9
+ ]
10
+ },
11
+ "observation.images.front": {
12
+ "type": "VISUAL",
13
+ "shape": [
14
+ 3,
15
+ 480,
16
+ 640
17
+ ]
18
+ }
19
+ },
20
+ "output_features": {
21
+ "action": {
22
+ "type": "ACTION",
23
+ "shape": [
24
+ 6
25
+ ]
26
+ }
27
+ },
28
+ "device": "cuda",
29
+ "use_amp": false,
30
+ "push_to_hub": true,
31
+ "repo_id": "Dimios45/smolvla1",
32
+ "private": null,
33
+ "tags": null,
34
+ "license": null,
35
+ "pretrained_path": null,
36
+ "chunk_size": 50,
37
+ "n_action_steps": 50,
38
+ "normalization_mapping": {
39
+ "VISUAL": "IDENTITY",
40
+ "STATE": "MEAN_STD",
41
+ "ACTION": "MEAN_STD"
42
+ },
43
+ "max_state_dim": 32,
44
+ "max_action_dim": 32,
45
+ "resize_imgs_with_padding": [
46
+ 512,
47
+ 512
48
+ ],
49
+ "empty_cameras": 0,
50
+ "adapt_to_pi_aloha": false,
51
+ "use_delta_joint_actions_aloha": false,
52
+ "tokenizer_max_length": 48,
53
+ "num_steps": 10,
54
+ "use_cache": true,
55
+ "freeze_vision_encoder": true,
56
+ "train_expert_only": true,
57
+ "train_state_proj": true,
58
+ "optimizer_lr": 0.0001,
59
+ "optimizer_betas": [
60
+ 0.9,
61
+ 0.95
62
+ ],
63
+ "optimizer_eps": 1e-08,
64
+ "optimizer_weight_decay": 1e-10,
65
+ "optimizer_grad_clip_norm": 10,
66
+ "scheduler_warmup_steps": 1000,
67
+ "scheduler_decay_steps": 30000,
68
+ "scheduler_decay_lr": 2.5e-06,
69
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
70
+ "load_vlm_weights": false,
71
+ "add_image_special_tokens": false,
72
+ "attention_mode": "cross_attn",
73
+ "prefix_length": -1,
74
+ "pad_language_to": "longest",
75
+ "num_expert_layers": -1,
76
+ "num_vlm_layers": 16,
77
+ "self_attn_every_n_layers": 2,
78
+ "expert_width_multiplier": 0.75,
79
+ "min_period": 0.004,
80
+ "max_period": 4.0
81
+ }
checkpoints/020000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee92ea7c584703562569df5e01300d8e518067f639a040ac1b837292a3684cad
3
+ size 1197789224
checkpoints/020000/pretrained_model/policy_postprocessor.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_postprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "unnormalizer_processor",
6
+ "config": {
7
+ "eps": 1e-08,
8
+ "features": {
9
+ "action": {
10
+ "type": "ACTION",
11
+ "shape": [
12
+ 6
13
+ ]
14
+ }
15
+ },
16
+ "norm_map": {
17
+ "VISUAL": "IDENTITY",
18
+ "STATE": "MEAN_STD",
19
+ "ACTION": "MEAN_STD"
20
+ }
21
+ },
22
+ "state_file": "policy_postprocessor_step_0_unnormalizer_processor.safetensors"
23
+ },
24
+ {
25
+ "registry_name": "device_processor",
26
+ "config": {
27
+ "device": "cpu",
28
+ "float_dtype": null
29
+ }
30
+ }
31
+ ]
32
+ }
checkpoints/020000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/020000/pretrained_model/policy_preprocessor.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_preprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "rename_observations_processor",
6
+ "config": {
7
+ "rename_map": {}
8
+ }
9
+ },
10
+ {
11
+ "registry_name": "to_batch_processor",
12
+ "config": {}
13
+ },
14
+ {
15
+ "registry_name": "smolvla_new_line_processor",
16
+ "config": {}
17
+ },
18
+ {
19
+ "registry_name": "tokenizer_processor",
20
+ "config": {
21
+ "max_length": 48,
22
+ "task_key": "task",
23
+ "padding_side": "right",
24
+ "padding": "longest",
25
+ "truncation": true,
26
+ "tokenizer_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct"
27
+ }
28
+ },
29
+ {
30
+ "registry_name": "device_processor",
31
+ "config": {
32
+ "device": "cuda",
33
+ "float_dtype": null
34
+ }
35
+ },
36
+ {
37
+ "registry_name": "normalizer_processor",
38
+ "config": {
39
+ "eps": 1e-08,
40
+ "features": {
41
+ "observation.state": {
42
+ "type": "STATE",
43
+ "shape": [
44
+ 6
45
+ ]
46
+ },
47
+ "observation.images.front": {
48
+ "type": "VISUAL",
49
+ "shape": [
50
+ 3,
51
+ 480,
52
+ 640
53
+ ]
54
+ },
55
+ "action": {
56
+ "type": "ACTION",
57
+ "shape": [
58
+ 6
59
+ ]
60
+ }
61
+ },
62
+ "norm_map": {
63
+ "VISUAL": "IDENTITY",
64
+ "STATE": "MEAN_STD",
65
+ "ACTION": "MEAN_STD"
66
+ }
67
+ },
68
+ "state_file": "policy_preprocessor_step_5_normalizer_processor.safetensors"
69
+ }
70
+ ]
71
+ }
checkpoints/020000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/020000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "Dimios45/santra",
4
+ "root": "data/santra/santra_demo_20251108_102726",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ },
61
+ "affine": {
62
+ "weight": 1.0,
63
+ "type": "RandomAffine",
64
+ "kwargs": {
65
+ "degrees": [
66
+ -5.0,
67
+ 5.0
68
+ ],
69
+ "translate": [
70
+ 0.05,
71
+ 0.05
72
+ ]
73
+ }
74
+ }
75
+ }
76
+ },
77
+ "revision": null,
78
+ "use_imagenet_stats": true,
79
+ "video_backend": "torchcodec",
80
+ "streaming": false
81
+ },
82
+ "env": null,
83
+ "policy": {
84
+ "type": "smolvla",
85
+ "n_obs_steps": 1,
86
+ "input_features": {
87
+ "observation.state": {
88
+ "type": "STATE",
89
+ "shape": [
90
+ 6
91
+ ]
92
+ },
93
+ "observation.images.front": {
94
+ "type": "VISUAL",
95
+ "shape": [
96
+ 3,
97
+ 480,
98
+ 640
99
+ ]
100
+ }
101
+ },
102
+ "output_features": {
103
+ "action": {
104
+ "type": "ACTION",
105
+ "shape": [
106
+ 6
107
+ ]
108
+ }
109
+ },
110
+ "device": "cuda",
111
+ "use_amp": false,
112
+ "push_to_hub": true,
113
+ "repo_id": "Dimios45/smolvla1",
114
+ "private": null,
115
+ "tags": null,
116
+ "license": null,
117
+ "pretrained_path": null,
118
+ "chunk_size": 50,
119
+ "n_action_steps": 50,
120
+ "normalization_mapping": {
121
+ "VISUAL": "IDENTITY",
122
+ "STATE": "MEAN_STD",
123
+ "ACTION": "MEAN_STD"
124
+ },
125
+ "max_state_dim": 32,
126
+ "max_action_dim": 32,
127
+ "resize_imgs_with_padding": [
128
+ 512,
129
+ 512
130
+ ],
131
+ "empty_cameras": 0,
132
+ "adapt_to_pi_aloha": false,
133
+ "use_delta_joint_actions_aloha": false,
134
+ "tokenizer_max_length": 48,
135
+ "num_steps": 10,
136
+ "use_cache": true,
137
+ "freeze_vision_encoder": true,
138
+ "train_expert_only": true,
139
+ "train_state_proj": true,
140
+ "optimizer_lr": 0.0001,
141
+ "optimizer_betas": [
142
+ 0.9,
143
+ 0.95
144
+ ],
145
+ "optimizer_eps": 1e-08,
146
+ "optimizer_weight_decay": 1e-10,
147
+ "optimizer_grad_clip_norm": 10,
148
+ "scheduler_warmup_steps": 1000,
149
+ "scheduler_decay_steps": 30000,
150
+ "scheduler_decay_lr": 2.5e-06,
151
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
152
+ "load_vlm_weights": false,
153
+ "add_image_special_tokens": false,
154
+ "attention_mode": "cross_attn",
155
+ "prefix_length": -1,
156
+ "pad_language_to": "longest",
157
+ "num_expert_layers": -1,
158
+ "num_vlm_layers": 16,
159
+ "self_attn_every_n_layers": 2,
160
+ "expert_width_multiplier": 0.75,
161
+ "min_period": 0.004,
162
+ "max_period": 4.0
163
+ },
164
+ "output_dir": "outputs/train/smolvla1",
165
+ "job_name": "smolvla_santra",
166
+ "resume": false,
167
+ "seed": 1000,
168
+ "num_workers": 4,
169
+ "batch_size": 8,
170
+ "steps": 100000,
171
+ "eval_freq": 20000,
172
+ "log_freq": 200,
173
+ "save_checkpoint": true,
174
+ "save_freq": 20000,
175
+ "use_policy_training_preset": true,
176
+ "optimizer": {
177
+ "type": "adamw",
178
+ "lr": 0.0001,
179
+ "weight_decay": 1e-10,
180
+ "grad_clip_norm": 10,
181
+ "betas": [
182
+ 0.9,
183
+ 0.95
184
+ ],
185
+ "eps": 1e-08
186
+ },
187
+ "scheduler": {
188
+ "type": "cosine_decay_with_warmup",
189
+ "num_warmup_steps": 1000,
190
+ "num_decay_steps": 30000,
191
+ "peak_lr": 0.0001,
192
+ "decay_lr": 2.5e-06
193
+ },
194
+ "eval": {
195
+ "n_episodes": 50,
196
+ "batch_size": 50,
197
+ "use_async_envs": false
198
+ },
199
+ "wandb": {
200
+ "enable": true,
201
+ "disable_artifact": false,
202
+ "project": "lerobot",
203
+ "entity": null,
204
+ "notes": null,
205
+ "run_id": "ba762whj",
206
+ "mode": null
207
+ },
208
+ "checkpoint_path": null,
209
+ "rename_map": {}
210
+ }
checkpoints/020000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,521 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.6875000000000013e-05,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "maximize": false,
12
+ "foreach": null,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "decoupled_weight_decay": true,
17
+ "initial_lr": 0.0001,
18
+ "params": [
19
+ 0,
20
+ 1,
21
+ 2,
22
+ 3,
23
+ 4,
24
+ 5,
25
+ 6,
26
+ 7,
27
+ 8,
28
+ 9,
29
+ 10,
30
+ 11,
31
+ 12,
32
+ 13,
33
+ 14,
34
+ 15,
35
+ 16,
36
+ 17,
37
+ 18,
38
+ 19,
39
+ 20,
40
+ 21,
41
+ 22,
42
+ 23,
43
+ 24,
44
+ 25,
45
+ 26,
46
+ 27,
47
+ 28,
48
+ 29,
49
+ 30,
50
+ 31,
51
+ 32,
52
+ 33,
53
+ 34,
54
+ 35,
55
+ 36,
56
+ 37,
57
+ 38,
58
+ 39,
59
+ 40,
60
+ 41,
61
+ 42,
62
+ 43,
63
+ 44,
64
+ 45,
65
+ 46,
66
+ 47,
67
+ 48,
68
+ 49,
69
+ 50,
70
+ 51,
71
+ 52,
72
+ 53,
73
+ 54,
74
+ 55,
75
+ 56,
76
+ 57,
77
+ 58,
78
+ 59,
79
+ 60,
80
+ 61,
81
+ 62,
82
+ 63,
83
+ 64,
84
+ 65,
85
+ 66,
86
+ 67,
87
+ 68,
88
+ 69,
89
+ 70,
90
+ 71,
91
+ 72,
92
+ 73,
93
+ 74,
94
+ 75,
95
+ 76,
96
+ 77,
97
+ 78,
98
+ 79,
99
+ 80,
100
+ 81,
101
+ 82,
102
+ 83,
103
+ 84,
104
+ 85,
105
+ 86,
106
+ 87,
107
+ 88,
108
+ 89,
109
+ 90,
110
+ 91,
111
+ 92,
112
+ 93,
113
+ 94,
114
+ 95,
115
+ 96,
116
+ 97,
117
+ 98,
118
+ 99,
119
+ 100,
120
+ 101,
121
+ 102,
122
+ 103,
123
+ 104,
124
+ 105,
125
+ 106,
126
+ 107,
127
+ 108,
128
+ 109,
129
+ 110,
130
+ 111,
131
+ 112,
132
+ 113,
133
+ 114,
134
+ 115,
135
+ 116,
136
+ 117,
137
+ 118,
138
+ 119,
139
+ 120,
140
+ 121,
141
+ 122,
142
+ 123,
143
+ 124,
144
+ 125,
145
+ 126,
146
+ 127,
147
+ 128,
148
+ 129,
149
+ 130,
150
+ 131,
151
+ 132,
152
+ 133,
153
+ 134,
154
+ 135,
155
+ 136,
156
+ 137,
157
+ 138,
158
+ 139,
159
+ 140,
160
+ 141,
161
+ 142,
162
+ 143,
163
+ 144,
164
+ 145,
165
+ 146,
166
+ 147,
167
+ 148,
168
+ 149,
169
+ 150,
170
+ 151,
171
+ 152,
172
+ 153,
173
+ 154,
174
+ 155,
175
+ 156,
176
+ 157,
177
+ 158,
178
+ 159,
179
+ 160,
180
+ 161,
181
+ 162,
182
+ 163,
183
+ 164,
184
+ 165,
185
+ 166,
186
+ 167,
187
+ 168,
188
+ 169,
189
+ 170,
190
+ 171,
191
+ 172,
192
+ 173,
193
+ 174,
194
+ 175,
195
+ 176,
196
+ 177,
197
+ 178,
198
+ 179,
199
+ 180,
200
+ 181,
201
+ 182,
202
+ 183,
203
+ 184,
204
+ 185,
205
+ 186,
206
+ 187,
207
+ 188,
208
+ 189,
209
+ 190,
210
+ 191,
211
+ 192,
212
+ 193,
213
+ 194,
214
+ 195,
215
+ 196,
216
+ 197,
217
+ 198,
218
+ 199,
219
+ 200,
220
+ 201,
221
+ 202,
222
+ 203,
223
+ 204,
224
+ 205,
225
+ 206,
226
+ 207,
227
+ 208,
228
+ 209,
229
+ 210,
230
+ 211,
231
+ 212,
232
+ 213,
233
+ 214,
234
+ 215,
235
+ 216,
236
+ 217,
237
+ 218,
238
+ 219,
239
+ 220,
240
+ 221,
241
+ 222,
242
+ 223,
243
+ 224,
244
+ 225,
245
+ 226,
246
+ 227,
247
+ 228,
248
+ 229,
249
+ 230,
250
+ 231,
251
+ 232,
252
+ 233,
253
+ 234,
254
+ 235,
255
+ 236,
256
+ 237,
257
+ 238,
258
+ 239,
259
+ 240,
260
+ 241,
261
+ 242,
262
+ 243,
263
+ 244,
264
+ 245,
265
+ 246,
266
+ 247,
267
+ 248,
268
+ 249,
269
+ 250,
270
+ 251,
271
+ 252,
272
+ 253,
273
+ 254,
274
+ 255,
275
+ 256,
276
+ 257,
277
+ 258,
278
+ 259,
279
+ 260,
280
+ 261,
281
+ 262,
282
+ 263,
283
+ 264,
284
+ 265,
285
+ 266,
286
+ 267,
287
+ 268,
288
+ 269,
289
+ 270,
290
+ 271,
291
+ 272,
292
+ 273,
293
+ 274,
294
+ 275,
295
+ 276,
296
+ 277,
297
+ 278,
298
+ 279,
299
+ 280,
300
+ 281,
301
+ 282,
302
+ 283,
303
+ 284,
304
+ 285,
305
+ 286,
306
+ 287,
307
+ 288,
308
+ 289,
309
+ 290,
310
+ 291,
311
+ 292,
312
+ 293,
313
+ 294,
314
+ 295,
315
+ 296,
316
+ 297,
317
+ 298,
318
+ 299,
319
+ 300,
320
+ 301,
321
+ 302,
322
+ 303,
323
+ 304,
324
+ 305,
325
+ 306,
326
+ 307,
327
+ 308,
328
+ 309,
329
+ 310,
330
+ 311,
331
+ 312,
332
+ 313,
333
+ 314,
334
+ 315,
335
+ 316,
336
+ 317,
337
+ 318,
338
+ 319,
339
+ 320,
340
+ 321,
341
+ 322,
342
+ 323,
343
+ 324,
344
+ 325,
345
+ 326,
346
+ 327,
347
+ 328,
348
+ 329,
349
+ 330,
350
+ 331,
351
+ 332,
352
+ 333,
353
+ 334,
354
+ 335,
355
+ 336,
356
+ 337,
357
+ 338,
358
+ 339,
359
+ 340,
360
+ 341,
361
+ 342,
362
+ 343,
363
+ 344,
364
+ 345,
365
+ 346,
366
+ 347,
367
+ 348,
368
+ 349,
369
+ 350,
370
+ 351,
371
+ 352,
372
+ 353,
373
+ 354,
374
+ 355,
375
+ 356,
376
+ 357,
377
+ 358,
378
+ 359,
379
+ 360,
380
+ 361,
381
+ 362,
382
+ 363,
383
+ 364,
384
+ 365,
385
+ 366,
386
+ 367,
387
+ 368,
388
+ 369,
389
+ 370,
390
+ 371,
391
+ 372,
392
+ 373,
393
+ 374,
394
+ 375,
395
+ 376,
396
+ 377,
397
+ 378,
398
+ 379,
399
+ 380,
400
+ 381,
401
+ 382,
402
+ 383,
403
+ 384,
404
+ 385,
405
+ 386,
406
+ 387,
407
+ 388,
408
+ 389,
409
+ 390,
410
+ 391,
411
+ 392,
412
+ 393,
413
+ 394,
414
+ 395,
415
+ 396,
416
+ 397,
417
+ 398,
418
+ 399,
419
+ 400,
420
+ 401,
421
+ 402,
422
+ 403,
423
+ 404,
424
+ 405,
425
+ 406,
426
+ 407,
427
+ 408,
428
+ 409,
429
+ 410,
430
+ 411,
431
+ 412,
432
+ 413,
433
+ 414,
434
+ 415,
435
+ 416,
436
+ 417,
437
+ 418,
438
+ 419,
439
+ 420,
440
+ 421,
441
+ 422,
442
+ 423,
443
+ 424,
444
+ 425,
445
+ 426,
446
+ 427,
447
+ 428,
448
+ 429,
449
+ 430,
450
+ 431,
451
+ 432,
452
+ 433,
453
+ 434,
454
+ 435,
455
+ 436,
456
+ 437,
457
+ 438,
458
+ 439,
459
+ 440,
460
+ 441,
461
+ 442,
462
+ 443,
463
+ 444,
464
+ 445,
465
+ 446,
466
+ 447,
467
+ 448,
468
+ 449,
469
+ 450,
470
+ 451,
471
+ 452,
472
+ 453,
473
+ 454,
474
+ 455,
475
+ 456,
476
+ 457,
477
+ 458,
478
+ 459,
479
+ 460,
480
+ 461,
481
+ 462,
482
+ 463,
483
+ 464,
484
+ 465,
485
+ 466,
486
+ 467,
487
+ 468,
488
+ 469,
489
+ 470,
490
+ 471,
491
+ 472,
492
+ 473,
493
+ 474,
494
+ 475,
495
+ 476,
496
+ 477,
497
+ 478,
498
+ 479,
499
+ 480,
500
+ 481,
501
+ 482,
502
+ 483,
503
+ 484,
504
+ 485,
505
+ 486,
506
+ 487,
507
+ 488,
508
+ 489,
509
+ 490,
510
+ 491,
511
+ 492,
512
+ 493,
513
+ 494,
514
+ 495,
515
+ 496,
516
+ 497,
517
+ 498,
518
+ 499
519
+ ]
520
+ }
521
+ ]
checkpoints/020000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad8fe07375f919e9869ee2fcb88123a5b33137b8097957f99c8be8dac25da99a
3
+ size 412659164
checkpoints/020000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0d21412f96be6221d1bf9960933852424aa861c5d536b6aad596173cfe94fde
3
+ size 15708
checkpoints/020000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 20000,
6
+ "_step_count": 20001,
7
+ "_get_lr_called_within_step": false,
8
+ "_last_lr": [
9
+ 2.6875000000000013e-05
10
+ ],
11
+ "lr_lambdas": [
12
+ null
13
+ ]
14
+ }
checkpoints/020000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 20000
3
+ }
checkpoints/040000/pretrained_model/config.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 6
9
+ ]
10
+ },
11
+ "observation.images.front": {
12
+ "type": "VISUAL",
13
+ "shape": [
14
+ 3,
15
+ 480,
16
+ 640
17
+ ]
18
+ }
19
+ },
20
+ "output_features": {
21
+ "action": {
22
+ "type": "ACTION",
23
+ "shape": [
24
+ 6
25
+ ]
26
+ }
27
+ },
28
+ "device": "cuda",
29
+ "use_amp": false,
30
+ "push_to_hub": true,
31
+ "repo_id": "Dimios45/smolvla1",
32
+ "private": null,
33
+ "tags": null,
34
+ "license": null,
35
+ "pretrained_path": "outputs/train/smolvla1/checkpoints/last/pretrained_model",
36
+ "chunk_size": 50,
37
+ "n_action_steps": 50,
38
+ "normalization_mapping": {
39
+ "VISUAL": "IDENTITY",
40
+ "STATE": "MEAN_STD",
41
+ "ACTION": "MEAN_STD"
42
+ },
43
+ "max_state_dim": 32,
44
+ "max_action_dim": 32,
45
+ "resize_imgs_with_padding": [
46
+ 512,
47
+ 512
48
+ ],
49
+ "empty_cameras": 0,
50
+ "adapt_to_pi_aloha": false,
51
+ "use_delta_joint_actions_aloha": false,
52
+ "tokenizer_max_length": 48,
53
+ "num_steps": 10,
54
+ "use_cache": true,
55
+ "freeze_vision_encoder": true,
56
+ "train_expert_only": true,
57
+ "train_state_proj": true,
58
+ "optimizer_lr": 0.0001,
59
+ "optimizer_betas": [
60
+ 0.9,
61
+ 0.95
62
+ ],
63
+ "optimizer_eps": 1e-08,
64
+ "optimizer_weight_decay": 1e-10,
65
+ "optimizer_grad_clip_norm": 10.0,
66
+ "scheduler_warmup_steps": 1000,
67
+ "scheduler_decay_steps": 30000,
68
+ "scheduler_decay_lr": 2.5e-06,
69
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
70
+ "load_vlm_weights": false,
71
+ "add_image_special_tokens": false,
72
+ "attention_mode": "cross_attn",
73
+ "prefix_length": -1,
74
+ "pad_language_to": "longest",
75
+ "num_expert_layers": -1,
76
+ "num_vlm_layers": 16,
77
+ "self_attn_every_n_layers": 2,
78
+ "expert_width_multiplier": 0.75,
79
+ "min_period": 0.004,
80
+ "max_period": 4.0
81
+ }
checkpoints/040000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef2b03f84fc8269ee2dacbab7c24099076984b6280d02b426670f96e80bb80d4
3
+ size 1197789224
checkpoints/040000/pretrained_model/policy_postprocessor.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_postprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "unnormalizer_processor",
6
+ "config": {
7
+ "eps": 1e-08,
8
+ "features": {
9
+ "action": {
10
+ "type": "ACTION",
11
+ "shape": [
12
+ 6
13
+ ]
14
+ }
15
+ },
16
+ "norm_map": {
17
+ "VISUAL": "IDENTITY",
18
+ "STATE": "MEAN_STD",
19
+ "ACTION": "MEAN_STD"
20
+ }
21
+ },
22
+ "state_file": "policy_postprocessor_step_0_unnormalizer_processor.safetensors"
23
+ },
24
+ {
25
+ "registry_name": "device_processor",
26
+ "config": {
27
+ "device": "cpu",
28
+ "float_dtype": null
29
+ }
30
+ }
31
+ ]
32
+ }
checkpoints/040000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/040000/pretrained_model/policy_preprocessor.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_preprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "rename_observations_processor",
6
+ "config": {
7
+ "rename_map": {}
8
+ }
9
+ },
10
+ {
11
+ "registry_name": "to_batch_processor",
12
+ "config": {}
13
+ },
14
+ {
15
+ "registry_name": "smolvla_new_line_processor",
16
+ "config": {}
17
+ },
18
+ {
19
+ "registry_name": "tokenizer_processor",
20
+ "config": {
21
+ "max_length": 48,
22
+ "task_key": "task",
23
+ "padding_side": "right",
24
+ "padding": "longest",
25
+ "truncation": true,
26
+ "tokenizer_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct"
27
+ }
28
+ },
29
+ {
30
+ "registry_name": "device_processor",
31
+ "config": {
32
+ "device": "cuda",
33
+ "float_dtype": null
34
+ }
35
+ },
36
+ {
37
+ "registry_name": "normalizer_processor",
38
+ "config": {
39
+ "eps": 1e-08,
40
+ "features": {
41
+ "observation.state": {
42
+ "type": "STATE",
43
+ "shape": [
44
+ 6
45
+ ]
46
+ },
47
+ "observation.images.front": {
48
+ "type": "VISUAL",
49
+ "shape": [
50
+ 3,
51
+ 480,
52
+ 640
53
+ ]
54
+ },
55
+ "action": {
56
+ "type": "ACTION",
57
+ "shape": [
58
+ 6
59
+ ]
60
+ }
61
+ },
62
+ "norm_map": {
63
+ "VISUAL": "IDENTITY",
64
+ "STATE": "MEAN_STD",
65
+ "ACTION": "MEAN_STD"
66
+ }
67
+ },
68
+ "state_file": "policy_preprocessor_step_5_normalizer_processor.safetensors"
69
+ }
70
+ ]
71
+ }
checkpoints/040000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/040000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "Dimios45/santra",
4
+ "root": "data/santra/santra_demo_20251108_102726",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ },
61
+ "affine": {
62
+ "weight": 1.0,
63
+ "type": "RandomAffine",
64
+ "kwargs": {
65
+ "degrees": [
66
+ -5.0,
67
+ 5.0
68
+ ],
69
+ "translate": [
70
+ 0.05,
71
+ 0.05
72
+ ]
73
+ }
74
+ }
75
+ }
76
+ },
77
+ "revision": null,
78
+ "use_imagenet_stats": true,
79
+ "video_backend": "torchcodec",
80
+ "streaming": false
81
+ },
82
+ "env": null,
83
+ "policy": {
84
+ "type": "smolvla",
85
+ "n_obs_steps": 1,
86
+ "input_features": {
87
+ "observation.state": {
88
+ "type": "STATE",
89
+ "shape": [
90
+ 6
91
+ ]
92
+ },
93
+ "observation.images.front": {
94
+ "type": "VISUAL",
95
+ "shape": [
96
+ 3,
97
+ 480,
98
+ 640
99
+ ]
100
+ }
101
+ },
102
+ "output_features": {
103
+ "action": {
104
+ "type": "ACTION",
105
+ "shape": [
106
+ 6
107
+ ]
108
+ }
109
+ },
110
+ "device": "cuda",
111
+ "use_amp": false,
112
+ "push_to_hub": true,
113
+ "repo_id": "Dimios45/smolvla1",
114
+ "private": null,
115
+ "tags": null,
116
+ "license": null,
117
+ "pretrained_path": "outputs/train/smolvla1/checkpoints/last/pretrained_model",
118
+ "chunk_size": 50,
119
+ "n_action_steps": 50,
120
+ "normalization_mapping": {
121
+ "VISUAL": "IDENTITY",
122
+ "STATE": "MEAN_STD",
123
+ "ACTION": "MEAN_STD"
124
+ },
125
+ "max_state_dim": 32,
126
+ "max_action_dim": 32,
127
+ "resize_imgs_with_padding": [
128
+ 512,
129
+ 512
130
+ ],
131
+ "empty_cameras": 0,
132
+ "adapt_to_pi_aloha": false,
133
+ "use_delta_joint_actions_aloha": false,
134
+ "tokenizer_max_length": 48,
135
+ "num_steps": 10,
136
+ "use_cache": true,
137
+ "freeze_vision_encoder": true,
138
+ "train_expert_only": true,
139
+ "train_state_proj": true,
140
+ "optimizer_lr": 0.0001,
141
+ "optimizer_betas": [
142
+ 0.9,
143
+ 0.95
144
+ ],
145
+ "optimizer_eps": 1e-08,
146
+ "optimizer_weight_decay": 1e-10,
147
+ "optimizer_grad_clip_norm": 10.0,
148
+ "scheduler_warmup_steps": 1000,
149
+ "scheduler_decay_steps": 30000,
150
+ "scheduler_decay_lr": 2.5e-06,
151
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
152
+ "load_vlm_weights": false,
153
+ "add_image_special_tokens": false,
154
+ "attention_mode": "cross_attn",
155
+ "prefix_length": -1,
156
+ "pad_language_to": "longest",
157
+ "num_expert_layers": -1,
158
+ "num_vlm_layers": 16,
159
+ "self_attn_every_n_layers": 2,
160
+ "expert_width_multiplier": 0.75,
161
+ "min_period": 0.004,
162
+ "max_period": 4.0
163
+ },
164
+ "output_dir": "outputs/train/smolvla1",
165
+ "job_name": "smolvla_santra",
166
+ "resume": true,
167
+ "seed": 1000,
168
+ "num_workers": 4,
169
+ "batch_size": 8,
170
+ "steps": 100000,
171
+ "eval_freq": 20000,
172
+ "log_freq": 200,
173
+ "save_checkpoint": true,
174
+ "save_freq": 20000,
175
+ "use_policy_training_preset": true,
176
+ "optimizer": {
177
+ "type": "adamw",
178
+ "lr": 0.0001,
179
+ "weight_decay": 1e-10,
180
+ "grad_clip_norm": 10.0,
181
+ "betas": [
182
+ 0.9,
183
+ 0.95
184
+ ],
185
+ "eps": 1e-08
186
+ },
187
+ "scheduler": {
188
+ "type": "cosine_decay_with_warmup",
189
+ "num_warmup_steps": 1000,
190
+ "num_decay_steps": 30000,
191
+ "peak_lr": 0.0001,
192
+ "decay_lr": 2.5e-06
193
+ },
194
+ "eval": {
195
+ "n_episodes": 50,
196
+ "batch_size": 50,
197
+ "use_async_envs": false
198
+ },
199
+ "wandb": {
200
+ "enable": true,
201
+ "disable_artifact": false,
202
+ "project": "lerobot",
203
+ "entity": null,
204
+ "notes": null,
205
+ "run_id": "ba762whj",
206
+ "mode": null
207
+ },
208
+ "checkpoint_path": "outputs/train/smolvla1/checkpoints/last",
209
+ "rename_map": {}
210
+ }
checkpoints/040000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,521 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.5e-06,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "maximize": false,
12
+ "foreach": null,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "decoupled_weight_decay": true,
17
+ "initial_lr": 0.0001,
18
+ "params": [
19
+ 0,
20
+ 1,
21
+ 2,
22
+ 3,
23
+ 4,
24
+ 5,
25
+ 6,
26
+ 7,
27
+ 8,
28
+ 9,
29
+ 10,
30
+ 11,
31
+ 12,
32
+ 13,
33
+ 14,
34
+ 15,
35
+ 16,
36
+ 17,
37
+ 18,
38
+ 19,
39
+ 20,
40
+ 21,
41
+ 22,
42
+ 23,
43
+ 24,
44
+ 25,
45
+ 26,
46
+ 27,
47
+ 28,
48
+ 29,
49
+ 30,
50
+ 31,
51
+ 32,
52
+ 33,
53
+ 34,
54
+ 35,
55
+ 36,
56
+ 37,
57
+ 38,
58
+ 39,
59
+ 40,
60
+ 41,
61
+ 42,
62
+ 43,
63
+ 44,
64
+ 45,
65
+ 46,
66
+ 47,
67
+ 48,
68
+ 49,
69
+ 50,
70
+ 51,
71
+ 52,
72
+ 53,
73
+ 54,
74
+ 55,
75
+ 56,
76
+ 57,
77
+ 58,
78
+ 59,
79
+ 60,
80
+ 61,
81
+ 62,
82
+ 63,
83
+ 64,
84
+ 65,
85
+ 66,
86
+ 67,
87
+ 68,
88
+ 69,
89
+ 70,
90
+ 71,
91
+ 72,
92
+ 73,
93
+ 74,
94
+ 75,
95
+ 76,
96
+ 77,
97
+ 78,
98
+ 79,
99
+ 80,
100
+ 81,
101
+ 82,
102
+ 83,
103
+ 84,
104
+ 85,
105
+ 86,
106
+ 87,
107
+ 88,
108
+ 89,
109
+ 90,
110
+ 91,
111
+ 92,
112
+ 93,
113
+ 94,
114
+ 95,
115
+ 96,
116
+ 97,
117
+ 98,
118
+ 99,
119
+ 100,
120
+ 101,
121
+ 102,
122
+ 103,
123
+ 104,
124
+ 105,
125
+ 106,
126
+ 107,
127
+ 108,
128
+ 109,
129
+ 110,
130
+ 111,
131
+ 112,
132
+ 113,
133
+ 114,
134
+ 115,
135
+ 116,
136
+ 117,
137
+ 118,
138
+ 119,
139
+ 120,
140
+ 121,
141
+ 122,
142
+ 123,
143
+ 124,
144
+ 125,
145
+ 126,
146
+ 127,
147
+ 128,
148
+ 129,
149
+ 130,
150
+ 131,
151
+ 132,
152
+ 133,
153
+ 134,
154
+ 135,
155
+ 136,
156
+ 137,
157
+ 138,
158
+ 139,
159
+ 140,
160
+ 141,
161
+ 142,
162
+ 143,
163
+ 144,
164
+ 145,
165
+ 146,
166
+ 147,
167
+ 148,
168
+ 149,
169
+ 150,
170
+ 151,
171
+ 152,
172
+ 153,
173
+ 154,
174
+ 155,
175
+ 156,
176
+ 157,
177
+ 158,
178
+ 159,
179
+ 160,
180
+ 161,
181
+ 162,
182
+ 163,
183
+ 164,
184
+ 165,
185
+ 166,
186
+ 167,
187
+ 168,
188
+ 169,
189
+ 170,
190
+ 171,
191
+ 172,
192
+ 173,
193
+ 174,
194
+ 175,
195
+ 176,
196
+ 177,
197
+ 178,
198
+ 179,
199
+ 180,
200
+ 181,
201
+ 182,
202
+ 183,
203
+ 184,
204
+ 185,
205
+ 186,
206
+ 187,
207
+ 188,
208
+ 189,
209
+ 190,
210
+ 191,
211
+ 192,
212
+ 193,
213
+ 194,
214
+ 195,
215
+ 196,
216
+ 197,
217
+ 198,
218
+ 199,
219
+ 200,
220
+ 201,
221
+ 202,
222
+ 203,
223
+ 204,
224
+ 205,
225
+ 206,
226
+ 207,
227
+ 208,
228
+ 209,
229
+ 210,
230
+ 211,
231
+ 212,
232
+ 213,
233
+ 214,
234
+ 215,
235
+ 216,
236
+ 217,
237
+ 218,
238
+ 219,
239
+ 220,
240
+ 221,
241
+ 222,
242
+ 223,
243
+ 224,
244
+ 225,
245
+ 226,
246
+ 227,
247
+ 228,
248
+ 229,
249
+ 230,
250
+ 231,
251
+ 232,
252
+ 233,
253
+ 234,
254
+ 235,
255
+ 236,
256
+ 237,
257
+ 238,
258
+ 239,
259
+ 240,
260
+ 241,
261
+ 242,
262
+ 243,
263
+ 244,
264
+ 245,
265
+ 246,
266
+ 247,
267
+ 248,
268
+ 249,
269
+ 250,
270
+ 251,
271
+ 252,
272
+ 253,
273
+ 254,
274
+ 255,
275
+ 256,
276
+ 257,
277
+ 258,
278
+ 259,
279
+ 260,
280
+ 261,
281
+ 262,
282
+ 263,
283
+ 264,
284
+ 265,
285
+ 266,
286
+ 267,
287
+ 268,
288
+ 269,
289
+ 270,
290
+ 271,
291
+ 272,
292
+ 273,
293
+ 274,
294
+ 275,
295
+ 276,
296
+ 277,
297
+ 278,
298
+ 279,
299
+ 280,
300
+ 281,
301
+ 282,
302
+ 283,
303
+ 284,
304
+ 285,
305
+ 286,
306
+ 287,
307
+ 288,
308
+ 289,
309
+ 290,
310
+ 291,
311
+ 292,
312
+ 293,
313
+ 294,
314
+ 295,
315
+ 296,
316
+ 297,
317
+ 298,
318
+ 299,
319
+ 300,
320
+ 301,
321
+ 302,
322
+ 303,
323
+ 304,
324
+ 305,
325
+ 306,
326
+ 307,
327
+ 308,
328
+ 309,
329
+ 310,
330
+ 311,
331
+ 312,
332
+ 313,
333
+ 314,
334
+ 315,
335
+ 316,
336
+ 317,
337
+ 318,
338
+ 319,
339
+ 320,
340
+ 321,
341
+ 322,
342
+ 323,
343
+ 324,
344
+ 325,
345
+ 326,
346
+ 327,
347
+ 328,
348
+ 329,
349
+ 330,
350
+ 331,
351
+ 332,
352
+ 333,
353
+ 334,
354
+ 335,
355
+ 336,
356
+ 337,
357
+ 338,
358
+ 339,
359
+ 340,
360
+ 341,
361
+ 342,
362
+ 343,
363
+ 344,
364
+ 345,
365
+ 346,
366
+ 347,
367
+ 348,
368
+ 349,
369
+ 350,
370
+ 351,
371
+ 352,
372
+ 353,
373
+ 354,
374
+ 355,
375
+ 356,
376
+ 357,
377
+ 358,
378
+ 359,
379
+ 360,
380
+ 361,
381
+ 362,
382
+ 363,
383
+ 364,
384
+ 365,
385
+ 366,
386
+ 367,
387
+ 368,
388
+ 369,
389
+ 370,
390
+ 371,
391
+ 372,
392
+ 373,
393
+ 374,
394
+ 375,
395
+ 376,
396
+ 377,
397
+ 378,
398
+ 379,
399
+ 380,
400
+ 381,
401
+ 382,
402
+ 383,
403
+ 384,
404
+ 385,
405
+ 386,
406
+ 387,
407
+ 388,
408
+ 389,
409
+ 390,
410
+ 391,
411
+ 392,
412
+ 393,
413
+ 394,
414
+ 395,
415
+ 396,
416
+ 397,
417
+ 398,
418
+ 399,
419
+ 400,
420
+ 401,
421
+ 402,
422
+ 403,
423
+ 404,
424
+ 405,
425
+ 406,
426
+ 407,
427
+ 408,
428
+ 409,
429
+ 410,
430
+ 411,
431
+ 412,
432
+ 413,
433
+ 414,
434
+ 415,
435
+ 416,
436
+ 417,
437
+ 418,
438
+ 419,
439
+ 420,
440
+ 421,
441
+ 422,
442
+ 423,
443
+ 424,
444
+ 425,
445
+ 426,
446
+ 427,
447
+ 428,
448
+ 429,
449
+ 430,
450
+ 431,
451
+ 432,
452
+ 433,
453
+ 434,
454
+ 435,
455
+ 436,
456
+ 437,
457
+ 438,
458
+ 439,
459
+ 440,
460
+ 441,
461
+ 442,
462
+ 443,
463
+ 444,
464
+ 445,
465
+ 446,
466
+ 447,
467
+ 448,
468
+ 449,
469
+ 450,
470
+ 451,
471
+ 452,
472
+ 453,
473
+ 454,
474
+ 455,
475
+ 456,
476
+ 457,
477
+ 458,
478
+ 459,
479
+ 460,
480
+ 461,
481
+ 462,
482
+ 463,
483
+ 464,
484
+ 465,
485
+ 466,
486
+ 467,
487
+ 468,
488
+ 469,
489
+ 470,
490
+ 471,
491
+ 472,
492
+ 473,
493
+ 474,
494
+ 475,
495
+ 476,
496
+ 477,
497
+ 478,
498
+ 479,
499
+ 480,
500
+ 481,
501
+ 482,
502
+ 483,
503
+ 484,
504
+ 485,
505
+ 486,
506
+ 487,
507
+ 488,
508
+ 489,
509
+ 490,
510
+ 491,
511
+ 492,
512
+ 493,
513
+ 494,
514
+ 495,
515
+ 496,
516
+ 497,
517
+ 498,
518
+ 499
519
+ ]
520
+ }
521
+ ]
checkpoints/040000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb279e3757486d196d1b5e5df6f36a42b66ebd4061b8eff50aa6cf7b0eec8a46
3
+ size 412659164
checkpoints/040000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f6c1f756d08f6cae5cf4a603904e857dd980c3eca85366071dcbdd83dc5b14d
3
+ size 15708
checkpoints/040000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 40000,
6
+ "_step_count": 40001,
7
+ "_get_lr_called_within_step": false,
8
+ "_last_lr": [
9
+ 2.5e-06
10
+ ],
11
+ "lr_lambdas": [
12
+ null
13
+ ]
14
+ }
checkpoints/040000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 40000
3
+ }
checkpoints/060000/pretrained_model/config.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 6
9
+ ]
10
+ },
11
+ "observation.images.front": {
12
+ "type": "VISUAL",
13
+ "shape": [
14
+ 3,
15
+ 480,
16
+ 640
17
+ ]
18
+ }
19
+ },
20
+ "output_features": {
21
+ "action": {
22
+ "type": "ACTION",
23
+ "shape": [
24
+ 6
25
+ ]
26
+ }
27
+ },
28
+ "device": "cuda",
29
+ "use_amp": false,
30
+ "push_to_hub": true,
31
+ "repo_id": "Dimios45/smolvla1",
32
+ "private": null,
33
+ "tags": null,
34
+ "license": null,
35
+ "pretrained_path": "outputs/train/smolvla1/checkpoints/last/pretrained_model",
36
+ "chunk_size": 50,
37
+ "n_action_steps": 50,
38
+ "normalization_mapping": {
39
+ "VISUAL": "IDENTITY",
40
+ "STATE": "MEAN_STD",
41
+ "ACTION": "MEAN_STD"
42
+ },
43
+ "max_state_dim": 32,
44
+ "max_action_dim": 32,
45
+ "resize_imgs_with_padding": [
46
+ 512,
47
+ 512
48
+ ],
49
+ "empty_cameras": 0,
50
+ "adapt_to_pi_aloha": false,
51
+ "use_delta_joint_actions_aloha": false,
52
+ "tokenizer_max_length": 48,
53
+ "num_steps": 10,
54
+ "use_cache": true,
55
+ "freeze_vision_encoder": true,
56
+ "train_expert_only": true,
57
+ "train_state_proj": true,
58
+ "optimizer_lr": 0.0001,
59
+ "optimizer_betas": [
60
+ 0.9,
61
+ 0.95
62
+ ],
63
+ "optimizer_eps": 1e-08,
64
+ "optimizer_weight_decay": 1e-10,
65
+ "optimizer_grad_clip_norm": 10.0,
66
+ "scheduler_warmup_steps": 1000,
67
+ "scheduler_decay_steps": 30000,
68
+ "scheduler_decay_lr": 2.5e-06,
69
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
70
+ "load_vlm_weights": false,
71
+ "add_image_special_tokens": false,
72
+ "attention_mode": "cross_attn",
73
+ "prefix_length": -1,
74
+ "pad_language_to": "longest",
75
+ "num_expert_layers": -1,
76
+ "num_vlm_layers": 16,
77
+ "self_attn_every_n_layers": 2,
78
+ "expert_width_multiplier": 0.75,
79
+ "min_period": 0.004,
80
+ "max_period": 4.0
81
+ }
checkpoints/060000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:291ad583d809c43f60344d57d09c9cc1ca28298ad4fa259f5590863f2aa78cbb
3
+ size 1197789224
checkpoints/060000/pretrained_model/policy_postprocessor.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_postprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "unnormalizer_processor",
6
+ "config": {
7
+ "eps": 1e-08,
8
+ "features": {
9
+ "action": {
10
+ "type": "ACTION",
11
+ "shape": [
12
+ 6
13
+ ]
14
+ }
15
+ },
16
+ "norm_map": {
17
+ "VISUAL": "IDENTITY",
18
+ "STATE": "MEAN_STD",
19
+ "ACTION": "MEAN_STD"
20
+ }
21
+ },
22
+ "state_file": "policy_postprocessor_step_0_unnormalizer_processor.safetensors"
23
+ },
24
+ {
25
+ "registry_name": "device_processor",
26
+ "config": {
27
+ "device": "cpu",
28
+ "float_dtype": null
29
+ }
30
+ }
31
+ ]
32
+ }
checkpoints/060000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/060000/pretrained_model/policy_preprocessor.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_preprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "rename_observations_processor",
6
+ "config": {
7
+ "rename_map": {}
8
+ }
9
+ },
10
+ {
11
+ "registry_name": "to_batch_processor",
12
+ "config": {}
13
+ },
14
+ {
15
+ "registry_name": "smolvla_new_line_processor",
16
+ "config": {}
17
+ },
18
+ {
19
+ "registry_name": "tokenizer_processor",
20
+ "config": {
21
+ "max_length": 48,
22
+ "task_key": "task",
23
+ "padding_side": "right",
24
+ "padding": "longest",
25
+ "truncation": true,
26
+ "tokenizer_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct"
27
+ }
28
+ },
29
+ {
30
+ "registry_name": "device_processor",
31
+ "config": {
32
+ "device": "cuda",
33
+ "float_dtype": null
34
+ }
35
+ },
36
+ {
37
+ "registry_name": "normalizer_processor",
38
+ "config": {
39
+ "eps": 1e-08,
40
+ "features": {
41
+ "observation.state": {
42
+ "type": "STATE",
43
+ "shape": [
44
+ 6
45
+ ]
46
+ },
47
+ "observation.images.front": {
48
+ "type": "VISUAL",
49
+ "shape": [
50
+ 3,
51
+ 480,
52
+ 640
53
+ ]
54
+ },
55
+ "action": {
56
+ "type": "ACTION",
57
+ "shape": [
58
+ 6
59
+ ]
60
+ }
61
+ },
62
+ "norm_map": {
63
+ "VISUAL": "IDENTITY",
64
+ "STATE": "MEAN_STD",
65
+ "ACTION": "MEAN_STD"
66
+ }
67
+ },
68
+ "state_file": "policy_preprocessor_step_5_normalizer_processor.safetensors"
69
+ }
70
+ ]
71
+ }
checkpoints/060000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/060000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "Dimios45/santra",
4
+ "root": "data/santra/santra_demo_20251108_102726",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ },
61
+ "affine": {
62
+ "weight": 1.0,
63
+ "type": "RandomAffine",
64
+ "kwargs": {
65
+ "degrees": [
66
+ -5.0,
67
+ 5.0
68
+ ],
69
+ "translate": [
70
+ 0.05,
71
+ 0.05
72
+ ]
73
+ }
74
+ }
75
+ }
76
+ },
77
+ "revision": null,
78
+ "use_imagenet_stats": true,
79
+ "video_backend": "torchcodec",
80
+ "streaming": false
81
+ },
82
+ "env": null,
83
+ "policy": {
84
+ "type": "smolvla",
85
+ "n_obs_steps": 1,
86
+ "input_features": {
87
+ "observation.state": {
88
+ "type": "STATE",
89
+ "shape": [
90
+ 6
91
+ ]
92
+ },
93
+ "observation.images.front": {
94
+ "type": "VISUAL",
95
+ "shape": [
96
+ 3,
97
+ 480,
98
+ 640
99
+ ]
100
+ }
101
+ },
102
+ "output_features": {
103
+ "action": {
104
+ "type": "ACTION",
105
+ "shape": [
106
+ 6
107
+ ]
108
+ }
109
+ },
110
+ "device": "cuda",
111
+ "use_amp": false,
112
+ "push_to_hub": true,
113
+ "repo_id": "Dimios45/smolvla1",
114
+ "private": null,
115
+ "tags": null,
116
+ "license": null,
117
+ "pretrained_path": "outputs/train/smolvla1/checkpoints/last/pretrained_model",
118
+ "chunk_size": 50,
119
+ "n_action_steps": 50,
120
+ "normalization_mapping": {
121
+ "VISUAL": "IDENTITY",
122
+ "STATE": "MEAN_STD",
123
+ "ACTION": "MEAN_STD"
124
+ },
125
+ "max_state_dim": 32,
126
+ "max_action_dim": 32,
127
+ "resize_imgs_with_padding": [
128
+ 512,
129
+ 512
130
+ ],
131
+ "empty_cameras": 0,
132
+ "adapt_to_pi_aloha": false,
133
+ "use_delta_joint_actions_aloha": false,
134
+ "tokenizer_max_length": 48,
135
+ "num_steps": 10,
136
+ "use_cache": true,
137
+ "freeze_vision_encoder": true,
138
+ "train_expert_only": true,
139
+ "train_state_proj": true,
140
+ "optimizer_lr": 0.0001,
141
+ "optimizer_betas": [
142
+ 0.9,
143
+ 0.95
144
+ ],
145
+ "optimizer_eps": 1e-08,
146
+ "optimizer_weight_decay": 1e-10,
147
+ "optimizer_grad_clip_norm": 10.0,
148
+ "scheduler_warmup_steps": 1000,
149
+ "scheduler_decay_steps": 30000,
150
+ "scheduler_decay_lr": 2.5e-06,
151
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
152
+ "load_vlm_weights": false,
153
+ "add_image_special_tokens": false,
154
+ "attention_mode": "cross_attn",
155
+ "prefix_length": -1,
156
+ "pad_language_to": "longest",
157
+ "num_expert_layers": -1,
158
+ "num_vlm_layers": 16,
159
+ "self_attn_every_n_layers": 2,
160
+ "expert_width_multiplier": 0.75,
161
+ "min_period": 0.004,
162
+ "max_period": 4.0
163
+ },
164
+ "output_dir": "outputs/train/smolvla1",
165
+ "job_name": "smolvla_santra",
166
+ "resume": true,
167
+ "seed": 1000,
168
+ "num_workers": 4,
169
+ "batch_size": 8,
170
+ "steps": 100000,
171
+ "eval_freq": 20000,
172
+ "log_freq": 200,
173
+ "save_checkpoint": true,
174
+ "save_freq": 20000,
175
+ "use_policy_training_preset": true,
176
+ "optimizer": {
177
+ "type": "adamw",
178
+ "lr": 0.0001,
179
+ "weight_decay": 1e-10,
180
+ "grad_clip_norm": 10.0,
181
+ "betas": [
182
+ 0.9,
183
+ 0.95
184
+ ],
185
+ "eps": 1e-08
186
+ },
187
+ "scheduler": {
188
+ "type": "cosine_decay_with_warmup",
189
+ "num_warmup_steps": 1000,
190
+ "num_decay_steps": 30000,
191
+ "peak_lr": 0.0001,
192
+ "decay_lr": 2.5e-06
193
+ },
194
+ "eval": {
195
+ "n_episodes": 50,
196
+ "batch_size": 50,
197
+ "use_async_envs": false
198
+ },
199
+ "wandb": {
200
+ "enable": true,
201
+ "disable_artifact": false,
202
+ "project": "lerobot",
203
+ "entity": null,
204
+ "notes": null,
205
+ "run_id": "ba762whj",
206
+ "mode": null
207
+ },
208
+ "checkpoint_path": "outputs/train/smolvla1/checkpoints/last",
209
+ "rename_map": {}
210
+ }
checkpoints/060000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,521 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.5e-06,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "maximize": false,
12
+ "foreach": null,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "decoupled_weight_decay": true,
17
+ "initial_lr": 0.0001,
18
+ "params": [
19
+ 0,
20
+ 1,
21
+ 2,
22
+ 3,
23
+ 4,
24
+ 5,
25
+ 6,
26
+ 7,
27
+ 8,
28
+ 9,
29
+ 10,
30
+ 11,
31
+ 12,
32
+ 13,
33
+ 14,
34
+ 15,
35
+ 16,
36
+ 17,
37
+ 18,
38
+ 19,
39
+ 20,
40
+ 21,
41
+ 22,
42
+ 23,
43
+ 24,
44
+ 25,
45
+ 26,
46
+ 27,
47
+ 28,
48
+ 29,
49
+ 30,
50
+ 31,
51
+ 32,
52
+ 33,
53
+ 34,
54
+ 35,
55
+ 36,
56
+ 37,
57
+ 38,
58
+ 39,
59
+ 40,
60
+ 41,
61
+ 42,
62
+ 43,
63
+ 44,
64
+ 45,
65
+ 46,
66
+ 47,
67
+ 48,
68
+ 49,
69
+ 50,
70
+ 51,
71
+ 52,
72
+ 53,
73
+ 54,
74
+ 55,
75
+ 56,
76
+ 57,
77
+ 58,
78
+ 59,
79
+ 60,
80
+ 61,
81
+ 62,
82
+ 63,
83
+ 64,
84
+ 65,
85
+ 66,
86
+ 67,
87
+ 68,
88
+ 69,
89
+ 70,
90
+ 71,
91
+ 72,
92
+ 73,
93
+ 74,
94
+ 75,
95
+ 76,
96
+ 77,
97
+ 78,
98
+ 79,
99
+ 80,
100
+ 81,
101
+ 82,
102
+ 83,
103
+ 84,
104
+ 85,
105
+ 86,
106
+ 87,
107
+ 88,
108
+ 89,
109
+ 90,
110
+ 91,
111
+ 92,
112
+ 93,
113
+ 94,
114
+ 95,
115
+ 96,
116
+ 97,
117
+ 98,
118
+ 99,
119
+ 100,
120
+ 101,
121
+ 102,
122
+ 103,
123
+ 104,
124
+ 105,
125
+ 106,
126
+ 107,
127
+ 108,
128
+ 109,
129
+ 110,
130
+ 111,
131
+ 112,
132
+ 113,
133
+ 114,
134
+ 115,
135
+ 116,
136
+ 117,
137
+ 118,
138
+ 119,
139
+ 120,
140
+ 121,
141
+ 122,
142
+ 123,
143
+ 124,
144
+ 125,
145
+ 126,
146
+ 127,
147
+ 128,
148
+ 129,
149
+ 130,
150
+ 131,
151
+ 132,
152
+ 133,
153
+ 134,
154
+ 135,
155
+ 136,
156
+ 137,
157
+ 138,
158
+ 139,
159
+ 140,
160
+ 141,
161
+ 142,
162
+ 143,
163
+ 144,
164
+ 145,
165
+ 146,
166
+ 147,
167
+ 148,
168
+ 149,
169
+ 150,
170
+ 151,
171
+ 152,
172
+ 153,
173
+ 154,
174
+ 155,
175
+ 156,
176
+ 157,
177
+ 158,
178
+ 159,
179
+ 160,
180
+ 161,
181
+ 162,
182
+ 163,
183
+ 164,
184
+ 165,
185
+ 166,
186
+ 167,
187
+ 168,
188
+ 169,
189
+ 170,
190
+ 171,
191
+ 172,
192
+ 173,
193
+ 174,
194
+ 175,
195
+ 176,
196
+ 177,
197
+ 178,
198
+ 179,
199
+ 180,
200
+ 181,
201
+ 182,
202
+ 183,
203
+ 184,
204
+ 185,
205
+ 186,
206
+ 187,
207
+ 188,
208
+ 189,
209
+ 190,
210
+ 191,
211
+ 192,
212
+ 193,
213
+ 194,
214
+ 195,
215
+ 196,
216
+ 197,
217
+ 198,
218
+ 199,
219
+ 200,
220
+ 201,
221
+ 202,
222
+ 203,
223
+ 204,
224
+ 205,
225
+ 206,
226
+ 207,
227
+ 208,
228
+ 209,
229
+ 210,
230
+ 211,
231
+ 212,
232
+ 213,
233
+ 214,
234
+ 215,
235
+ 216,
236
+ 217,
237
+ 218,
238
+ 219,
239
+ 220,
240
+ 221,
241
+ 222,
242
+ 223,
243
+ 224,
244
+ 225,
245
+ 226,
246
+ 227,
247
+ 228,
248
+ 229,
249
+ 230,
250
+ 231,
251
+ 232,
252
+ 233,
253
+ 234,
254
+ 235,
255
+ 236,
256
+ 237,
257
+ 238,
258
+ 239,
259
+ 240,
260
+ 241,
261
+ 242,
262
+ 243,
263
+ 244,
264
+ 245,
265
+ 246,
266
+ 247,
267
+ 248,
268
+ 249,
269
+ 250,
270
+ 251,
271
+ 252,
272
+ 253,
273
+ 254,
274
+ 255,
275
+ 256,
276
+ 257,
277
+ 258,
278
+ 259,
279
+ 260,
280
+ 261,
281
+ 262,
282
+ 263,
283
+ 264,
284
+ 265,
285
+ 266,
286
+ 267,
287
+ 268,
288
+ 269,
289
+ 270,
290
+ 271,
291
+ 272,
292
+ 273,
293
+ 274,
294
+ 275,
295
+ 276,
296
+ 277,
297
+ 278,
298
+ 279,
299
+ 280,
300
+ 281,
301
+ 282,
302
+ 283,
303
+ 284,
304
+ 285,
305
+ 286,
306
+ 287,
307
+ 288,
308
+ 289,
309
+ 290,
310
+ 291,
311
+ 292,
312
+ 293,
313
+ 294,
314
+ 295,
315
+ 296,
316
+ 297,
317
+ 298,
318
+ 299,
319
+ 300,
320
+ 301,
321
+ 302,
322
+ 303,
323
+ 304,
324
+ 305,
325
+ 306,
326
+ 307,
327
+ 308,
328
+ 309,
329
+ 310,
330
+ 311,
331
+ 312,
332
+ 313,
333
+ 314,
334
+ 315,
335
+ 316,
336
+ 317,
337
+ 318,
338
+ 319,
339
+ 320,
340
+ 321,
341
+ 322,
342
+ 323,
343
+ 324,
344
+ 325,
345
+ 326,
346
+ 327,
347
+ 328,
348
+ 329,
349
+ 330,
350
+ 331,
351
+ 332,
352
+ 333,
353
+ 334,
354
+ 335,
355
+ 336,
356
+ 337,
357
+ 338,
358
+ 339,
359
+ 340,
360
+ 341,
361
+ 342,
362
+ 343,
363
+ 344,
364
+ 345,
365
+ 346,
366
+ 347,
367
+ 348,
368
+ 349,
369
+ 350,
370
+ 351,
371
+ 352,
372
+ 353,
373
+ 354,
374
+ 355,
375
+ 356,
376
+ 357,
377
+ 358,
378
+ 359,
379
+ 360,
380
+ 361,
381
+ 362,
382
+ 363,
383
+ 364,
384
+ 365,
385
+ 366,
386
+ 367,
387
+ 368,
388
+ 369,
389
+ 370,
390
+ 371,
391
+ 372,
392
+ 373,
393
+ 374,
394
+ 375,
395
+ 376,
396
+ 377,
397
+ 378,
398
+ 379,
399
+ 380,
400
+ 381,
401
+ 382,
402
+ 383,
403
+ 384,
404
+ 385,
405
+ 386,
406
+ 387,
407
+ 388,
408
+ 389,
409
+ 390,
410
+ 391,
411
+ 392,
412
+ 393,
413
+ 394,
414
+ 395,
415
+ 396,
416
+ 397,
417
+ 398,
418
+ 399,
419
+ 400,
420
+ 401,
421
+ 402,
422
+ 403,
423
+ 404,
424
+ 405,
425
+ 406,
426
+ 407,
427
+ 408,
428
+ 409,
429
+ 410,
430
+ 411,
431
+ 412,
432
+ 413,
433
+ 414,
434
+ 415,
435
+ 416,
436
+ 417,
437
+ 418,
438
+ 419,
439
+ 420,
440
+ 421,
441
+ 422,
442
+ 423,
443
+ 424,
444
+ 425,
445
+ 426,
446
+ 427,
447
+ 428,
448
+ 429,
449
+ 430,
450
+ 431,
451
+ 432,
452
+ 433,
453
+ 434,
454
+ 435,
455
+ 436,
456
+ 437,
457
+ 438,
458
+ 439,
459
+ 440,
460
+ 441,
461
+ 442,
462
+ 443,
463
+ 444,
464
+ 445,
465
+ 446,
466
+ 447,
467
+ 448,
468
+ 449,
469
+ 450,
470
+ 451,
471
+ 452,
472
+ 453,
473
+ 454,
474
+ 455,
475
+ 456,
476
+ 457,
477
+ 458,
478
+ 459,
479
+ 460,
480
+ 461,
481
+ 462,
482
+ 463,
483
+ 464,
484
+ 465,
485
+ 466,
486
+ 467,
487
+ 468,
488
+ 469,
489
+ 470,
490
+ 471,
491
+ 472,
492
+ 473,
493
+ 474,
494
+ 475,
495
+ 476,
496
+ 477,
497
+ 478,
498
+ 479,
499
+ 480,
500
+ 481,
501
+ 482,
502
+ 483,
503
+ 484,
504
+ 485,
505
+ 486,
506
+ 487,
507
+ 488,
508
+ 489,
509
+ 490,
510
+ 491,
511
+ 492,
512
+ 493,
513
+ 494,
514
+ 495,
515
+ 496,
516
+ 497,
517
+ 498,
518
+ 499
519
+ ]
520
+ }
521
+ ]
checkpoints/060000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cbda48d71dde2c9dde603310ee109f02fec46cd9152ff7471501bb84b1f62a2
3
+ size 412659164
checkpoints/060000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55f7202f8a82b08c9113c0ec6db0c3d93db7aac2777ba072142de9e45df6fbb2
3
+ size 15708
checkpoints/060000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 60000,
6
+ "_step_count": 60001,
7
+ "_get_lr_called_within_step": false,
8
+ "_last_lr": [
9
+ 2.5e-06
10
+ ],
11
+ "lr_lambdas": [
12
+ null
13
+ ]
14
+ }
checkpoints/060000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 60000
3
+ }
checkpoints/080000/pretrained_model/config.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 6
9
+ ]
10
+ },
11
+ "observation.images.front": {
12
+ "type": "VISUAL",
13
+ "shape": [
14
+ 3,
15
+ 480,
16
+ 640
17
+ ]
18
+ }
19
+ },
20
+ "output_features": {
21
+ "action": {
22
+ "type": "ACTION",
23
+ "shape": [
24
+ 6
25
+ ]
26
+ }
27
+ },
28
+ "device": "cuda",
29
+ "use_amp": false,
30
+ "push_to_hub": true,
31
+ "repo_id": "Dimios45/smolvla1",
32
+ "private": null,
33
+ "tags": null,
34
+ "license": null,
35
+ "pretrained_path": "outputs/train/smolvla1/checkpoints/last/pretrained_model",
36
+ "chunk_size": 50,
37
+ "n_action_steps": 50,
38
+ "normalization_mapping": {
39
+ "VISUAL": "IDENTITY",
40
+ "STATE": "MEAN_STD",
41
+ "ACTION": "MEAN_STD"
42
+ },
43
+ "max_state_dim": 32,
44
+ "max_action_dim": 32,
45
+ "resize_imgs_with_padding": [
46
+ 512,
47
+ 512
48
+ ],
49
+ "empty_cameras": 0,
50
+ "adapt_to_pi_aloha": false,
51
+ "use_delta_joint_actions_aloha": false,
52
+ "tokenizer_max_length": 48,
53
+ "num_steps": 10,
54
+ "use_cache": true,
55
+ "freeze_vision_encoder": true,
56
+ "train_expert_only": true,
57
+ "train_state_proj": true,
58
+ "optimizer_lr": 0.0001,
59
+ "optimizer_betas": [
60
+ 0.9,
61
+ 0.95
62
+ ],
63
+ "optimizer_eps": 1e-08,
64
+ "optimizer_weight_decay": 1e-10,
65
+ "optimizer_grad_clip_norm": 10.0,
66
+ "scheduler_warmup_steps": 1000,
67
+ "scheduler_decay_steps": 30000,
68
+ "scheduler_decay_lr": 2.5e-06,
69
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
70
+ "load_vlm_weights": false,
71
+ "add_image_special_tokens": false,
72
+ "attention_mode": "cross_attn",
73
+ "prefix_length": -1,
74
+ "pad_language_to": "longest",
75
+ "num_expert_layers": -1,
76
+ "num_vlm_layers": 16,
77
+ "self_attn_every_n_layers": 2,
78
+ "expert_width_multiplier": 0.75,
79
+ "min_period": 0.004,
80
+ "max_period": 4.0
81
+ }
checkpoints/080000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9182bcf46c5d9ab25c65b350b1f4acc8f2e27f8c709647dc2db63ad6983a9523
3
+ size 1197789224
checkpoints/080000/pretrained_model/policy_postprocessor.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_postprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "unnormalizer_processor",
6
+ "config": {
7
+ "eps": 1e-08,
8
+ "features": {
9
+ "action": {
10
+ "type": "ACTION",
11
+ "shape": [
12
+ 6
13
+ ]
14
+ }
15
+ },
16
+ "norm_map": {
17
+ "VISUAL": "IDENTITY",
18
+ "STATE": "MEAN_STD",
19
+ "ACTION": "MEAN_STD"
20
+ }
21
+ },
22
+ "state_file": "policy_postprocessor_step_0_unnormalizer_processor.safetensors"
23
+ },
24
+ {
25
+ "registry_name": "device_processor",
26
+ "config": {
27
+ "device": "cpu",
28
+ "float_dtype": null
29
+ }
30
+ }
31
+ ]
32
+ }
checkpoints/080000/pretrained_model/policy_postprocessor_step_0_unnormalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/080000/pretrained_model/policy_preprocessor.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "name": "policy_preprocessor",
3
+ "steps": [
4
+ {
5
+ "registry_name": "rename_observations_processor",
6
+ "config": {
7
+ "rename_map": {}
8
+ }
9
+ },
10
+ {
11
+ "registry_name": "to_batch_processor",
12
+ "config": {}
13
+ },
14
+ {
15
+ "registry_name": "smolvla_new_line_processor",
16
+ "config": {}
17
+ },
18
+ {
19
+ "registry_name": "tokenizer_processor",
20
+ "config": {
21
+ "max_length": 48,
22
+ "task_key": "task",
23
+ "padding_side": "right",
24
+ "padding": "longest",
25
+ "truncation": true,
26
+ "tokenizer_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct"
27
+ }
28
+ },
29
+ {
30
+ "registry_name": "device_processor",
31
+ "config": {
32
+ "device": "cuda",
33
+ "float_dtype": null
34
+ }
35
+ },
36
+ {
37
+ "registry_name": "normalizer_processor",
38
+ "config": {
39
+ "eps": 1e-08,
40
+ "features": {
41
+ "observation.state": {
42
+ "type": "STATE",
43
+ "shape": [
44
+ 6
45
+ ]
46
+ },
47
+ "observation.images.front": {
48
+ "type": "VISUAL",
49
+ "shape": [
50
+ 3,
51
+ 480,
52
+ 640
53
+ ]
54
+ },
55
+ "action": {
56
+ "type": "ACTION",
57
+ "shape": [
58
+ 6
59
+ ]
60
+ }
61
+ },
62
+ "norm_map": {
63
+ "VISUAL": "IDENTITY",
64
+ "STATE": "MEAN_STD",
65
+ "ACTION": "MEAN_STD"
66
+ }
67
+ },
68
+ "state_file": "policy_preprocessor_step_5_normalizer_processor.safetensors"
69
+ }
70
+ ]
71
+ }
checkpoints/080000/pretrained_model/policy_preprocessor_step_5_normalizer_processor.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93e38dbbaf8f561886dfc5aa9497d5cd263387d20cab724c84b5be152b85dea2
3
+ size 6560
checkpoints/080000/pretrained_model/train_config.json ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset": {
3
+ "repo_id": "Dimios45/santra",
4
+ "root": "data/santra/santra_demo_20251108_102726",
5
+ "episodes": null,
6
+ "image_transforms": {
7
+ "enable": false,
8
+ "max_num_transforms": 3,
9
+ "random_order": false,
10
+ "tfs": {
11
+ "brightness": {
12
+ "weight": 1.0,
13
+ "type": "ColorJitter",
14
+ "kwargs": {
15
+ "brightness": [
16
+ 0.8,
17
+ 1.2
18
+ ]
19
+ }
20
+ },
21
+ "contrast": {
22
+ "weight": 1.0,
23
+ "type": "ColorJitter",
24
+ "kwargs": {
25
+ "contrast": [
26
+ 0.8,
27
+ 1.2
28
+ ]
29
+ }
30
+ },
31
+ "saturation": {
32
+ "weight": 1.0,
33
+ "type": "ColorJitter",
34
+ "kwargs": {
35
+ "saturation": [
36
+ 0.5,
37
+ 1.5
38
+ ]
39
+ }
40
+ },
41
+ "hue": {
42
+ "weight": 1.0,
43
+ "type": "ColorJitter",
44
+ "kwargs": {
45
+ "hue": [
46
+ -0.05,
47
+ 0.05
48
+ ]
49
+ }
50
+ },
51
+ "sharpness": {
52
+ "weight": 1.0,
53
+ "type": "SharpnessJitter",
54
+ "kwargs": {
55
+ "sharpness": [
56
+ 0.5,
57
+ 1.5
58
+ ]
59
+ }
60
+ },
61
+ "affine": {
62
+ "weight": 1.0,
63
+ "type": "RandomAffine",
64
+ "kwargs": {
65
+ "degrees": [
66
+ -5.0,
67
+ 5.0
68
+ ],
69
+ "translate": [
70
+ 0.05,
71
+ 0.05
72
+ ]
73
+ }
74
+ }
75
+ }
76
+ },
77
+ "revision": null,
78
+ "use_imagenet_stats": true,
79
+ "video_backend": "torchcodec",
80
+ "streaming": false
81
+ },
82
+ "env": null,
83
+ "policy": {
84
+ "type": "smolvla",
85
+ "n_obs_steps": 1,
86
+ "input_features": {
87
+ "observation.state": {
88
+ "type": "STATE",
89
+ "shape": [
90
+ 6
91
+ ]
92
+ },
93
+ "observation.images.front": {
94
+ "type": "VISUAL",
95
+ "shape": [
96
+ 3,
97
+ 480,
98
+ 640
99
+ ]
100
+ }
101
+ },
102
+ "output_features": {
103
+ "action": {
104
+ "type": "ACTION",
105
+ "shape": [
106
+ 6
107
+ ]
108
+ }
109
+ },
110
+ "device": "cuda",
111
+ "use_amp": false,
112
+ "push_to_hub": true,
113
+ "repo_id": "Dimios45/smolvla1",
114
+ "private": null,
115
+ "tags": null,
116
+ "license": null,
117
+ "pretrained_path": "outputs/train/smolvla1/checkpoints/last/pretrained_model",
118
+ "chunk_size": 50,
119
+ "n_action_steps": 50,
120
+ "normalization_mapping": {
121
+ "VISUAL": "IDENTITY",
122
+ "STATE": "MEAN_STD",
123
+ "ACTION": "MEAN_STD"
124
+ },
125
+ "max_state_dim": 32,
126
+ "max_action_dim": 32,
127
+ "resize_imgs_with_padding": [
128
+ 512,
129
+ 512
130
+ ],
131
+ "empty_cameras": 0,
132
+ "adapt_to_pi_aloha": false,
133
+ "use_delta_joint_actions_aloha": false,
134
+ "tokenizer_max_length": 48,
135
+ "num_steps": 10,
136
+ "use_cache": true,
137
+ "freeze_vision_encoder": true,
138
+ "train_expert_only": true,
139
+ "train_state_proj": true,
140
+ "optimizer_lr": 0.0001,
141
+ "optimizer_betas": [
142
+ 0.9,
143
+ 0.95
144
+ ],
145
+ "optimizer_eps": 1e-08,
146
+ "optimizer_weight_decay": 1e-10,
147
+ "optimizer_grad_clip_norm": 10.0,
148
+ "scheduler_warmup_steps": 1000,
149
+ "scheduler_decay_steps": 30000,
150
+ "scheduler_decay_lr": 2.5e-06,
151
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
152
+ "load_vlm_weights": false,
153
+ "add_image_special_tokens": false,
154
+ "attention_mode": "cross_attn",
155
+ "prefix_length": -1,
156
+ "pad_language_to": "longest",
157
+ "num_expert_layers": -1,
158
+ "num_vlm_layers": 16,
159
+ "self_attn_every_n_layers": 2,
160
+ "expert_width_multiplier": 0.75,
161
+ "min_period": 0.004,
162
+ "max_period": 4.0
163
+ },
164
+ "output_dir": "outputs/train/smolvla1",
165
+ "job_name": "smolvla_santra",
166
+ "resume": true,
167
+ "seed": 1000,
168
+ "num_workers": 4,
169
+ "batch_size": 8,
170
+ "steps": 100000,
171
+ "eval_freq": 20000,
172
+ "log_freq": 200,
173
+ "save_checkpoint": true,
174
+ "save_freq": 20000,
175
+ "use_policy_training_preset": true,
176
+ "optimizer": {
177
+ "type": "adamw",
178
+ "lr": 0.0001,
179
+ "weight_decay": 1e-10,
180
+ "grad_clip_norm": 10.0,
181
+ "betas": [
182
+ 0.9,
183
+ 0.95
184
+ ],
185
+ "eps": 1e-08
186
+ },
187
+ "scheduler": {
188
+ "type": "cosine_decay_with_warmup",
189
+ "num_warmup_steps": 1000,
190
+ "num_decay_steps": 30000,
191
+ "peak_lr": 0.0001,
192
+ "decay_lr": 2.5e-06
193
+ },
194
+ "eval": {
195
+ "n_episodes": 50,
196
+ "batch_size": 50,
197
+ "use_async_envs": false
198
+ },
199
+ "wandb": {
200
+ "enable": true,
201
+ "disable_artifact": false,
202
+ "project": "lerobot",
203
+ "entity": null,
204
+ "notes": null,
205
+ "run_id": "ba762whj",
206
+ "mode": null
207
+ },
208
+ "checkpoint_path": "outputs/train/smolvla1/checkpoints/last",
209
+ "rename_map": {}
210
+ }
checkpoints/080000/training_state/optimizer_param_groups.json ADDED
@@ -0,0 +1,521 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "lr": 2.5e-06,
4
+ "betas": [
5
+ 0.9,
6
+ 0.95
7
+ ],
8
+ "eps": 1e-08,
9
+ "weight_decay": 1e-10,
10
+ "amsgrad": false,
11
+ "maximize": false,
12
+ "foreach": null,
13
+ "capturable": false,
14
+ "differentiable": false,
15
+ "fused": null,
16
+ "decoupled_weight_decay": true,
17
+ "initial_lr": 0.0001,
18
+ "params": [
19
+ 0,
20
+ 1,
21
+ 2,
22
+ 3,
23
+ 4,
24
+ 5,
25
+ 6,
26
+ 7,
27
+ 8,
28
+ 9,
29
+ 10,
30
+ 11,
31
+ 12,
32
+ 13,
33
+ 14,
34
+ 15,
35
+ 16,
36
+ 17,
37
+ 18,
38
+ 19,
39
+ 20,
40
+ 21,
41
+ 22,
42
+ 23,
43
+ 24,
44
+ 25,
45
+ 26,
46
+ 27,
47
+ 28,
48
+ 29,
49
+ 30,
50
+ 31,
51
+ 32,
52
+ 33,
53
+ 34,
54
+ 35,
55
+ 36,
56
+ 37,
57
+ 38,
58
+ 39,
59
+ 40,
60
+ 41,
61
+ 42,
62
+ 43,
63
+ 44,
64
+ 45,
65
+ 46,
66
+ 47,
67
+ 48,
68
+ 49,
69
+ 50,
70
+ 51,
71
+ 52,
72
+ 53,
73
+ 54,
74
+ 55,
75
+ 56,
76
+ 57,
77
+ 58,
78
+ 59,
79
+ 60,
80
+ 61,
81
+ 62,
82
+ 63,
83
+ 64,
84
+ 65,
85
+ 66,
86
+ 67,
87
+ 68,
88
+ 69,
89
+ 70,
90
+ 71,
91
+ 72,
92
+ 73,
93
+ 74,
94
+ 75,
95
+ 76,
96
+ 77,
97
+ 78,
98
+ 79,
99
+ 80,
100
+ 81,
101
+ 82,
102
+ 83,
103
+ 84,
104
+ 85,
105
+ 86,
106
+ 87,
107
+ 88,
108
+ 89,
109
+ 90,
110
+ 91,
111
+ 92,
112
+ 93,
113
+ 94,
114
+ 95,
115
+ 96,
116
+ 97,
117
+ 98,
118
+ 99,
119
+ 100,
120
+ 101,
121
+ 102,
122
+ 103,
123
+ 104,
124
+ 105,
125
+ 106,
126
+ 107,
127
+ 108,
128
+ 109,
129
+ 110,
130
+ 111,
131
+ 112,
132
+ 113,
133
+ 114,
134
+ 115,
135
+ 116,
136
+ 117,
137
+ 118,
138
+ 119,
139
+ 120,
140
+ 121,
141
+ 122,
142
+ 123,
143
+ 124,
144
+ 125,
145
+ 126,
146
+ 127,
147
+ 128,
148
+ 129,
149
+ 130,
150
+ 131,
151
+ 132,
152
+ 133,
153
+ 134,
154
+ 135,
155
+ 136,
156
+ 137,
157
+ 138,
158
+ 139,
159
+ 140,
160
+ 141,
161
+ 142,
162
+ 143,
163
+ 144,
164
+ 145,
165
+ 146,
166
+ 147,
167
+ 148,
168
+ 149,
169
+ 150,
170
+ 151,
171
+ 152,
172
+ 153,
173
+ 154,
174
+ 155,
175
+ 156,
176
+ 157,
177
+ 158,
178
+ 159,
179
+ 160,
180
+ 161,
181
+ 162,
182
+ 163,
183
+ 164,
184
+ 165,
185
+ 166,
186
+ 167,
187
+ 168,
188
+ 169,
189
+ 170,
190
+ 171,
191
+ 172,
192
+ 173,
193
+ 174,
194
+ 175,
195
+ 176,
196
+ 177,
197
+ 178,
198
+ 179,
199
+ 180,
200
+ 181,
201
+ 182,
202
+ 183,
203
+ 184,
204
+ 185,
205
+ 186,
206
+ 187,
207
+ 188,
208
+ 189,
209
+ 190,
210
+ 191,
211
+ 192,
212
+ 193,
213
+ 194,
214
+ 195,
215
+ 196,
216
+ 197,
217
+ 198,
218
+ 199,
219
+ 200,
220
+ 201,
221
+ 202,
222
+ 203,
223
+ 204,
224
+ 205,
225
+ 206,
226
+ 207,
227
+ 208,
228
+ 209,
229
+ 210,
230
+ 211,
231
+ 212,
232
+ 213,
233
+ 214,
234
+ 215,
235
+ 216,
236
+ 217,
237
+ 218,
238
+ 219,
239
+ 220,
240
+ 221,
241
+ 222,
242
+ 223,
243
+ 224,
244
+ 225,
245
+ 226,
246
+ 227,
247
+ 228,
248
+ 229,
249
+ 230,
250
+ 231,
251
+ 232,
252
+ 233,
253
+ 234,
254
+ 235,
255
+ 236,
256
+ 237,
257
+ 238,
258
+ 239,
259
+ 240,
260
+ 241,
261
+ 242,
262
+ 243,
263
+ 244,
264
+ 245,
265
+ 246,
266
+ 247,
267
+ 248,
268
+ 249,
269
+ 250,
270
+ 251,
271
+ 252,
272
+ 253,
273
+ 254,
274
+ 255,
275
+ 256,
276
+ 257,
277
+ 258,
278
+ 259,
279
+ 260,
280
+ 261,
281
+ 262,
282
+ 263,
283
+ 264,
284
+ 265,
285
+ 266,
286
+ 267,
287
+ 268,
288
+ 269,
289
+ 270,
290
+ 271,
291
+ 272,
292
+ 273,
293
+ 274,
294
+ 275,
295
+ 276,
296
+ 277,
297
+ 278,
298
+ 279,
299
+ 280,
300
+ 281,
301
+ 282,
302
+ 283,
303
+ 284,
304
+ 285,
305
+ 286,
306
+ 287,
307
+ 288,
308
+ 289,
309
+ 290,
310
+ 291,
311
+ 292,
312
+ 293,
313
+ 294,
314
+ 295,
315
+ 296,
316
+ 297,
317
+ 298,
318
+ 299,
319
+ 300,
320
+ 301,
321
+ 302,
322
+ 303,
323
+ 304,
324
+ 305,
325
+ 306,
326
+ 307,
327
+ 308,
328
+ 309,
329
+ 310,
330
+ 311,
331
+ 312,
332
+ 313,
333
+ 314,
334
+ 315,
335
+ 316,
336
+ 317,
337
+ 318,
338
+ 319,
339
+ 320,
340
+ 321,
341
+ 322,
342
+ 323,
343
+ 324,
344
+ 325,
345
+ 326,
346
+ 327,
347
+ 328,
348
+ 329,
349
+ 330,
350
+ 331,
351
+ 332,
352
+ 333,
353
+ 334,
354
+ 335,
355
+ 336,
356
+ 337,
357
+ 338,
358
+ 339,
359
+ 340,
360
+ 341,
361
+ 342,
362
+ 343,
363
+ 344,
364
+ 345,
365
+ 346,
366
+ 347,
367
+ 348,
368
+ 349,
369
+ 350,
370
+ 351,
371
+ 352,
372
+ 353,
373
+ 354,
374
+ 355,
375
+ 356,
376
+ 357,
377
+ 358,
378
+ 359,
379
+ 360,
380
+ 361,
381
+ 362,
382
+ 363,
383
+ 364,
384
+ 365,
385
+ 366,
386
+ 367,
387
+ 368,
388
+ 369,
389
+ 370,
390
+ 371,
391
+ 372,
392
+ 373,
393
+ 374,
394
+ 375,
395
+ 376,
396
+ 377,
397
+ 378,
398
+ 379,
399
+ 380,
400
+ 381,
401
+ 382,
402
+ 383,
403
+ 384,
404
+ 385,
405
+ 386,
406
+ 387,
407
+ 388,
408
+ 389,
409
+ 390,
410
+ 391,
411
+ 392,
412
+ 393,
413
+ 394,
414
+ 395,
415
+ 396,
416
+ 397,
417
+ 398,
418
+ 399,
419
+ 400,
420
+ 401,
421
+ 402,
422
+ 403,
423
+ 404,
424
+ 405,
425
+ 406,
426
+ 407,
427
+ 408,
428
+ 409,
429
+ 410,
430
+ 411,
431
+ 412,
432
+ 413,
433
+ 414,
434
+ 415,
435
+ 416,
436
+ 417,
437
+ 418,
438
+ 419,
439
+ 420,
440
+ 421,
441
+ 422,
442
+ 423,
443
+ 424,
444
+ 425,
445
+ 426,
446
+ 427,
447
+ 428,
448
+ 429,
449
+ 430,
450
+ 431,
451
+ 432,
452
+ 433,
453
+ 434,
454
+ 435,
455
+ 436,
456
+ 437,
457
+ 438,
458
+ 439,
459
+ 440,
460
+ 441,
461
+ 442,
462
+ 443,
463
+ 444,
464
+ 445,
465
+ 446,
466
+ 447,
467
+ 448,
468
+ 449,
469
+ 450,
470
+ 451,
471
+ 452,
472
+ 453,
473
+ 454,
474
+ 455,
475
+ 456,
476
+ 457,
477
+ 458,
478
+ 459,
479
+ 460,
480
+ 461,
481
+ 462,
482
+ 463,
483
+ 464,
484
+ 465,
485
+ 466,
486
+ 467,
487
+ 468,
488
+ 469,
489
+ 470,
490
+ 471,
491
+ 472,
492
+ 473,
493
+ 474,
494
+ 475,
495
+ 476,
496
+ 477,
497
+ 478,
498
+ 479,
499
+ 480,
500
+ 481,
501
+ 482,
502
+ 483,
503
+ 484,
504
+ 485,
505
+ 486,
506
+ 487,
507
+ 488,
508
+ 489,
509
+ 490,
510
+ 491,
511
+ 492,
512
+ 493,
513
+ 494,
514
+ 495,
515
+ 496,
516
+ 497,
517
+ 498,
518
+ 499
519
+ ]
520
+ }
521
+ ]
checkpoints/080000/training_state/optimizer_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:423c1c02ccd7753f3232685c6daf8acf606a67cece3ce330573968aa5f5b4b41
3
+ size 412659164
checkpoints/080000/training_state/rng_state.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba5f2a94aaf02f3cc692290e428bc96b87921107c1520ae18f6ce1e42b5d399e
3
+ size 15708
checkpoints/080000/training_state/scheduler_state.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_lrs": [
3
+ 0.0001
4
+ ],
5
+ "last_epoch": 80000,
6
+ "_step_count": 80001,
7
+ "_get_lr_called_within_step": false,
8
+ "_last_lr": [
9
+ 2.5e-06
10
+ ],
11
+ "lr_lambdas": [
12
+ null
13
+ ]
14
+ }
checkpoints/080000/training_state/training_step.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "step": 80000
3
+ }
checkpoints/100000/pretrained_model/config.json ADDED
@@ -0,0 +1,81 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "type": "smolvla",
3
+ "n_obs_steps": 1,
4
+ "input_features": {
5
+ "observation.state": {
6
+ "type": "STATE",
7
+ "shape": [
8
+ 6
9
+ ]
10
+ },
11
+ "observation.images.front": {
12
+ "type": "VISUAL",
13
+ "shape": [
14
+ 3,
15
+ 480,
16
+ 640
17
+ ]
18
+ }
19
+ },
20
+ "output_features": {
21
+ "action": {
22
+ "type": "ACTION",
23
+ "shape": [
24
+ 6
25
+ ]
26
+ }
27
+ },
28
+ "device": "cuda",
29
+ "use_amp": false,
30
+ "push_to_hub": true,
31
+ "repo_id": "Dimios45/smolvla1",
32
+ "private": null,
33
+ "tags": null,
34
+ "license": null,
35
+ "pretrained_path": "outputs/train/smolvla1/checkpoints/last/pretrained_model",
36
+ "chunk_size": 50,
37
+ "n_action_steps": 50,
38
+ "normalization_mapping": {
39
+ "VISUAL": "IDENTITY",
40
+ "STATE": "MEAN_STD",
41
+ "ACTION": "MEAN_STD"
42
+ },
43
+ "max_state_dim": 32,
44
+ "max_action_dim": 32,
45
+ "resize_imgs_with_padding": [
46
+ 512,
47
+ 512
48
+ ],
49
+ "empty_cameras": 0,
50
+ "adapt_to_pi_aloha": false,
51
+ "use_delta_joint_actions_aloha": false,
52
+ "tokenizer_max_length": 48,
53
+ "num_steps": 10,
54
+ "use_cache": true,
55
+ "freeze_vision_encoder": true,
56
+ "train_expert_only": true,
57
+ "train_state_proj": true,
58
+ "optimizer_lr": 0.0001,
59
+ "optimizer_betas": [
60
+ 0.9,
61
+ 0.95
62
+ ],
63
+ "optimizer_eps": 1e-08,
64
+ "optimizer_weight_decay": 1e-10,
65
+ "optimizer_grad_clip_norm": 10.0,
66
+ "scheduler_warmup_steps": 1000,
67
+ "scheduler_decay_steps": 30000,
68
+ "scheduler_decay_lr": 2.5e-06,
69
+ "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
70
+ "load_vlm_weights": false,
71
+ "add_image_special_tokens": false,
72
+ "attention_mode": "cross_attn",
73
+ "prefix_length": -1,
74
+ "pad_language_to": "longest",
75
+ "num_expert_layers": -1,
76
+ "num_vlm_layers": 16,
77
+ "self_attn_every_n_layers": 2,
78
+ "expert_width_multiplier": 0.75,
79
+ "min_period": 0.004,
80
+ "max_period": 4.0
81
+ }
checkpoints/100000/pretrained_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32579685d79e6272fb6e65e4a5f28a70cc4a74fd215448b9972d8d19c7d9373e
3
+ size 1197789224