Robotics
LeRobot
Safetensors
smolvla
qualiaadmin commited on
Commit
dd70844
·
verified ·
1 Parent(s): 662f629

Upload policy weights, train config and readme

Browse files
Files changed (4) hide show
  1. README.md +1 -1
  2. config.json +7 -7
  3. model.safetensors +2 -2
  4. train_config.json +12 -12
README.md CHANGED
@@ -6,8 +6,8 @@ license: apache-2.0
6
  model_name: smolvla
7
  pipeline_tag: robotics
8
  tags:
9
- - smolvla
10
  - lerobot
 
11
  - robotics
12
  ---
13
 
 
6
  model_name: smolvla
7
  pipeline_tag: robotics
8
  tags:
 
9
  - lerobot
10
+ - smolvla
11
  - robotics
12
  ---
13
 
config.json CHANGED
@@ -40,7 +40,7 @@
40
  "private": null,
41
  "tags": null,
42
  "license": null,
43
- "pretrained_path": "./lerobot_smolvla_base_migrated",
44
  "chunk_size": 50,
45
  "n_action_steps": 50,
46
  "normalization_mapping": {
@@ -70,17 +70,17 @@
70
  ],
71
  "optimizer_eps": 1e-08,
72
  "optimizer_weight_decay": 1e-10,
73
- "optimizer_grad_clip_norm": 10.0,
74
  "scheduler_warmup_steps": 1000,
75
  "scheduler_decay_steps": 30000,
76
- "scheduler_decay_lr": 2.5e-06,
77
  "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
78
- "load_vlm_weights": true,
79
  "add_image_special_tokens": false,
80
  "attention_mode": "cross_attn",
81
- "prefix_length": 0,
82
- "pad_language_to": "max_length",
83
- "num_expert_layers": 0,
84
  "num_vlm_layers": 16,
85
  "self_attn_every_n_layers": 2,
86
  "expert_width_multiplier": 0.75,
 
40
  "private": null,
41
  "tags": null,
42
  "license": null,
43
+ "pretrained_path": null,
44
  "chunk_size": 50,
45
  "n_action_steps": 50,
46
  "normalization_mapping": {
 
70
  ],
71
  "optimizer_eps": 1e-08,
72
  "optimizer_weight_decay": 1e-10,
73
+ "optimizer_grad_clip_norm": 10,
74
  "scheduler_warmup_steps": 1000,
75
  "scheduler_decay_steps": 30000,
76
+ "scheduler_decay_lr": 0.025,
77
  "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
78
+ "load_vlm_weights": false,
79
  "add_image_special_tokens": false,
80
  "attention_mode": "cross_attn",
81
+ "prefix_length": -1,
82
+ "pad_language_to": "longest",
83
+ "num_expert_layers": -1,
84
  "num_vlm_layers": 16,
85
  "self_attn_every_n_layers": 2,
86
  "expert_width_multiplier": 0.75,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea3eab3e960358290831fe5c23193dd5a94d6ef907b50d443c83b1d58f5864bf
3
- size 906712520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a99f156bd0a564124b32f77323bdb8b9bbcd1691f8e5db9481f70aef768238fb
3
+ size 1197789224
train_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "dataset": {
3
  "repo_id": "Calvert0921/SmolVLA_LiftBlackCube5_Franka_100",
4
- "root": null,
5
  "episodes": null,
6
  "image_transforms": {
7
  "enable": false,
@@ -108,7 +108,7 @@
108
  "private": null,
109
  "tags": null,
110
  "license": null,
111
- "pretrained_path": "./lerobot_smolvla_base_migrated",
112
  "chunk_size": 50,
113
  "n_action_steps": 50,
114
  "normalization_mapping": {
@@ -138,17 +138,17 @@
138
  ],
139
  "optimizer_eps": 1e-08,
140
  "optimizer_weight_decay": 1e-10,
141
- "optimizer_grad_clip_norm": 10.0,
142
  "scheduler_warmup_steps": 1000,
143
  "scheduler_decay_steps": 30000,
144
- "scheduler_decay_lr": 2.5e-06,
145
  "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
146
- "load_vlm_weights": true,
147
  "add_image_special_tokens": false,
148
  "attention_mode": "cross_attn",
149
- "prefix_length": 0,
150
- "pad_language_to": "max_length",
151
- "num_expert_layers": 0,
152
  "num_vlm_layers": 16,
153
  "self_attn_every_n_layers": 2,
154
  "expert_width_multiplier": 0.75,
@@ -161,9 +161,9 @@
161
  "seed": 1000,
162
  "num_workers": 4,
163
  "batch_size": 1,
164
- "steps": 1000,
165
  "eval_freq": 20000,
166
- "log_freq": 200,
167
  "save_checkpoint": true,
168
  "save_freq": 20000,
169
  "use_policy_training_preset": true,
@@ -171,7 +171,7 @@
171
  "type": "adamw",
172
  "lr": 0.0001,
173
  "weight_decay": 1e-10,
174
- "grad_clip_norm": 10.0,
175
  "betas": [
176
  0.9,
177
  0.95
@@ -183,7 +183,7 @@
183
  "num_warmup_steps": 1000,
184
  "num_decay_steps": 30000,
185
  "peak_lr": 0.0001,
186
- "decay_lr": 2.5e-06
187
  },
188
  "eval": {
189
  "n_episodes": 50,
 
1
  {
2
  "dataset": {
3
  "repo_id": "Calvert0921/SmolVLA_LiftBlackCube5_Franka_100",
4
+ "root": "./dataset/Calvert0921/SmolVLA_LiftBlackCube5_Franka_100",
5
  "episodes": null,
6
  "image_transforms": {
7
  "enable": false,
 
108
  "private": null,
109
  "tags": null,
110
  "license": null,
111
+ "pretrained_path": null,
112
  "chunk_size": 50,
113
  "n_action_steps": 50,
114
  "normalization_mapping": {
 
138
  ],
139
  "optimizer_eps": 1e-08,
140
  "optimizer_weight_decay": 1e-10,
141
+ "optimizer_grad_clip_norm": 10,
142
  "scheduler_warmup_steps": 1000,
143
  "scheduler_decay_steps": 30000,
144
+ "scheduler_decay_lr": 0.025,
145
  "vlm_model_name": "HuggingFaceTB/SmolVLM2-500M-Video-Instruct",
146
+ "load_vlm_weights": false,
147
  "add_image_special_tokens": false,
148
  "attention_mode": "cross_attn",
149
+ "prefix_length": -1,
150
+ "pad_language_to": "longest",
151
+ "num_expert_layers": -1,
152
  "num_vlm_layers": 16,
153
  "self_attn_every_n_layers": 2,
154
  "expert_width_multiplier": 0.75,
 
161
  "seed": 1000,
162
  "num_workers": 4,
163
  "batch_size": 1,
164
+ "steps": 10000000,
165
  "eval_freq": 20000,
166
+ "log_freq": 10,
167
  "save_checkpoint": true,
168
  "save_freq": 20000,
169
  "use_policy_training_preset": true,
 
171
  "type": "adamw",
172
  "lr": 0.0001,
173
  "weight_decay": 1e-10,
174
+ "grad_clip_norm": 10,
175
  "betas": [
176
  0.9,
177
  0.95
 
183
  "num_warmup_steps": 1000,
184
  "num_decay_steps": 30000,
185
  "peak_lr": 0.0001,
186
+ "decay_lr": 0.025
187
  },
188
  "eval": {
189
  "n_episodes": 50,