Gavin1411 commited on
Commit
acde485
·
verified ·
1 Parent(s): 078c572

Upload policy weights, train config and readme

Browse files
README.md CHANGED
@@ -6,8 +6,8 @@ model_name: pi05
6
  pipeline_tag: robotics
7
  tags:
8
  - pi05
9
- - robotics
10
  - lerobot
 
11
  ---
12
 
13
  # Model Card for pi05
 
6
  pipeline_tag: robotics
7
  tags:
8
  - pi05
 
9
  - lerobot
10
+ - robotics
11
  ---
12
 
13
  # Model Card for pi05
adapter_config.json CHANGED
@@ -24,14 +24,16 @@
24
  "lora_dropout": 0.0,
25
  "megatron_config": null,
26
  "megatron_core": "megatron.core",
27
- "modules_to_save": [],
 
 
28
  "peft_type": "LORA",
29
  "peft_version": "0.18.1",
30
  "qalora_group_size": 16,
31
  "r": 16,
32
  "rank_pattern": {},
33
  "revision": null,
34
- "target_modules": "(.*\\.gemma_expert\\..*\\.self_attn\\.(q|v)_proj|model\\.(state_proj|action_in_proj|action_out_proj|action_time_mlp_in|action_time_mlp_out))",
35
  "target_parameters": null,
36
  "task_type": null,
37
  "trainable_token_indices": null,
 
24
  "lora_dropout": 0.0,
25
  "megatron_config": null,
26
  "megatron_core": "megatron.core",
27
+ "modules_to_save": [
28
+ ".*\\.vision_tower\\..*"
29
+ ],
30
  "peft_type": "LORA",
31
  "peft_version": "0.18.1",
32
  "qalora_group_size": 16,
33
  "r": 16,
34
  "rank_pattern": {},
35
  "revision": null,
36
+ "target_modules": ".*\\.paligemma\\.model\\.language_model\\..*\\.self_attn\\.(q|v)_proj|.*\\.gemma_expert\\..*\\.self_attn\\.(q|v)_proj|model\\.(state_proj|action_in_proj|action_out_proj|action_time_mlp_in|action_time_mlp_out)",
37
  "target_parameters": null,
38
  "task_type": null,
39
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:12eb3b98c04791dfde7b6b4df727958df15b1be762658b52181eeaeab6de76e0
3
- size 5161624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5a24ffc6cdad9fc08b833675654a64d1fcdb3522b5a111ce9dfc5e47847899c
3
+ size 12547808
train_config.json CHANGED
@@ -210,7 +210,7 @@
210
  "scheduler_decay_steps": 30000,
211
  "scheduler_decay_lr": 2.5e-06
212
  },
213
- "output_dir": "/data/taojiachen/Continuous_VLA/exp/sequential/spatial/seed42/task30/train",
214
  "job_name": "pi05",
215
  "resume": false,
216
  "seed": 42,
@@ -257,8 +257,10 @@
257
  "mode": null
258
  },
259
  "peft": {
260
- "target_modules": null,
261
- "full_training_modules": null,
 
 
262
  "method_type": "LORA",
263
  "init_type": null,
264
  "r": 16
 
210
  "scheduler_decay_steps": 30000,
211
  "scheduler_decay_lr": 2.5e-06
212
  },
213
+ "output_dir": "/data/taojiachen/Continuous_VLA/exp/er/spatial/seed42/task30/train",
214
  "job_name": "pi05",
215
  "resume": false,
216
  "seed": 42,
 
257
  "mode": null
258
  },
259
  "peft": {
260
+ "target_modules": ".*\\.paligemma\\.model\\.language_model\\..*\\.self_attn\\.(q|v)_proj|.*\\.gemma_expert\\..*\\.self_attn\\.(q|v)_proj|model\\.(state_proj|action_in_proj|action_out_proj|action_time_mlp_in|action_time_mlp_out)",
261
+ "full_training_modules": [
262
+ ".*\\.vision_tower\\..*"
263
+ ],
264
  "method_type": "LORA",
265
  "init_type": null,
266
  "r": 16