Upload folder using huggingface_hub
Browse files- checkpoints/finetune_lora_task33_2000step/checkpoints/steps_1000_pytorch_model.pt +3 -0
- checkpoints/finetune_lora_task33_2000step/checkpoints/steps_1500_pytorch_model.pt +3 -0
- checkpoints/finetune_lora_task33_2000step/checkpoints/steps_2000_pytorch_model.pt +3 -0
- checkpoints/finetune_lora_task33_2000step/checkpoints/steps_500_pytorch_model.pt +3 -0
- checkpoints/finetune_lora_task33_2000step/config.yaml +58 -0
- checkpoints/finetune_lora_task33_2000step/dataset_statistics.json +133 -0
- checkpoints/finetune_lora_task33_2000step/final_model/pytorch_model.pt +3 -0
- checkpoints/finetune_lora_task33_2000step/summary.jsonl +4 -0
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_1000_pytorch_model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6e3181b89b99e764882fe6fcfd5a41b586d61300ed5eeba697baebd87ba82b0
|
| 3 |
+
size 29605718
|
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_1500_pytorch_model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5ed0f5a284fd054c5af1b35bbdae34930febba5acd2ea039900bf632ec99e1ca
|
| 3 |
+
size 29605718
|
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_2000_pytorch_model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e4d0ca206384408fc8b0c5d69a20baf5549ea479935b5d43a5dae86c91e5873b
|
| 3 |
+
size 29605718
|
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_500_pytorch_model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c6ce393d900b98e31cc5446934c5a60c6937c1a5a7dcefdf518263cd15623648
|
| 3 |
+
size 29605426
|
checkpoints/finetune_lora_task33_2000step/config.yaml
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
datasets:
|
| 2 |
+
vla_data:
|
| 3 |
+
CoT_prompt: Your task is {instruction}. To identify the key objects for your task.
|
| 4 |
+
Locate their bounding boxes in [x1,y1,x2,y2] format.
|
| 5 |
+
data_mix: libero_90_task_33
|
| 6 |
+
data_root_dir: playground/Datasets/LEROBOT_LIBERO_DATA
|
| 7 |
+
dataset_py: lerobot_datasets
|
| 8 |
+
per_device_batch_size: 1
|
| 9 |
+
sequential_step_sampling: false
|
| 10 |
+
video_backend: torchvision_av
|
| 11 |
+
framework:
|
| 12 |
+
action_model:
|
| 13 |
+
action_dim: 7
|
| 14 |
+
future_action_window_size: 7
|
| 15 |
+
past_action_window_size: 0
|
| 16 |
+
name: QwenFast
|
| 17 |
+
qwenvl:
|
| 18 |
+
base_vlm: playground/Pretrained_models/Qwen2.5-VL-3B-Instruct-Action
|
| 19 |
+
lora_config:
|
| 20 |
+
lora_alpha: 32
|
| 21 |
+
lora_dropout: 0.05
|
| 22 |
+
r: 16
|
| 23 |
+
target_modules:
|
| 24 |
+
- q_proj
|
| 25 |
+
- v_proj
|
| 26 |
+
- k_proj
|
| 27 |
+
- o_proj
|
| 28 |
+
output_dir: ./results/Checkpoints/finetune_lora_task33_2000step
|
| 29 |
+
run_id: finetune_lora_task33_2000step
|
| 30 |
+
run_root_dir: ./results/Checkpoints
|
| 31 |
+
seed: 42
|
| 32 |
+
trainer:
|
| 33 |
+
eval_interval: 100
|
| 34 |
+
freeze_modules: qwen_vl_interface.model.model.visual,dino_encoder
|
| 35 |
+
gradient_accumulation_steps: 1
|
| 36 |
+
gradient_clipping: 1.0
|
| 37 |
+
is_resume: true
|
| 38 |
+
learning_rate:
|
| 39 |
+
action_model: 0.0001
|
| 40 |
+
base: 2.5e-05
|
| 41 |
+
qwen_vl_interface: 1.0e-05
|
| 42 |
+
logging_frequency: 100
|
| 43 |
+
lr_scheduler_type: cosine_with_min_lr
|
| 44 |
+
max_train_steps: 2000
|
| 45 |
+
num_warmup_steps: 5000
|
| 46 |
+
optimizer:
|
| 47 |
+
betas:
|
| 48 |
+
- 0.9
|
| 49 |
+
- 0.95
|
| 50 |
+
eps: 1.0e-08
|
| 51 |
+
weight_decay: 1.0e-08
|
| 52 |
+
pretrained_checkpoint: /content/starVLA_r/results/Checkpoints/Qwen2.5-VL-FAST-LIBERO-4in1/checkpoints/steps_30000_pytorch_model.pt
|
| 53 |
+
save_interval: 500
|
| 54 |
+
scheduler_specific_kwargs:
|
| 55 |
+
min_lr: 1.0e-06
|
| 56 |
+
use_lora: true
|
| 57 |
+
wandb_entity: michellelin9102-usc
|
| 58 |
+
wandb_project: starVLA_Libero
|
checkpoints/finetune_lora_task33_2000step/dataset_statistics.json
ADDED
|
@@ -0,0 +1,133 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"franka": {
|
| 3 |
+
"action": {
|
| 4 |
+
"mean": [
|
| 5 |
+
0.03962625563144684,
|
| 6 |
+
0.13655489683151245,
|
| 7 |
+
-0.0458662211894989,
|
| 8 |
+
-0.0012228331761434674,
|
| 9 |
+
0.0017042774707078934,
|
| 10 |
+
-0.005507396999746561,
|
| 11 |
+
0.5082296133041382
|
| 12 |
+
],
|
| 13 |
+
"std": [
|
| 14 |
+
0.22318312525749207,
|
| 15 |
+
0.43824875354766846,
|
| 16 |
+
0.41923093795776367,
|
| 17 |
+
0.034977223724126816,
|
| 18 |
+
0.04809708148241043,
|
| 19 |
+
0.024638118222355843,
|
| 20 |
+
0.49992871284484863
|
| 21 |
+
],
|
| 22 |
+
"max": [
|
| 23 |
+
0.7232142686843872,
|
| 24 |
+
0.9160714149475098,
|
| 25 |
+
0.9375,
|
| 26 |
+
0.16285714507102966,
|
| 27 |
+
0.23571428656578064,
|
| 28 |
+
0.12107142806053162,
|
| 29 |
+
1.0
|
| 30 |
+
],
|
| 31 |
+
"min": [
|
| 32 |
+
-0.5758928656578064,
|
| 33 |
+
-0.8357142806053162,
|
| 34 |
+
-0.9375,
|
| 35 |
+
-0.1842857152223587,
|
| 36 |
+
-0.23678570985794067,
|
| 37 |
+
-0.11999999731779099,
|
| 38 |
+
0.0
|
| 39 |
+
],
|
| 40 |
+
"q01": [
|
| 41 |
+
-0.4258928596973419,
|
| 42 |
+
-0.7071428298950195,
|
| 43 |
+
-0.8016428303718567,
|
| 44 |
+
-0.11892857402563095,
|
| 45 |
+
-0.11785714328289032,
|
| 46 |
+
-0.08250000327825546,
|
| 47 |
+
0.0
|
| 48 |
+
],
|
| 49 |
+
"q99": [
|
| 50 |
+
0.5625,
|
| 51 |
+
0.8866071701049805,
|
| 52 |
+
0.9375,
|
| 53 |
+
0.08892857283353806,
|
| 54 |
+
0.12857143580913544,
|
| 55 |
+
0.07178571075201035,
|
| 56 |
+
1.0
|
| 57 |
+
],
|
| 58 |
+
"mask": [
|
| 59 |
+
true,
|
| 60 |
+
true,
|
| 61 |
+
true,
|
| 62 |
+
true,
|
| 63 |
+
true,
|
| 64 |
+
true,
|
| 65 |
+
false
|
| 66 |
+
]
|
| 67 |
+
},
|
| 68 |
+
"state": {
|
| 69 |
+
"mean": [
|
| 70 |
+
0.054813556373119354,
|
| 71 |
+
-0.018270745873451233,
|
| 72 |
+
0.6020475029945374,
|
| 73 |
+
3.1282541751861572,
|
| 74 |
+
-0.09854517877101898,
|
| 75 |
+
0.032103825360536575,
|
| 76 |
+
0.030025552958250046,
|
| 77 |
+
-0.031449686735868454
|
| 78 |
+
],
|
| 79 |
+
"std": [
|
| 80 |
+
0.053855083882808685,
|
| 81 |
+
0.1741994470357895,
|
| 82 |
+
0.07851995527744293,
|
| 83 |
+
0.06309114396571937,
|
| 84 |
+
0.13087928295135498,
|
| 85 |
+
0.1720549315214157,
|
| 86 |
+
0.008607113733887672,
|
| 87 |
+
0.007780037354677916
|
| 88 |
+
],
|
| 89 |
+
"max": [
|
| 90 |
+
0.16465522348880768,
|
| 91 |
+
0.3209042251110077,
|
| 92 |
+
0.756892204284668,
|
| 93 |
+
3.3454883098602295,
|
| 94 |
+
0.30037251114845276,
|
| 95 |
+
0.44232118129730225,
|
| 96 |
+
0.04052062705159187,
|
| 97 |
+
-0.011507377959787846
|
| 98 |
+
],
|
| 99 |
+
"min": [
|
| 100 |
+
-0.06677521765232086,
|
| 101 |
+
-0.26465433835983276,
|
| 102 |
+
0.44561758637428284,
|
| 103 |
+
2.911552906036377,
|
| 104 |
+
-0.6348567008972168,
|
| 105 |
+
-0.7116587162017822,
|
| 106 |
+
0.013648548163473606,
|
| 107 |
+
-0.04069427400827408
|
| 108 |
+
],
|
| 109 |
+
"q01": [
|
| 110 |
+
-0.05987565651535988,
|
| 111 |
+
-0.22870126426219942,
|
| 112 |
+
0.44699052095413205,
|
| 113 |
+
2.9767065334320066,
|
| 114 |
+
-0.44803553462028506,
|
| 115 |
+
-0.4113043105602264,
|
| 116 |
+
0.014268653728067874,
|
| 117 |
+
-0.03993676871061325
|
| 118 |
+
],
|
| 119 |
+
"q99": [
|
| 120 |
+
0.14320090532302857,
|
| 121 |
+
0.28736546516418454,
|
| 122 |
+
0.7209196949005127,
|
| 123 |
+
3.2833758640289306,
|
| 124 |
+
0.17867330253124214,
|
| 125 |
+
0.3982699787616729,
|
| 126 |
+
0.03994086429476738,
|
| 127 |
+
-0.013643189668655397
|
| 128 |
+
]
|
| 129 |
+
},
|
| 130 |
+
"num_transitions": 7473,
|
| 131 |
+
"num_trajectories": 49
|
| 132 |
+
}
|
| 133 |
+
}
|
checkpoints/finetune_lora_task33_2000step/final_model/pytorch_model.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:736ba8329c8be9ad390be023ee2c4543783e0e84027879640fcaa99909988c88
|
| 3 |
+
size 29602442
|
checkpoints/finetune_lora_task33_2000step/summary.jsonl
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"steps": 500}
|
| 2 |
+
{"steps": 1000}
|
| 3 |
+
{"steps": 1500}
|
| 4 |
+
{"steps": 2000}
|