rrmura commited on
Commit
5fd0a5f
·
verified ·
1 Parent(s): 5f5fe41

Upload folder using huggingface_hub

Browse files
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_1000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6e3181b89b99e764882fe6fcfd5a41b586d61300ed5eeba697baebd87ba82b0
3
+ size 29605718
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_1500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ed0f5a284fd054c5af1b35bbdae34930febba5acd2ea039900bf632ec99e1ca
3
+ size 29605718
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_2000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4d0ca206384408fc8b0c5d69a20baf5549ea479935b5d43a5dae86c91e5873b
3
+ size 29605718
checkpoints/finetune_lora_task33_2000step/checkpoints/steps_500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6ce393d900b98e31cc5446934c5a60c6937c1a5a7dcefdf518263cd15623648
3
+ size 29605426
checkpoints/finetune_lora_task33_2000step/config.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ datasets:
2
+ vla_data:
3
+ CoT_prompt: Your task is {instruction}. To identify the key objects for your task.
4
+ Locate their bounding boxes in [x1,y1,x2,y2] format.
5
+ data_mix: libero_90_task_33
6
+ data_root_dir: playground/Datasets/LEROBOT_LIBERO_DATA
7
+ dataset_py: lerobot_datasets
8
+ per_device_batch_size: 1
9
+ sequential_step_sampling: false
10
+ video_backend: torchvision_av
11
+ framework:
12
+ action_model:
13
+ action_dim: 7
14
+ future_action_window_size: 7
15
+ past_action_window_size: 0
16
+ name: QwenFast
17
+ qwenvl:
18
+ base_vlm: playground/Pretrained_models/Qwen2.5-VL-3B-Instruct-Action
19
+ lora_config:
20
+ lora_alpha: 32
21
+ lora_dropout: 0.05
22
+ r: 16
23
+ target_modules:
24
+ - q_proj
25
+ - v_proj
26
+ - k_proj
27
+ - o_proj
28
+ output_dir: ./results/Checkpoints/finetune_lora_task33_2000step
29
+ run_id: finetune_lora_task33_2000step
30
+ run_root_dir: ./results/Checkpoints
31
+ seed: 42
32
+ trainer:
33
+ eval_interval: 100
34
+ freeze_modules: qwen_vl_interface.model.model.visual,dino_encoder
35
+ gradient_accumulation_steps: 1
36
+ gradient_clipping: 1.0
37
+ is_resume: true
38
+ learning_rate:
39
+ action_model: 0.0001
40
+ base: 2.5e-05
41
+ qwen_vl_interface: 1.0e-05
42
+ logging_frequency: 100
43
+ lr_scheduler_type: cosine_with_min_lr
44
+ max_train_steps: 2000
45
+ num_warmup_steps: 5000
46
+ optimizer:
47
+ betas:
48
+ - 0.9
49
+ - 0.95
50
+ eps: 1.0e-08
51
+ weight_decay: 1.0e-08
52
+ pretrained_checkpoint: /content/starVLA_r/results/Checkpoints/Qwen2.5-VL-FAST-LIBERO-4in1/checkpoints/steps_30000_pytorch_model.pt
53
+ save_interval: 500
54
+ scheduler_specific_kwargs:
55
+ min_lr: 1.0e-06
56
+ use_lora: true
57
+ wandb_entity: michellelin9102-usc
58
+ wandb_project: starVLA_Libero
checkpoints/finetune_lora_task33_2000step/dataset_statistics.json ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "franka": {
3
+ "action": {
4
+ "mean": [
5
+ 0.03962625563144684,
6
+ 0.13655489683151245,
7
+ -0.0458662211894989,
8
+ -0.0012228331761434674,
9
+ 0.0017042774707078934,
10
+ -0.005507396999746561,
11
+ 0.5082296133041382
12
+ ],
13
+ "std": [
14
+ 0.22318312525749207,
15
+ 0.43824875354766846,
16
+ 0.41923093795776367,
17
+ 0.034977223724126816,
18
+ 0.04809708148241043,
19
+ 0.024638118222355843,
20
+ 0.49992871284484863
21
+ ],
22
+ "max": [
23
+ 0.7232142686843872,
24
+ 0.9160714149475098,
25
+ 0.9375,
26
+ 0.16285714507102966,
27
+ 0.23571428656578064,
28
+ 0.12107142806053162,
29
+ 1.0
30
+ ],
31
+ "min": [
32
+ -0.5758928656578064,
33
+ -0.8357142806053162,
34
+ -0.9375,
35
+ -0.1842857152223587,
36
+ -0.23678570985794067,
37
+ -0.11999999731779099,
38
+ 0.0
39
+ ],
40
+ "q01": [
41
+ -0.4258928596973419,
42
+ -0.7071428298950195,
43
+ -0.8016428303718567,
44
+ -0.11892857402563095,
45
+ -0.11785714328289032,
46
+ -0.08250000327825546,
47
+ 0.0
48
+ ],
49
+ "q99": [
50
+ 0.5625,
51
+ 0.8866071701049805,
52
+ 0.9375,
53
+ 0.08892857283353806,
54
+ 0.12857143580913544,
55
+ 0.07178571075201035,
56
+ 1.0
57
+ ],
58
+ "mask": [
59
+ true,
60
+ true,
61
+ true,
62
+ true,
63
+ true,
64
+ true,
65
+ false
66
+ ]
67
+ },
68
+ "state": {
69
+ "mean": [
70
+ 0.054813556373119354,
71
+ -0.018270745873451233,
72
+ 0.6020475029945374,
73
+ 3.1282541751861572,
74
+ -0.09854517877101898,
75
+ 0.032103825360536575,
76
+ 0.030025552958250046,
77
+ -0.031449686735868454
78
+ ],
79
+ "std": [
80
+ 0.053855083882808685,
81
+ 0.1741994470357895,
82
+ 0.07851995527744293,
83
+ 0.06309114396571937,
84
+ 0.13087928295135498,
85
+ 0.1720549315214157,
86
+ 0.008607113733887672,
87
+ 0.007780037354677916
88
+ ],
89
+ "max": [
90
+ 0.16465522348880768,
91
+ 0.3209042251110077,
92
+ 0.756892204284668,
93
+ 3.3454883098602295,
94
+ 0.30037251114845276,
95
+ 0.44232118129730225,
96
+ 0.04052062705159187,
97
+ -0.011507377959787846
98
+ ],
99
+ "min": [
100
+ -0.06677521765232086,
101
+ -0.26465433835983276,
102
+ 0.44561758637428284,
103
+ 2.911552906036377,
104
+ -0.6348567008972168,
105
+ -0.7116587162017822,
106
+ 0.013648548163473606,
107
+ -0.04069427400827408
108
+ ],
109
+ "q01": [
110
+ -0.05987565651535988,
111
+ -0.22870126426219942,
112
+ 0.44699052095413205,
113
+ 2.9767065334320066,
114
+ -0.44803553462028506,
115
+ -0.4113043105602264,
116
+ 0.014268653728067874,
117
+ -0.03993676871061325
118
+ ],
119
+ "q99": [
120
+ 0.14320090532302857,
121
+ 0.28736546516418454,
122
+ 0.7209196949005127,
123
+ 3.2833758640289306,
124
+ 0.17867330253124214,
125
+ 0.3982699787616729,
126
+ 0.03994086429476738,
127
+ -0.013643189668655397
128
+ ]
129
+ },
130
+ "num_transitions": 7473,
131
+ "num_trajectories": 49
132
+ }
133
+ }
checkpoints/finetune_lora_task33_2000step/final_model/pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:736ba8329c8be9ad390be023ee2c4543783e0e84027879640fcaa99909988c88
3
+ size 29602442
checkpoints/finetune_lora_task33_2000step/summary.jsonl ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {"steps": 500}
2
+ {"steps": 1000}
3
+ {"steps": 1500}
4
+ {"steps": 2000}