ChihHanShen commited on
Commit
74a858c
·
verified ·
1 Parent(s): 9c3e8ee

Upload folder using huggingface_hub

Browse files
merging-checkpoints/checkpoints/task33-merging/config.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ datasets:
2
+ vla_data:
3
+ CoT_prompt: Your task is {instruction}. To identify the key objects for your task.
4
+ Locate their bounding boxes in [x1,y1,x2,y2] format.
5
+ data_mix: libero_90_task_33
6
+ data_root_dir: playground/Datasets/LEROBOT_LIBERO_DATA
7
+ dataset_py: lerobot_datasets
8
+ per_device_batch_size: 1
9
+ sequential_step_sampling: false
10
+ video_backend: torchvision_av
11
+ framework:
12
+ action_model:
13
+ action_dim: 7
14
+ future_action_window_size: 7
15
+ past_action_window_size: 0
16
+ name: QwenFast
17
+ qwenvl:
18
+ base_vlm: playground/Pretrained_models/Qwen2.5-VL-3B-Instruct-Action
19
+ output_dir: ./results/Checkpoints/finetune_task33_2000step
20
+ run_id: finetune_task33_2000step
21
+ run_root_dir: ./results/Checkpoints
22
+ seed: 42
23
+ trainer:
24
+ eval_interval: 100
25
+ freeze_modules: qwen_vl_interface.model.model.visual,dino_encoder
26
+ gradient_accumulation_steps: 1
27
+ gradient_clipping: 1.0
28
+ is_resume: true
29
+ learning_rate:
30
+ action_model: 0.0001
31
+ base: 2.5e-05
32
+ qwen_vl_interface: 1.0e-05
33
+ logging_frequency: 100
34
+ lr_scheduler_type: cosine_with_min_lr
35
+ max_train_steps: 2000
36
+ num_warmup_steps: 5000
37
+ optimizer:
38
+ betas:
39
+ - 0.9
40
+ - 0.95
41
+ eps: 1.0e-08
42
+ weight_decay: 1.0e-08
43
+ pretrained_checkpoint: /content/starVLA_r/results/Checkpoints/Qwen2.5-VL-FAST-LIBERO-4in1/checkpoints/steps_30000_pytorch_model.pt
44
+ save_interval: 500
45
+ scheduler_specific_kwargs:
46
+ min_lr: 1.0e-06
47
+ wandb_entity: michellelin9102-usc
48
+ wandb_project: starVLA_Libero
merging-checkpoints/checkpoints/task33-merging/dataset_statistics.json ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "franka": {
3
+ "action": {
4
+ "mean": [
5
+ 0.03962625563144684,
6
+ 0.13655489683151245,
7
+ -0.0458662211894989,
8
+ -0.0012228331761434674,
9
+ 0.0017042774707078934,
10
+ -0.005507396999746561,
11
+ 0.5082296133041382
12
+ ],
13
+ "std": [
14
+ 0.22318312525749207,
15
+ 0.43824875354766846,
16
+ 0.41923093795776367,
17
+ 0.034977223724126816,
18
+ 0.04809708148241043,
19
+ 0.024638118222355843,
20
+ 0.49992871284484863
21
+ ],
22
+ "max": [
23
+ 0.7232142686843872,
24
+ 0.9160714149475098,
25
+ 0.9375,
26
+ 0.16285714507102966,
27
+ 0.23571428656578064,
28
+ 0.12107142806053162,
29
+ 1.0
30
+ ],
31
+ "min": [
32
+ -0.5758928656578064,
33
+ -0.8357142806053162,
34
+ -0.9375,
35
+ -0.1842857152223587,
36
+ -0.23678570985794067,
37
+ -0.11999999731779099,
38
+ 0.0
39
+ ],
40
+ "q01": [
41
+ -0.4258928596973419,
42
+ -0.7071428298950195,
43
+ -0.8016428351402283,
44
+ -0.11892857402563095,
45
+ -0.11785714328289032,
46
+ -0.08250000327825546,
47
+ 0.0
48
+ ],
49
+ "q99": [
50
+ 0.5625,
51
+ 0.8866071701049805,
52
+ 0.9375,
53
+ 0.08892857283353806,
54
+ 0.12857143580913544,
55
+ 0.07178571075201035,
56
+ 1.0
57
+ ],
58
+ "mask": [
59
+ true,
60
+ true,
61
+ true,
62
+ true,
63
+ true,
64
+ true,
65
+ false
66
+ ]
67
+ },
68
+ "state": {
69
+ "mean": [
70
+ 0.054813556373119354,
71
+ -0.018270745873451233,
72
+ 0.6020475029945374,
73
+ 3.1282541751861572,
74
+ -0.09854517877101898,
75
+ 0.032103825360536575,
76
+ 0.030025552958250046,
77
+ -0.031449686735868454
78
+ ],
79
+ "std": [
80
+ 0.053855083882808685,
81
+ 0.1741994470357895,
82
+ 0.07851995527744293,
83
+ 0.06309114396571937,
84
+ 0.13087928295135498,
85
+ 0.1720549315214157,
86
+ 0.008607113733887672,
87
+ 0.007780037354677916
88
+ ],
89
+ "max": [
90
+ 0.16465522348880768,
91
+ 0.3209042251110077,
92
+ 0.756892204284668,
93
+ 3.3454883098602295,
94
+ 0.30037251114845276,
95
+ 0.44232118129730225,
96
+ 0.04052062705159187,
97
+ -0.011507377959787846
98
+ ],
99
+ "min": [
100
+ -0.06677521765232086,
101
+ -0.26465433835983276,
102
+ 0.44561758637428284,
103
+ 2.911552906036377,
104
+ -0.6348567008972168,
105
+ -0.7116587162017822,
106
+ 0.013648548163473606,
107
+ -0.04069427400827408
108
+ ],
109
+ "q01": [
110
+ -0.05987565591931343,
111
+ -0.22870126366615295,
112
+ 0.4469905197620392,
113
+ 2.9767065048217773,
114
+ -0.4480355381965637,
115
+ -0.4113042950630188,
116
+ 0.014268653467297554,
117
+ -0.039936769753694534
118
+ ],
119
+ "q99": [
120
+ 0.14320091903209686,
121
+ 0.28736546635627747,
122
+ 0.720919668674469,
123
+ 3.2833759784698486,
124
+ 0.17867356538772583,
125
+ 0.3982700705528259,
126
+ 0.039940863847732544,
127
+ -0.013643189333379269
128
+ ]
129
+ },
130
+ "num_transitions": 7473,
131
+ "num_trajectories": 49
132
+ }
133
+ }
merging-checkpoints/checkpoints/task33-merging/models/merged_task33_3over10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9fdf116570bbb76feee7adcb50f12db76273ab36b88c91622b2ee8e71d40b1e2
3
+ size 8146435699
merging-checkpoints/checkpoints/task33-merging/models/merged_task33_7over10.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bc305216beac60161574227b5d086862487a5461fded1d1a9da1c96ea9da67d
3
+ size 8146435699