rrmura commited on
Commit
a6c7f74
·
verified ·
1 Parent(s): 5fd0a5f

Upload folder using huggingface_hub

Browse files
checkpoints/finetune_lora_task2_2000step/checkpoints/steps_1000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d94f509bc0579b3e910c8a43fbe07b6872fedd6c6fce99cecedec5559103b22
3
+ size 29605718
checkpoints/finetune_lora_task2_2000step/checkpoints/steps_1500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93690cc1cd4fb407263b9547216b125b22adbfcf499f0d6d257ad5a2bf06d19f
3
+ size 29605718
checkpoints/finetune_lora_task2_2000step/checkpoints/steps_2000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36f33c154ac2b28c75d45bdaa759ac77ef0a7db3c0c064304ed59563d67581a6
3
+ size 29605718
checkpoints/finetune_lora_task2_2000step/checkpoints/steps_500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6a7b4c004188c4cd9eb44d6745f876eb4e708281c98b64005e316e41767ca03
3
+ size 29605426
checkpoints/finetune_lora_task2_2000step/config.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ datasets:
2
+ vla_data:
3
+ CoT_prompt: Your task is {instruction}. To identify the key objects for your task.
4
+ Locate their bounding boxes in [x1,y1,x2,y2] format.
5
+ data_mix: libero_90_task_2
6
+ data_root_dir: playground/Datasets/LEROBOT_LIBERO_DATA
7
+ dataset_py: lerobot_datasets
8
+ per_device_batch_size: 1
9
+ sequential_step_sampling: false
10
+ video_backend: torchvision_av
11
+ framework:
12
+ action_model:
13
+ action_dim: 7
14
+ future_action_window_size: 7
15
+ past_action_window_size: 0
16
+ name: QwenFast
17
+ qwenvl:
18
+ base_vlm: playground/Pretrained_models/Qwen2.5-VL-3B-Instruct-Action
19
+ lora_config:
20
+ lora_alpha: 32
21
+ lora_dropout: 0.05
22
+ r: 16
23
+ target_modules:
24
+ - q_proj
25
+ - v_proj
26
+ - k_proj
27
+ - o_proj
28
+ output_dir: ./results/Checkpoints/finetune_lora_task2_2000step
29
+ run_id: finetune_lora_task2_2000step
30
+ run_root_dir: ./results/Checkpoints
31
+ seed: 42
32
+ trainer:
33
+ eval_interval: 100
34
+ freeze_modules: qwen_vl_interface.model.model.visual,dino_encoder
35
+ gradient_accumulation_steps: 1
36
+ gradient_clipping: 1.0
37
+ is_resume: true
38
+ learning_rate:
39
+ action_model: 0.0001
40
+ base: 2.5e-05
41
+ qwen_vl_interface: 1.0e-05
42
+ logging_frequency: 100
43
+ lr_scheduler_type: cosine_with_min_lr
44
+ max_train_steps: 2000
45
+ num_warmup_steps: 5000
46
+ optimizer:
47
+ betas:
48
+ - 0.9
49
+ - 0.95
50
+ eps: 1.0e-08
51
+ weight_decay: 1.0e-08
52
+ pretrained_checkpoint: /content/starVLA_r/results/Checkpoints/Qwen2.5-VL-FAST-LIBERO-4in1/checkpoints/steps_30000_pytorch_model.pt
53
+ save_interval: 500
54
+ scheduler_specific_kwargs:
55
+ min_lr: 1.0e-06
56
+ use_lora: true
57
+ wandb_entity: michellelin9102-usc
58
+ wandb_project: starVLA_Libero
checkpoints/finetune_lora_task2_2000step/dataset_statistics.json ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "franka": {
3
+ "action": {
4
+ "mean": [
5
+ 0.03965260088443756,
6
+ 0.13710077106952667,
7
+ -0.04964581876993179,
8
+ -0.00436883419752121,
9
+ 0.0031783515587449074,
10
+ -0.00018181839550379664,
11
+ 0.5124579071998596
12
+ ],
13
+ "std": [
14
+ 0.24294555187225342,
15
+ 0.44865477085113525,
16
+ 0.44734615087509155,
17
+ 0.0339176170527935,
18
+ 0.04405592009425163,
19
+ 0.029885200783610344,
20
+ 0.49982890486717224
21
+ ],
22
+ "max": [
23
+ 0.7794643044471741,
24
+ 0.9375,
25
+ 0.9375,
26
+ 0.19499999284744263,
27
+ 0.1907142847776413,
28
+ 0.19928571581840515,
29
+ 1.0
30
+ ],
31
+ "min": [
32
+ -0.7151785492897034,
33
+ -0.8999999761581421,
34
+ -0.9241071343421936,
35
+ -0.16821429133415222,
36
+ -0.167142853140831,
37
+ -0.12964285910129547,
38
+ 0.0
39
+ ],
40
+ "q01": [
41
+ -0.46875,
42
+ -0.7232142686843872,
43
+ -0.8169642686843872,
44
+ -0.10821428894996643,
45
+ -0.11571428924798965,
46
+ -0.08142857253551483,
47
+ 0.0
48
+ ],
49
+ "q99": [
50
+ 0.5839285850524902,
51
+ 0.8919642567634583,
52
+ 0.9375,
53
+ 0.09535714238882065,
54
+ 0.13984285354614304,
55
+ 0.1039285734295845,
56
+ 1.0
57
+ ],
58
+ "mask": [
59
+ true,
60
+ true,
61
+ true,
62
+ true,
63
+ true,
64
+ true,
65
+ false
66
+ ]
67
+ },
68
+ "state": {
69
+ "mean": [
70
+ 0.06304012984037399,
71
+ -0.02723514847457409,
72
+ 0.5950468182563782,
73
+ 3.1040256023406982,
74
+ -0.0479881688952446,
75
+ -0.014697893522679806,
76
+ 0.029381589964032173,
77
+ -0.030202925205230713
78
+ ],
79
+ "std": [
80
+ 0.05494280904531479,
81
+ 0.17417463660240173,
82
+ 0.08279268443584442,
83
+ 0.06757557392120361,
84
+ 0.16604064404964447,
85
+ 0.1603231579065323,
86
+ 0.00942574255168438,
87
+ 0.009197638370096684
88
+ ],
89
+ "max": [
90
+ 0.17418493330478668,
91
+ 0.30584609508514404,
92
+ 0.7395508289337158,
93
+ 3.3254528045654297,
94
+ 0.5380978584289551,
95
+ 0.45999088883399963,
96
+ 0.04025300219655037,
97
+ -0.008219979703426361
98
+ ],
99
+ "min": [
100
+ -0.08505505323410034,
101
+ -0.24681705236434937,
102
+ 0.4457172751426697,
103
+ 2.8618643283843994,
104
+ -0.6842642426490784,
105
+ -0.5939062833786011,
106
+ 0.0075335511937737465,
107
+ -0.04111039638519287
108
+ ],
109
+ "q01": [
110
+ -0.06130984306335449,
111
+ -0.23173466563224793,
112
+ 0.4463089668750763,
113
+ 2.898547372817993,
114
+ -0.5309021449089051,
115
+ -0.40839494705200197,
116
+ 0.009174905717372894,
117
+ -0.04018907129764557
118
+ ],
119
+ "q99": [
120
+ 0.154893668293953,
121
+ 0.2796362054347992,
122
+ 0.719877290725708,
123
+ 3.2510776233673098,
124
+ 0.383401458263398,
125
+ 0.3866866183280948,
126
+ 0.03991854026913643,
127
+ -0.008571043685078621
128
+ ]
129
+ },
130
+ "num_transitions": 7425,
131
+ "num_trajectories": 49
132
+ }
133
+ }
checkpoints/finetune_lora_task2_2000step/final_model/pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5b6da0a551d45671264cfa4f450a885094b44fac00f0d4e8fa89b030c3e4b8c
3
+ size 29602442
checkpoints/finetune_lora_task2_2000step/summary.jsonl ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {"steps": 500}
2
+ {"steps": 1000}
3
+ {"steps": 1500}
4
+ {"steps": 2000}