rrmura commited on
Commit
5f5fe41
·
verified ·
1 Parent(s): 883d93b

Upload folder using huggingface_hub

Browse files
checkpoints/finetune_lora_task48_2000step/checkpoints/steps_1000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:472afe4f3c132c3288ece1548ae0a855ead1d8fb5779b883bdd0d514d9967f0e
3
+ size 29605718
checkpoints/finetune_lora_task48_2000step/checkpoints/steps_1500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23f9bc42bc118a80fdbe706a2827e8adbe13ee1b13cff642b65129c806a27917
3
+ size 29605718
checkpoints/finetune_lora_task48_2000step/checkpoints/steps_2000_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:662d89444c29d10e59e36053fc1daedba0a1eafe47b823a6ea2f58f7f22f3b58
3
+ size 29605718
checkpoints/finetune_lora_task48_2000step/checkpoints/steps_500_pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8471699093614a5630c9fba07925ad5a460e10e9cd53d9d97502be9b1052f7cf
3
+ size 29605426
checkpoints/finetune_lora_task48_2000step/config.yaml ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ datasets:
2
+ vla_data:
3
+ CoT_prompt: Your task is {instruction}. To identify the key objects for your task.
4
+ Locate their bounding boxes in [x1,y1,x2,y2] format.
5
+ data_mix: libero_90_task_48
6
+ data_root_dir: playground/Datasets/LEROBOT_LIBERO_DATA
7
+ dataset_py: lerobot_datasets
8
+ per_device_batch_size: 1
9
+ sequential_step_sampling: false
10
+ video_backend: torchvision_av
11
+ framework:
12
+ action_model:
13
+ action_dim: 7
14
+ future_action_window_size: 7
15
+ past_action_window_size: 0
16
+ name: QwenFast
17
+ qwenvl:
18
+ base_vlm: playground/Pretrained_models/Qwen2.5-VL-3B-Instruct-Action
19
+ lora_config:
20
+ lora_alpha: 32
21
+ lora_dropout: 0.05
22
+ r: 16
23
+ target_modules:
24
+ - q_proj
25
+ - v_proj
26
+ - k_proj
27
+ - o_proj
28
+ output_dir: ./results/Checkpoints/finetune_lora_task48_2000step
29
+ run_id: finetune_lora_task48_2000step
30
+ run_root_dir: ./results/Checkpoints
31
+ seed: 42
32
+ trainer:
33
+ eval_interval: 100
34
+ freeze_modules: qwen_vl_interface.model.model.visual,dino_encoder
35
+ gradient_accumulation_steps: 1
36
+ gradient_clipping: 1.0
37
+ is_resume: true
38
+ learning_rate:
39
+ action_model: 0.0001
40
+ base: 2.5e-05
41
+ qwen_vl_interface: 1.0e-05
42
+ logging_frequency: 100
43
+ lr_scheduler_type: cosine_with_min_lr
44
+ max_train_steps: 2000
45
+ num_warmup_steps: 5000
46
+ optimizer:
47
+ betas:
48
+ - 0.9
49
+ - 0.95
50
+ eps: 1.0e-08
51
+ weight_decay: 1.0e-08
52
+ pretrained_checkpoint: /content/starVLA_r/results/Checkpoints/Qwen2.5-VL-FAST-LIBERO-4in1/checkpoints/steps_30000_pytorch_model.pt
53
+ save_interval: 500
54
+ scheduler_specific_kwargs:
55
+ min_lr: 1.0e-06
56
+ use_lora: true
57
+ wandb_entity: michellelin9102-usc
58
+ wandb_project: starVLA_Libero
checkpoints/finetune_lora_task48_2000step/dataset_statistics.json ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "franka": {
3
+ "action": {
4
+ "mean": [
5
+ 0.04281630367040634,
6
+ 0.16635118424892426,
7
+ -0.10021582990884781,
8
+ 0.0008796637994237244,
9
+ 0.009483124129474163,
10
+ -0.0018370545003563166,
11
+ 0.48476770520210266
12
+ ],
13
+ "std": [
14
+ 0.23814928531646729,
15
+ 0.48973721265792847,
16
+ 0.4945138692855835,
17
+ 0.03829231485724449,
18
+ 0.048698168247938156,
19
+ 0.031572677195072174,
20
+ 0.4997621178627014
21
+ ],
22
+ "max": [
23
+ 0.8142856955528259,
24
+ 0.9375,
25
+ 0.9375,
26
+ 0.1971428543329239,
27
+ 0.3385714292526245,
28
+ 0.1703571379184723,
29
+ 1.0
30
+ ],
31
+ "min": [
32
+ -0.7392857074737549,
33
+ -0.9375,
34
+ -0.9375,
35
+ -0.17571428418159485,
36
+ -0.21642857789993286,
37
+ -0.2442857176065445,
38
+ 0.0
39
+ ],
40
+ "q01": [
41
+ -0.545062488913536,
42
+ -0.8571428656578064,
43
+ -0.9375,
44
+ -0.10607142746448517,
45
+ -0.11142857372760773,
46
+ -0.09857142716646194,
47
+ 0.0
48
+ ],
49
+ "q99": [
50
+ 0.6401785612106323,
51
+ 0.9375,
52
+ 0.9375,
53
+ 0.11249999701976776,
54
+ 0.16338214248418784,
55
+ 0.07981071211397625,
56
+ 1.0
57
+ ],
58
+ "mask": [
59
+ true,
60
+ true,
61
+ true,
62
+ true,
63
+ true,
64
+ true,
65
+ false
66
+ ]
67
+ },
68
+ "state": {
69
+ "mean": [
70
+ -0.0446692518889904,
71
+ -0.009891818277537823,
72
+ 0.5988359451293945,
73
+ 3.129826545715332,
74
+ -0.04531480744481087,
75
+ -0.08344470709562302,
76
+ 0.03376416116952896,
77
+ -0.03510740399360657
78
+ ],
79
+ "std": [
80
+ 0.03938385099172592,
81
+ 0.15407362580299377,
82
+ 0.07102718949317932,
83
+ 0.057501792907714844,
84
+ 0.13009506464004517,
85
+ 0.13546207547187805,
86
+ 0.005857877433300018,
87
+ 0.005015654023736715
88
+ ],
89
+ "max": [
90
+ 0.08624760061502457,
91
+ 0.32907697558403015,
92
+ 0.7157319784164429,
93
+ 3.3233284950256348,
94
+ 0.5133177638053894,
95
+ 0.3517029881477356,
96
+ 0.04034854471683502,
97
+ -0.017706822603940964
98
+ ],
99
+ "min": [
100
+ -0.13770002126693726,
101
+ -0.1906675398349762,
102
+ 0.47950780391693115,
103
+ 2.9164979457855225,
104
+ -0.6462156176567078,
105
+ -0.6787144541740417,
106
+ 0.013480174355208874,
107
+ -0.041075896471738815
108
+ ],
109
+ "q01": [
110
+ -0.122961335927248,
111
+ -0.18305308774113654,
112
+ 0.4812150990962982,
113
+ 2.990795373916626,
114
+ -0.43866351157426836,
115
+ -0.4504926311969757,
116
+ 0.017378225158900022,
117
+ -0.040015368908643725
118
+ ],
119
+ "q99": [
120
+ 0.05706421278417107,
121
+ 0.28637341052293763,
122
+ 0.694216803908348,
123
+ 3.2584143328666686,
124
+ 0.28781155496835636,
125
+ 0.28255320578813553,
126
+ 0.0398839446529746,
127
+ -0.019760548584163196
128
+ ]
129
+ },
130
+ "num_transitions": 5252,
131
+ "num_trajectories": 44
132
+ }
133
+ }
checkpoints/finetune_lora_task48_2000step/final_model/pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6522b7d0f06bff4ef703036afb14d6ea5689f4c9218d4b9befa69cea4ea2a97b
3
+ size 29602442
checkpoints/finetune_lora_task48_2000step/summary.jsonl ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {"steps": 500}
2
+ {"steps": 1000}
3
+ {"steps": 1500}
4
+ {"steps": 2000}