SII-LibAI commited on
Commit
63e2cf8
·
verified ·
1 Parent(s): 0e1ee2f

upload model directory

Browse files
config.yaml ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ datasets:
2
+ vla_data:
3
+ CoT_prompt: Your task is {instruction}. To identify the key objects for your task.
4
+ Locate their bounding boxes in [x1,y1,x2,y2] format.
5
+ data_mix: libero_all
6
+ data_root_dir: /inspire/ssd/project/embodied-basic-model/zhangjianing-253108140206/experiment/starVLA/playground/Datasets/LEROBOT_LIBERO_DATA/libero
7
+ dataset_py: lerobot_datasets
8
+ per_device_batch_size: 8
9
+ video_backend: torchvision_av
10
+ framework:
11
+ action_model:
12
+ action_dim: 7
13
+ action_hidden_dim: 2560
14
+ action_model_type: DiT-B
15
+ future_action_window_size: 7
16
+ past_action_window_size: 0
17
+ name: QwenOFT
18
+ qwenvl:
19
+ base_vlm: /inspire/qb-ilm/project/embodied-basic-model/zhangjianing-253108140206/model/cubev0-200000-Qwen3-VL
20
+ output_dir: ./results/Checkpoints/125_cube_oft_gr00t
21
+ run_id: 125_cube_oft_gr00t
22
+ run_root_dir: ./results/Checkpoints
23
+ seed: 42
24
+ trainer:
25
+ eval_interval: 1000
26
+ freeze_modules: true
27
+ gradient_accumulation_steps: 1
28
+ gradient_clipping: 1.0
29
+ is_resume: false
30
+ learning_rate:
31
+ action_model: 0.0001
32
+ base: 2.5e-05
33
+ qwen_vl_interface: 1.0e-05
34
+ logging_frequency: 10
35
+ lr_scheduler_type: cosine_with_min_lr
36
+ max_train_steps: 30000
37
+ num_warmup_steps: 100
38
+ optimizer:
39
+ betas:
40
+ - 0.9
41
+ - 0.95
42
+ eps: 1.0e-08
43
+ weight_decay: 1.0e-08
44
+ save_interval: 5000
45
+ scheduler_specific_kwargs:
46
+ min_lr: 1.0e-06
47
+ wandb_entity: 1732949190-tongji-university
48
+ wandb_project: wallx4libero
dataset_statistics.json ADDED
@@ -0,0 +1,133 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "franka": {
3
+ "action": {
4
+ "mean": [
5
+ 0.07237596483901143,
6
+ 0.08987006871029735,
7
+ -0.10144743137061596,
8
+ -0.00045383188989944756,
9
+ 0.006273590726777911,
10
+ -0.003878799732774496,
11
+ 0.524486355483532
12
+ ],
13
+ "std": [
14
+ 0.3498823308902479,
15
+ 0.37794140366375184,
16
+ 0.460084266976933,
17
+ 0.0403885784928603,
18
+ 0.06616144248501059,
19
+ 0.07763074391911857,
20
+ 0.4994683356809767
21
+ ],
22
+ "max": [
23
+ 0.9375,
24
+ 0.9375,
25
+ 0.9375,
26
+ 0.3557142913341522,
27
+ 0.375,
28
+ 0.375,
29
+ 1.0
30
+ ],
31
+ "min": [
32
+ -0.9375,
33
+ -0.9375,
34
+ -0.9375,
35
+ -0.2582142949104309,
36
+ -0.375,
37
+ -0.3675000071525574,
38
+ 0.0
39
+ ],
40
+ "q01": [
41
+ -0.8785714507102966,
42
+ -0.8758928775787354,
43
+ -0.9375,
44
+ -0.1510714292526245,
45
+ -0.20678570866584778,
46
+ -0.2742857038974762,
47
+ 0.0
48
+ ],
49
+ "q99": [
50
+ 0.9375,
51
+ 0.9107142686843872,
52
+ 0.9375,
53
+ 0.20357142388820648,
54
+ 0.26357144117355347,
55
+ 0.375,
56
+ 1.0
57
+ ],
58
+ "mask": [
59
+ true,
60
+ true,
61
+ true,
62
+ true,
63
+ true,
64
+ true,
65
+ false
66
+ ]
67
+ },
68
+ "state": {
69
+ "mean": [
70
+ -0.04889854742214084,
71
+ 0.03689368185587227,
72
+ 0.7890402488410473,
73
+ 2.9771945476531982,
74
+ -0.1417286954820156,
75
+ -0.11769362539052963,
76
+ 0.026436020154505968,
77
+ -0.02665513101965189
78
+ ],
79
+ "std": [
80
+ 0.10639013941746686,
81
+ 0.15115733130675715,
82
+ 0.38406895599530033,
83
+ 0.3530238395244304,
84
+ 0.8227341427331599,
85
+ 0.32357567121520087,
86
+ 0.014583991652936385,
87
+ 0.014467005007200339
88
+ ],
89
+ "max": [
90
+ 0.21031762659549713,
91
+ 0.39128610491752625,
92
+ 1.3660105466842651,
93
+ 3.6714255809783936,
94
+ 3.560650587081909,
95
+ 1.386339545249939,
96
+ 0.04233968257904053,
97
+ 0.0013633022317662835
98
+ ],
99
+ "min": [
100
+ -0.4828203022480011,
101
+ -0.3255046010017395,
102
+ 0.008128180168569088,
103
+ 0.35277295112609863,
104
+ -3.641430377960205,
105
+ -1.842738389968872,
106
+ -0.0013586411951109767,
107
+ -0.042040832340717316
108
+ ],
109
+ "q01": [
110
+ -0.42401049643754957,
111
+ -0.2838300323486328,
112
+ 0.009925739830359817,
113
+ 1.3085840785503386,
114
+ -2.886677579879761,
115
+ -1.1599004411697387,
116
+ 0.001503719249740243,
117
+ -0.040336399003863335
118
+ ],
119
+ "q99": [
120
+ 0.1530261474847791,
121
+ 0.3629165390133857,
122
+ 1.2910678112506866,
123
+ 3.303542451858519,
124
+ 2.7496529006957933,
125
+ 0.6893712210655194,
126
+ 0.040610933862626555,
127
+ -0.0015016929572448147
128
+ ]
129
+ },
130
+ "num_transitions": 273465,
131
+ "num_trajectories": 1693
132
+ }
133
+ }
final_model/pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1b72852b8f4873c3d8734192115b391e711a0bd913b851b590ac45f1deff289
3
+ size 9803309835
run_libero_train.sh ADDED
@@ -0,0 +1,79 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ # export NCCL_SOCKET_IFNAME=bond0
4
+ # export NCCL_IB_HCA=mlx5_2,mlx5_3
5
+ # export NCCL_DEBUG=INFO # 输出调试信息,帮助查找问题
6
+ # export NCCL_IB_DISABLE=1 # 禁用 InfiniBand,防止某些网络设备问题
7
+ # export NCCL_SOCKET_IFNAME=eth0 # 设置网络接口
8
+
9
+ # # used for check save when communication
10
+ # export NCCL_BLOCKING_WAIT=1
11
+ # export NCCL_ASYNC_ERROR_HANDLING=1
12
+ # export NCCL_TIMEOUT=10000 # timeout set to 1 hour (unit: seconds)
13
+ # export NCCL_SOCKET_TIMEOUT_MS=360000
14
+ ###########################################################################################
15
+ # === Please modify the following paths according to hf_iukkofmmRdUqCdqdqclmFjSOktKYvSrOjMyour environment ===
16
+ ##Gr00t是125cubelr
17
+ #oft是125cubeoftlibero
18
+ Framework_name=QwenOFT
19
+ freeze_module_list=''
20
+ base_vlm=/inspire/qb-ilm/project/embodied-basic-model/zhangjianing-253108140206/model/cubev0-200000-Qwen3-VL
21
+ config_yaml=./examples/LIBERO/train_files/starvla_cotrain_libero.yaml
22
+ libero_data_root=/inspire/ssd/project/embodied-basic-model/zhangjianing-253108140206/experiment/starVLA/playground/Datasets/LEROBOT_LIBERO_DATA/libero
23
+ data_mix=libero_all
24
+ run_root_dir=./results/Checkpoints
25
+ run_id=125_cube_oft_gr00t
26
+ # === End of environment variable configuration ===
27
+ ###########################################################################################
28
+ export WANDB_MODE=offline
29
+
30
+ # export WANDB_MODE=disabled
31
+ #examples/LIBERO/train_files/run_libero_train.sh
32
+ output_dir=${run_root_dir}/${run_id}
33
+ mkdir -p ${output_dir}
34
+ # mv this script to the output dir
35
+ cp $0 ${output_dir}/
36
+
37
+
38
+ accelerate launch \
39
+ --config_file starVLA/config/deepseeds/deepspeed_zero2.yaml \
40
+ --num_processes 4 \
41
+ starVLA/training/train_starvla.py \
42
+ --config_yaml ${config_yaml} \
43
+ --framework.name ${Framework_name} \
44
+ --framework.qwenvl.base_vlm ${base_vlm} \
45
+ --datasets.vla_data.data_root_dir ${libero_data_root}\
46
+ --datasets.vla_data.data_mix ${data_mix} \
47
+ --datasets.vla_data.per_device_batch_size 8 \
48
+ --trainer.vla_data.video_backend torchvision_av \
49
+ --trainer.freeze_modules ${freeze_module_list} \
50
+ --trainer.max_train_steps 30000 \
51
+ --trainer.save_interval 5000 \
52
+ --trainer.logging_frequency 10 \
53
+ --trainer.eval_interval 1000 \
54
+ --run_root_dir ${run_root_dir} \
55
+ --run_id ${run_id} \
56
+ --wandb_project wallx4libero \
57
+ --wandb_entity 1732949190-tongji-university \
58
+ # --is_debug True
59
+
60
+
61
+
62
+ # #### Multi-Server Multi-GPU training script #####
63
+ # accelerate launch \
64
+
65
+ # --config_file starVLA/config/deepseeds/deepspeed_zero2.yaml \
66
+ # --main_process_ip $MASTER_ADDR \
67
+ # --main_process_port $MASTER_PORT \
68
+ # --machine_rank $SLURM_PROCID \
69
+ # --num_machines $SLURM_NNODES \
70
+ # --num_processes=${TOTAL_GPUS} \
71
+ # starVLA/training/train_starvla.py \
72
+ # --config_yaml ${config_yaml} \
73
+ # --framework.name ${Framework_name} \
74
+ # --framework.qwenvl.base_vlm ${base_vlm} \
75
+ # --run_root_dir ${run_root_dir} \
76
+ # --run_id ${run_id} \
77
+ # --wandb_project your_project \
78
+ # --wandb_entity your_name
79
+ # ##### Multi-Server Multi-GPU training script #####
summary.jsonl ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"steps": 5000}
2
+ {"steps": 10000}
3
+ {"steps": 15000}
4
+ {"steps": 20000}
5
+ {"steps": 25000}
6
+ {"steps": 30000}