hjx1995 commited on
Commit
c1de8f8
·
verified ·
1 Parent(s): 27fb584

Upload folder using huggingface_hub

Browse files
Files changed (2) hide show
  1. checkpoints/pytorch_model.pt +3 -0
  2. config.yaml +95 -0
checkpoints/pytorch_model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a82a5ce5494ce85a3c5dca1f381195520f0f60824aafbbbe56ccf8d39f21f33
3
+ size 18982885927
config.yaml ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ seed: 42
2
+ trackers:
3
+ - jsonl
4
+ - wandb
5
+ wandb_entity: zbdz
6
+ wandb_project: vla_jiang
7
+ is_debug: false
8
+ framework:
9
+ name: QwenFM
10
+ qwenvl:
11
+ base_vlm: /root/Unifolm-VLM-0
12
+ attn_implementation: flash_attention_2
13
+ vl_hidden_dim: 2048
14
+ model_type: qwen2_5_vl
15
+ action_model:
16
+ input_embedding_dim: 1536
17
+ hidden_size: 1024
18
+ add_pos_embed: true
19
+ max_seq_len: 1024
20
+ action_dim: 7
21
+ state_dim: 7
22
+ future_action_window_size: 15
23
+ action_horizon: 16
24
+ past_action_window_size: 0
25
+ repeated_diffusion_steps: 8
26
+ noise_beta_alpha: 1.5
27
+ noise_beta_beta: 1.0
28
+ noise_s: 0.999
29
+ num_timestep_buckets: 1000
30
+ num_inference_timesteps: 4
31
+ num_target_vision_tokens: 32
32
+ diffusion_model_cfg:
33
+ cross_attention_dim: 2048
34
+ attention_head_dim: 48
35
+ num_attention_heads: 32
36
+ dropout: 0.2
37
+ final_dropout: true
38
+ interleave_self_attention: true
39
+ norm_type: ada_norm
40
+ num_layers: 16
41
+ output_dim: 1024
42
+ positional_embeddings: null
43
+ use_relative_action: false
44
+ framework_py: unifolm_vla
45
+ datasets:
46
+ vla_data:
47
+ data_root_dir: /DATA/disk0/rlds_format_datasets/g1
48
+ data_mix: Unitree_all_task
49
+ per_device_batch_size: 16
50
+ load_all_data_for_training: true
51
+ image_size:
52
+ - 224
53
+ - 224
54
+ window_size: 1
55
+ trainer:
56
+ epochs: 100
57
+ max_train_steps: 300000
58
+ num_warmup_steps: 5000
59
+ save_interval: 10000
60
+ eval_interval: 500
61
+ learning_rate:
62
+ base: 4.0e-05
63
+ qwen_vl_interface: 1.0e-05
64
+ action_model: 0.0004
65
+ lr_scheduler_type: cosine_with_min_lr
66
+ scheduler_specific_kwargs:
67
+ min_lr: 5.0e-07
68
+ freeze_modules: null
69
+ loss_scale:
70
+ vla: 1.0
71
+ vlm: 0.1
72
+ repeated_diffusion_steps: 4
73
+ max_grad_norm: 1.0
74
+ warmup_ratio: 0.1
75
+ weight_decay: 0.0
76
+ logging_frequency: 500
77
+ gradient_clipping: 1.0
78
+ gradient_accumulation_steps: 1
79
+ optimizer:
80
+ name: AdamW
81
+ betas:
82
+ - 0.9
83
+ - 0.95
84
+ eps: 1.0e-08
85
+ weight_decay: 1.0e-08
86
+ is_resume: false
87
+ resume_epoch: null
88
+ resume_step: null
89
+ enable_gradient_checkpointing: true
90
+ enable_mixed_precision_training: true
91
+ shuffle_buffer_size: 10000
92
+ use_wrist_image: true
93
+ use_proprio: true
94
+ action_queries: 0
95
+