D-eon commited on
Commit
b42718f
·
verified ·
1 Parent(s): 1834e39

Delete rc_robomind_alohas_ft_put_opener_in_drawer

Browse files
rc_robomind_alohas_ft_put_opener_in_drawer/step_33000/config.yaml DELETED
@@ -1,343 +0,0 @@
1
- run_name: molmo_7b_robomind_alohas_ft_aloha_put_opener_in_drawer_20260105_024701
2
- seed: 6198
3
- epoch: null
4
- dry_run: false
5
- model:
6
- d_model: 3584
7
- n_heads: 28
8
- n_kv_heads: 4
9
- qkv_bias: true
10
- clip_qkv: null
11
- n_layers: 28
12
- mlp_ratio: 4
13
- mlp_hidden_size: 37888
14
- activation_type: swiglu
15
- block_type: sequential
16
- block_group_size: 1
17
- rope: true
18
- rope_full_precision: true
19
- rope_theta: 1000000.0
20
- vision_backbone:
21
- image_model_type: openai
22
- image_default_input_size:
23
- - 336
24
- - 336
25
- image_patch_size: 14
26
- image_pos_patch_size: 14
27
- image_emb_dim: 1024
28
- image_num_heads: 16
29
- image_num_key_value_heads: 16
30
- image_num_layers: 23
31
- image_head_dim: 64
32
- image_mlp_dim: 4096
33
- image_mlp_activations: quick_gelu
34
- image_dropout_rate: 0.0
35
- image_num_pos: 577
36
- image_norm_eps: 1.0e-05
37
- attention_dropout: 0.0
38
- residual_dropout: 0.0
39
- initializer_range: 0.02
40
- fsdp_wrap: false
41
- resize_mode: default
42
- vit_load_path: /weka/oe-training-default/mm-olmo/pretrained_image_encoders/vit-l-14-336.pt
43
- llm_load_path: /weka/oe-training-default/mm-olmo/pretrained_llms/qwen2-7b.pt
44
- low_cpu_fsdp: false
45
- attention_type: sdpa
46
- float32_attention: true
47
- attention_dropout: 0.0
48
- attention_layer_norm: false
49
- residual_dropout: 0.1
50
- response_residual_dropout: 0.0
51
- embedding_dropout: 0.0
52
- layer_norm_type: rms
53
- layer_norm_with_affine: true
54
- layer_norm_eps: 1.0e-06
55
- attention_layer_norm_with_affine: true
56
- max_sequence_length: 4096
57
- max_position_embeddings: null
58
- include_bias: false
59
- bias_for_layer_norm: null
60
- scale_logits: false
61
- vocab_size: 152064
62
- embedding_size: 152064
63
- ff_out_size: null
64
- additional_vocab_size: 128
65
- new_embedding_init_range: 0.02
66
- weight_tying: false
67
- init_device: null
68
- init_fn: normal
69
- init_std: 0.02
70
- init_cutoff_factor: null
71
- norm_after: false
72
- precision: amp_bf16
73
- max_crops: 3
74
- crop_mode: resize
75
- use_col_tokens: true
76
- prompt_type: uber_model
77
- system_prompt_kind: demo_or_style
78
- message_formatting: role
79
- always_start_with_space: true
80
- multi_annotation_weighting: root_subsegments
81
- default_inference_len: 65
82
- overlap_margins:
83
- - 4
84
- - 4
85
- pad_value: 0.0
86
- image_padding_embed: pad_and_partial_pad
87
- fix_image_padding: true
88
- vit_layers:
89
- - -2
90
- - -9
91
- image_pooling_h: 2
92
- image_pooling_w: 2
93
- image_pooling_2d: attention_meanq
94
- image_projector: mlp
95
- image_feature_dropout: 0.0
96
- initializer_range: 0.02
97
- normalize_input_embeds: false
98
- use_position_ids: true
99
- head_dim: null
100
- action_tokenizer:
101
- identifier: physical-intelligence/fast
102
- tokenizer_dir: null
103
- tokenizer:
104
- identifier: Qwen/Qwen2-7B
105
- tokenizer_dir: null
106
- pad_tokenizer: true
107
- moe_num_experts: 8
108
- moe_top_k: 2
109
- moe_mlp_impl: sparse
110
- moe_log_expert_assignment: false
111
- moe_shared_expert: false
112
- moe_lbl_in_fp32: false
113
- moe_interleave: false
114
- moe_loss_weight: 0.1
115
- moe_zloss_weight: null
116
- moe_dropless: true
117
- moe_capacity_factor: 1.25
118
- action_head: flow_matching
119
- action_dim: 32
120
- fixed_action_dim: 32
121
- right_end_effector_dim: 7
122
- left_end_effector_dim: 7
123
- mobile_base_dim: 0
124
- num_actions_chunk: 50
125
- proprio_dim: 32
126
- num_diffusion_steps: 1000
127
- num_diffusion_inference_steps: 10
128
- use_proprio: true
129
- action_head_dit_hidden_size: 1152
130
- action_head_dit_depth: 28
131
- action_head_dit_num_heads: 16
132
- action_head_flow_matching_dim: 1024
133
- action_head_flow_matching_layers: 28
134
- action_head_flow_matching_heads: 8
135
- action_head_flow_matching_intermediate_size: 2048
136
- action_head_flow_matching_pvf_function: 2d_attn_mask
137
- llm_causal_attention: false
138
- action_use_left_eef: false
139
- action_use_mobile_base: false
140
- allow_resume: false
141
- ft_llm: true
142
- ft_vit: false
143
- ft_connector: false
144
- ft_embedding: lm_head
145
- lora: false
146
- use_lora: false
147
- lora_rank: 8
148
- lora_llm: false
149
- lora_vit: false
150
- lora_connector: false
151
- early_exit: false
152
- train_exit_random_layer: false
153
- optimizer:
154
- name: adamw
155
- learning_rate: 0.0001
156
- weight_decay: 0.01
157
- betas:
158
- - 0.9
159
- - 0.95
160
- eps: 1.0e-05
161
- connector_learning_rate: 2.0e-06
162
- vit_learning_rate: 2.0e-06
163
- llm_learning_rate: 5.0e-06
164
- action_head_learning_rate: 5.0e-05
165
- connector_weight_decay: 0.0
166
- vit_weight_decay: 0.0
167
- llm_weight_decay: 0.0
168
- action_head_weight_decay: 0.0
169
- connector_betas:
170
- - 0.9
171
- - 0.95
172
- vit_betas:
173
- - 0.9
174
- - 0.95
175
- llm_betas:
176
- - 0.9
177
- - 0.95
178
- action_head_betas:
179
- - 0.9
180
- - 0.95
181
- connector_eps: 1.0e-06
182
- vit_eps: 1.0e-06
183
- llm_eps: 1.0e-06
184
- action_head_eps: 1.0e-06
185
- metrics_log_interval: 50
186
- scheduler:
187
- name: multimodal
188
- units: steps
189
- t_warmup: 100
190
- t_max: null
191
- alpha_f: 0.1
192
- connector_t_warmup: 2000
193
- vit_t_warmup: 2000
194
- action_head_t_warmup: 2000
195
- llm_t_warmup: 2000
196
- connector_freeze_steps: 1000
197
- vit_freeze_steps: 1000
198
- action_head_freeze_steps: 0
199
- llm_freeze_steps: 1000
200
- grad_clip_warmup_steps: null
201
- grad_clip_warmup_factor: null
202
- warmup_min_lr: 0.0
203
- data:
204
- dataset: vla_dataset_rc
205
- mixture: null
206
- root_size_mixture: null
207
- split: train
208
- seed: 95818
209
- shuffle_messages: false
210
- pad: to_max
211
- sequence_length: 600
212
- shuffle: true
213
- for_inference: false
214
- multi_modal: torch
215
- num_workers: 4
216
- drop_last: true
217
- pin_memory: true
218
- prefetch_factor: null
219
- persistent_workers: false
220
- timeout: 0
221
- rlds_dataset_name: libero_4_task_suites_no_noops
222
- rlds_data_root_dir: /vast/users/xiaodan/zhangkaidong/datasets/rc_datasets/rc_aloha_stack_bowls
223
- use_wrist_image: true
224
- use_proprio: true
225
- rlds_shuffle_buffer_size: 100000
226
- rlds_traj_threads: 2
227
- rlds_read_threads: 2
228
- lerobot_episode_index_start: null
229
- lerobot_episode_index_end: null
230
- restore_dataloader: true
231
- fast_forward_batches: null
232
- evaluators:
233
- - label: val
234
- data:
235
- dataset: vla_dataset_rc
236
- mixture: null
237
- root_size_mixture: null
238
- split: validation
239
- seed: null
240
- shuffle_messages: false
241
- pad: to_max
242
- sequence_length: 600
243
- shuffle: false
244
- for_inference: false
245
- multi_modal: torch
246
- num_workers: 0
247
- drop_last: true
248
- pin_memory: true
249
- prefetch_factor: null
250
- persistent_workers: true
251
- timeout: 0
252
- rlds_dataset_name: libero_4_task_suites_no_noops
253
- rlds_data_root_dir: /mnt/data/zhangjian/dataset/Simulation/datasets--openvla--modified_libero_rlds
254
- use_wrist_image: true
255
- use_proprio: true
256
- rlds_shuffle_buffer_size: 256000
257
- rlds_traj_threads: 8
258
- rlds_read_threads: 8
259
- lerobot_episode_index_start: 353
260
- lerobot_episode_index_end: 765
261
- device_eval_batch_size: null
262
- subset_num_batches: 64
263
- max_examples: null
264
- max_new_tokens: 448
265
- mm_evaluator: null
266
- save_dir: null
267
- save_to_checkpoint_dir: false
268
- eval_name: null
269
- skip_if_metrics_cached: true
270
- eval_interval: 0
271
- inf_eval_interval: -1
272
- inf_evaluators: []
273
- save_folder: ./model/checkpoints/molmo_7b_robomind_alohas_ft_aloha_put_opener_in_drawer
274
- remote_save_folder: null
275
- canceled_check_interval: 50
276
- save_interval: 1000
277
- save_interval_unsharded: 1000
278
- save_interval_ephemeral: null
279
- save_interval_action_head: 1000
280
- save_num_checkpoints_to_keep: 1
281
- save_num_unsharded_checkpoints_to_keep: 1
282
- save_num_action_head_checkpoints_to_keep: 2
283
- save_overwrite: true
284
- force_save_unsharded: false
285
- no_pre_train_checkpoint: true
286
- initial_model_checkpoint: model/checkpoints/molmo_7b_rc_robomind_rc_aloha/step30000-unsharded-test
287
- load_model_config: null
288
- checkpoint_dir: model/checkpoints/molmo_7b_rc_robomind_rc_aloha/step30000-unsharded-test
289
- load_path: null
290
- load_path_sharded_checkpointer: null
291
- reset_optimizer_state: false
292
- reset_trainer_state: false
293
- save_dataloader_state: false
294
- reset_dataloader_state: false
295
- keep_lr_on_load: true
296
- sharded_checkpointer: torch_legacy
297
- max_duration: 500000
298
- global_train_batch_size: 64
299
- device_train_batch_size: 8
300
- device_train_microbatch_size: 8
301
- device_eval_batch_size: 4
302
- eval_subset_num_batches: -1
303
- eval_on_load: false
304
- device_inf_eval_batch_size: 16
305
- inf_eval_subset_num_batches: -1
306
- device_train_grad_accum: 1
307
- max_grad_norm: 1.0
308
- multi_component_grad_norm: true
309
- batch_divisor: global_batch
310
- max_grad_norm_ratio: null
311
- precision: amp_bf16
312
- wandb:
313
- project: rc_training
314
- entity: zhangkd3
315
- group: null
316
- name: molmo_7b_robomind_alohas_ft_aloha_put_opener_in_drawer_20260105_024701
317
- tags:
318
- - watching
319
- log_artifacts: false
320
- rank_zero_only: true
321
- log_interval: 50
322
- speed_monitor:
323
- window_size: 20
324
- gpu_flops_available: null
325
- console_log_interval: 50
326
- gen1_gc_interval: 1
327
- compile: null
328
- fsdp:
329
- use_orig_params: true
330
- sharding_strategy: FULL_SHARD
331
- wrapping_strategy: by_block_and_size
332
- precision: float
333
- hybrid_sharding_num_model_replicas: null
334
- softmax_auxiliary_loss: true
335
- softmax_auxiliary_loss_scale: 0.0001
336
- time_limit: null
337
- extra_steps_after_cancel: 10
338
- python_profiling: false
339
- torch_profiling: false
340
- stop_at: 500000
341
- stop_after: null
342
- activation_checkpointing: whole_layer
343
- fused_loss: null
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
rc_robomind_alohas_ft_put_opener_in_drawer/step_33000/model.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3f2d9645993219aadfb6a814b2fc316ac2da0e352368493eb71c68055449024
3
- size 33841822503