JianZhangAI commited on
Commit
f829de8
·
verified ·
1 Parent(s): 50bafb2

Upload folder using huggingface_hub

Browse files
model/Molmo-7B-D-0924_expanded_vocab_128/config.yaml ADDED
@@ -0,0 +1,803 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ run_name: multitask_train
2
+ seed: 6198
3
+ epoch: null
4
+ dry_run: false
5
+ model:
6
+ d_model: 3584
7
+ n_heads: 28
8
+ n_kv_heads: 4
9
+ qkv_bias: true
10
+ clip_qkv: null
11
+ n_layers: 28
12
+ mlp_ratio: 4
13
+ mlp_hidden_size: 37888
14
+ activation_type: swiglu
15
+ block_type: sequential
16
+ block_group_size: 1
17
+ rope: true
18
+ rope_full_precision: true
19
+ rope_theta: 1000000.0
20
+ vision_backbone:
21
+ image_model_type: openai
22
+ image_default_input_size:
23
+ - 336
24
+ - 336
25
+ image_patch_size: 14
26
+ image_pos_patch_size: 14
27
+ image_emb_dim: 1024
28
+ image_num_heads: 16
29
+ image_num_key_value_heads: 16
30
+ image_num_layers: 23
31
+ image_head_dim: 64
32
+ image_mlp_dim: 4096
33
+ image_mlp_activations: quick_gelu
34
+ image_dropout_rate: 0.0
35
+ image_num_pos: 577
36
+ image_norm_eps: 1.0e-05
37
+ attention_dropout: 0.0
38
+ residual_dropout: 0.0
39
+ initializer_range: 0.02
40
+ fsdp_wrap: false
41
+ resize_mode: default
42
+ vit_load_path: /weka/oe-training-default/mm-olmo/pretrained_image_encoders/vit-l-14-336.pt
43
+ llm_load_path: /weka/oe-training-default/mm-olmo/pretrained_llms/qwen2-7b.pt
44
+ low_cpu_fsdp: true
45
+ attention_type: sdpa
46
+ float32_attention: true
47
+ attention_dropout: 0.0
48
+ attention_layer_norm: false
49
+ residual_dropout: 0.1
50
+ response_residual_dropout: 0.0
51
+ embedding_dropout: 0.0
52
+ layer_norm_type: rms
53
+ layer_norm_with_affine: true
54
+ layer_norm_eps: 1.0e-06
55
+ attention_layer_norm_with_affine: true
56
+ max_sequence_length: 4096
57
+ max_position_embeddings: null
58
+ include_bias: false
59
+ bias_for_layer_norm: null
60
+ scale_logits: false
61
+ vocab_size: 152064
62
+ embedding_size: 152064
63
+ ff_out_size: 152192
64
+ additional_vocab_size: 128
65
+ new_embedding_init_range: 0.02
66
+ weight_tying: false
67
+ init_device: cpu
68
+ init_fn: normal
69
+ init_std: 0.02
70
+ init_cutoff_factor: null
71
+ norm_after: false
72
+ precision: amp_bf16
73
+ max_crops: 12
74
+ crop_mode: overlap-and-resize-c2
75
+ use_col_tokens: true
76
+ prompt_type: uber_model
77
+ system_prompt_kind: demo_or_style
78
+ message_formatting: role
79
+ always_start_with_space: true
80
+ multi_annotation_weighting: root_subsegments
81
+ default_inference_len: 65
82
+ overlap_margins:
83
+ - 4
84
+ - 4
85
+ pad_value: 0.0
86
+ image_padding_embed: pad_and_partial_pad
87
+ fix_image_padding: true
88
+ vit_layers:
89
+ - -2
90
+ - -9
91
+ image_pooling_h: 2
92
+ image_pooling_w: 2
93
+ image_pooling_2d: attention_meanq
94
+ image_projector: mlp
95
+ image_feature_dropout: 0.0
96
+ initializer_range: 0.02
97
+ normalize_input_embeds: false
98
+ use_position_ids: true
99
+ head_dim: null
100
+ action_tokenizer:
101
+ identifier: physical-intelligence/fast
102
+ tokenizer_dir: null
103
+ action_dim: 7
104
+ horizon: 8
105
+ tokenizer:
106
+ identifier: Qwen/Qwen2-7B
107
+ tokenizer_dir: null
108
+ pad_tokenizer: true
109
+ moe_num_experts: 8
110
+ moe_top_k: 2
111
+ moe_mlp_impl: sparse
112
+ moe_log_expert_assignment: false
113
+ moe_shared_expert: false
114
+ moe_lbl_in_fp32: false
115
+ moe_interleave: false
116
+ moe_loss_weight: 0.1
117
+ moe_zloss_weight: null
118
+ moe_dropless: true
119
+ moe_capacity_factor: 1.25
120
+ action_head: l1_regression
121
+ num_diffusion_steps: 1000
122
+ num_diffusion_inference_steps: 30
123
+ use_proprio: false
124
+ action_head_dit_hidden_size: 1024
125
+ action_head_dit_depth: 14
126
+ action_head_dit_num_heads: 16
127
+ llm_causal_attention: false
128
+ action_use_left_eef: false
129
+ action_use_mobile_base: false
130
+ ft_llm: true
131
+ ft_vit: true
132
+ ft_connector: true
133
+ ft_embedding: lm_head
134
+ optimizer:
135
+ name: adamw
136
+ learning_rate: 0.0001
137
+ weight_decay: 0.01
138
+ betas:
139
+ - 0.9
140
+ - 0.95
141
+ eps: 1.0e-05
142
+ connector_learning_rate: 5.0e-06
143
+ vit_learning_rate: 5.0e-06
144
+ llm_learning_rate: 1.0e-05
145
+ connector_weight_decay: 0.0
146
+ vit_weight_decay: 0.0
147
+ llm_weight_decay: 0.0
148
+ connector_betas:
149
+ - 0.9
150
+ - 0.95
151
+ vit_betas:
152
+ - 0.9
153
+ - 0.95
154
+ llm_betas:
155
+ - 0.9
156
+ - 0.95
157
+ connector_eps: 1.0e-06
158
+ vit_eps: 1.0e-06
159
+ llm_eps: 1.0e-06
160
+ no_decay_norm_and_bias: null
161
+ decay_norm_and_bias: false
162
+ decay_embeddings: false
163
+ metrics_log_interval: 20
164
+ scheduler:
165
+ name: multimodal
166
+ units: steps
167
+ t_warmup: 100
168
+ t_max: null
169
+ alpha_f: 0.1
170
+ connector_t_warmup: 200
171
+ vit_t_warmup: 200
172
+ llm_t_warmup: 200
173
+ grad_clip_warmup_steps: null
174
+ grad_clip_warmup_factor: null
175
+ warmup_min_lr: 0.0
176
+ data:
177
+ multi_modal: true
178
+ mixture_or_task_name: null
179
+ paths: null
180
+ datasets: null
181
+ label_mask_paths: null
182
+ pad_direction: right
183
+ generate_attention_mask: false
184
+ num_workers: 0
185
+ drop_last: true
186
+ pin_memory: false
187
+ prefetch_factor: null
188
+ persistent_workers: false
189
+ timeout: 0
190
+ seed: null
191
+ instance_filter: null
192
+ mixture:
193
+ user_qa: 3.772204620047811
194
+ cockatoo_712k_sept6: 3.161736404536387
195
+ synthetic_qa_v3_as_user_qa: 5.70612401246295
196
+ point_qa: 2.359934962952852
197
+ coco_2014_vqa_multi: 3.067124765028162
198
+ text_vqa: 1.982948627824974
199
+ okvqa: 1.011810438733119
200
+ chart_qa_weighted: 1.793272988041348
201
+ doc_qa: 2.1176584276768065
202
+ info_qa: 1.6495950748744952
203
+ ai2_diagram_v2_mix_transparent: 1.307415575130731
204
+ a_okvqa_mc: 1.3921930323623322
205
+ a_okvqa_da: 1.3921930323623322
206
+ android_control: 2.91381420538637
207
+ science_qa_img: 0.8405938747465043
208
+ tabwmp_da: 1.618754903035765
209
+ st_qa: 1.6871928539865735
210
+ tally_qa: 3.8873680368227395
211
+ clocks: 5.330044888030191
212
+ scifi_charts_qa: 5.152496802868815
213
+ scifi_table_qa: 3.251519607143135
214
+ scifi_document_qa: 4.024928615904329
215
+ scifi_diagram_qa: 1.9394919410872649
216
+ dv_qa: 1.0660089776060382
217
+ figure_qa: 1.0660089776060382
218
+ plot_qa: 1.507564353741936
219
+ pointing: 8.963869696866821
220
+ pointing_high_freq: 5.438911590336815
221
+ fast_flickr_count_qa_pointing: 3.0972187127963617
222
+ point_count: 8.963869696866821
223
+ point_count_high_freq: 5.438911590336815
224
+ fast_flickr_count_qa_point_count: 3.0972187127963617
225
+ sequence_length: 2304
226
+ shuffle: true
227
+ for_inference: false
228
+ split: train
229
+ use_memory_cache: false
230
+ num_epochs: null
231
+ shuffle_buffer_size: 200
232
+ per_node_data_loader: null
233
+ restore_dataloader: true
234
+ fast_forward_batches: null
235
+ evaluators: []
236
+ eval_interval: 2000
237
+ inf_eval_interval: 2000
238
+ inf_evaluators:
239
+ - label: chart_qa
240
+ type: multi_modal
241
+ data:
242
+ multi_modal: true
243
+ mixture_or_task_name: chart_qa
244
+ paths: null
245
+ datasets: null
246
+ label_mask_paths: null
247
+ pad_direction: right
248
+ generate_attention_mask: false
249
+ num_workers: 0
250
+ drop_last: false
251
+ pin_memory: false
252
+ prefetch_factor: null
253
+ persistent_workers: false
254
+ timeout: 0
255
+ seed: null
256
+ instance_filter: null
257
+ mixture: null
258
+ sequence_length: 1792
259
+ shuffle: false
260
+ for_inference: true
261
+ split: validation
262
+ use_memory_cache: false
263
+ num_epochs: 1
264
+ shuffle_buffer_size: 1000
265
+ per_node_data_loader: null
266
+ device_eval_batch_size: null
267
+ subset_num_batches: -1
268
+ max_new_tokens: 16
269
+ mm_evaluator:
270
+ cider: ''
271
+ num_wandb_examples: 32
272
+ ptb_tokenizer: false
273
+ save_html: 0
274
+ save_predictions: null
275
+ named_entity_eval: false
276
+ save_tokens: false
277
+ vqa_eval: scifi_relaxed_correctness,relaxed_correctness,em
278
+ n_to_log: 0
279
+ mme_eval: false
280
+ mmbench_eval: false
281
+ sugar_crepe_eval: false
282
+ pointing_eval: false
283
+ count_eval: false
284
+ point_count_eval: false
285
+ refexp_eval: false
286
+ pointing: false
287
+ android_eval: false
288
+ clock_eval: false
289
+ gpt_eval: null
290
+ save_dir: null
291
+ save_to_checkpoint_dir: false
292
+ eval_name: null
293
+ skip_if_metrics_cached: true
294
+ - label: info_qa
295
+ type: multi_modal
296
+ data:
297
+ multi_modal: true
298
+ mixture_or_task_name: info_qa
299
+ paths: null
300
+ datasets: null
301
+ label_mask_paths: null
302
+ pad_direction: right
303
+ generate_attention_mask: false
304
+ num_workers: 0
305
+ drop_last: true
306
+ pin_memory: false
307
+ prefetch_factor: null
308
+ persistent_workers: false
309
+ timeout: 0
310
+ seed: null
311
+ instance_filter: null
312
+ mixture: null
313
+ sequence_length: 1792
314
+ shuffle: false
315
+ for_inference: true
316
+ split: validation
317
+ use_memory_cache: false
318
+ num_epochs: null
319
+ shuffle_buffer_size: 200
320
+ per_node_data_loader: null
321
+ device_eval_batch_size: null
322
+ subset_num_batches: 4
323
+ max_new_tokens: 16
324
+ mm_evaluator:
325
+ cider: ''
326
+ num_wandb_examples: 32
327
+ ptb_tokenizer: false
328
+ save_html: 0
329
+ save_predictions: null
330
+ named_entity_eval: false
331
+ save_tokens: false
332
+ vqa_eval: ansl,em
333
+ n_to_log: 0
334
+ mme_eval: false
335
+ mmbench_eval: false
336
+ sugar_crepe_eval: false
337
+ pointing_eval: false
338
+ count_eval: false
339
+ point_count_eval: false
340
+ refexp_eval: false
341
+ pointing: false
342
+ android_eval: false
343
+ clock_eval: false
344
+ gpt_eval: null
345
+ save_dir: null
346
+ save_to_checkpoint_dir: false
347
+ eval_name: null
348
+ skip_if_metrics_cached: true
349
+ - label: doc_qa
350
+ type: multi_modal
351
+ data:
352
+ multi_modal: true
353
+ mixture_or_task_name: doc_qa
354
+ paths: null
355
+ datasets: null
356
+ label_mask_paths: null
357
+ pad_direction: right
358
+ generate_attention_mask: false
359
+ num_workers: 0
360
+ drop_last: true
361
+ pin_memory: false
362
+ prefetch_factor: null
363
+ persistent_workers: false
364
+ timeout: 0
365
+ seed: null
366
+ instance_filter: null
367
+ mixture: null
368
+ sequence_length: 1792
369
+ shuffle: false
370
+ for_inference: true
371
+ split: validation
372
+ use_memory_cache: false
373
+ num_epochs: null
374
+ shuffle_buffer_size: 200
375
+ per_node_data_loader: null
376
+ device_eval_batch_size: null
377
+ subset_num_batches: 4
378
+ max_new_tokens: 16
379
+ mm_evaluator:
380
+ cider: ''
381
+ num_wandb_examples: 32
382
+ ptb_tokenizer: false
383
+ save_html: 0
384
+ save_predictions: null
385
+ named_entity_eval: false
386
+ save_tokens: false
387
+ vqa_eval: ansl,em
388
+ n_to_log: 0
389
+ mme_eval: false
390
+ mmbench_eval: false
391
+ sugar_crepe_eval: false
392
+ pointing_eval: false
393
+ count_eval: false
394
+ point_count_eval: false
395
+ refexp_eval: false
396
+ pointing: false
397
+ android_eval: false
398
+ clock_eval: false
399
+ gpt_eval: null
400
+ save_dir: null
401
+ save_to_checkpoint_dir: false
402
+ eval_name: null
403
+ skip_if_metrics_cached: true
404
+ - label: a_okvqa_da
405
+ type: multi_modal
406
+ data:
407
+ multi_modal: true
408
+ mixture_or_task_name: a_okvqa_da
409
+ paths: null
410
+ datasets: null
411
+ label_mask_paths: null
412
+ pad_direction: right
413
+ generate_attention_mask: false
414
+ num_workers: 0
415
+ drop_last: false
416
+ pin_memory: false
417
+ prefetch_factor: null
418
+ persistent_workers: false
419
+ timeout: 0
420
+ seed: null
421
+ instance_filter: null
422
+ mixture: null
423
+ sequence_length: 1792
424
+ shuffle: false
425
+ for_inference: true
426
+ split: validation
427
+ use_memory_cache: false
428
+ num_epochs: 1
429
+ shuffle_buffer_size: 1000
430
+ per_node_data_loader: null
431
+ device_eval_batch_size: null
432
+ subset_num_batches: -1
433
+ max_new_tokens: 16
434
+ mm_evaluator:
435
+ cider: ''
436
+ num_wandb_examples: 32
437
+ ptb_tokenizer: false
438
+ save_html: 0
439
+ save_predictions: null
440
+ named_entity_eval: false
441
+ save_tokens: false
442
+ vqa_eval: a_okvqa_score
443
+ n_to_log: 0
444
+ mme_eval: false
445
+ mmbench_eval: false
446
+ sugar_crepe_eval: false
447
+ pointing_eval: false
448
+ count_eval: false
449
+ point_count_eval: false
450
+ refexp_eval: false
451
+ pointing: false
452
+ android_eval: false
453
+ clock_eval: false
454
+ gpt_eval: null
455
+ save_dir: null
456
+ save_to_checkpoint_dir: false
457
+ eval_name: null
458
+ skip_if_metrics_cached: true
459
+ - label: ai2_diagram
460
+ type: multi_modal
461
+ data:
462
+ multi_modal: true
463
+ mixture_or_task_name: ai2_diagram_v2_mix_transparent
464
+ paths: null
465
+ datasets: null
466
+ label_mask_paths: null
467
+ pad_direction: right
468
+ generate_attention_mask: false
469
+ num_workers: 0
470
+ drop_last: false
471
+ pin_memory: false
472
+ prefetch_factor: null
473
+ persistent_workers: false
474
+ timeout: 0
475
+ seed: null
476
+ instance_filter: null
477
+ mixture: null
478
+ sequence_length: 1792
479
+ shuffle: false
480
+ for_inference: true
481
+ split: validation
482
+ use_memory_cache: false
483
+ num_epochs: 1
484
+ shuffle_buffer_size: 1000
485
+ per_node_data_loader: null
486
+ device_eval_batch_size: null
487
+ subset_num_batches: -1
488
+ max_new_tokens: 16
489
+ mm_evaluator:
490
+ cider: ''
491
+ num_wandb_examples: 32
492
+ ptb_tokenizer: false
493
+ save_html: 0
494
+ save_predictions: null
495
+ named_entity_eval: false
496
+ save_tokens: false
497
+ vqa_eval: mc_ai2d_opaque,mc_ai2d_transparent
498
+ n_to_log: 0
499
+ mme_eval: false
500
+ mmbench_eval: false
501
+ sugar_crepe_eval: false
502
+ pointing_eval: false
503
+ count_eval: false
504
+ point_count_eval: false
505
+ refexp_eval: false
506
+ pointing: false
507
+ android_eval: false
508
+ clock_eval: false
509
+ gpt_eval: null
510
+ save_dir: null
511
+ save_to_checkpoint_dir: false
512
+ eval_name: null
513
+ skip_if_metrics_cached: true
514
+ - label: clocks
515
+ type: multi_modal
516
+ data:
517
+ multi_modal: true
518
+ mixture_or_task_name: clocks
519
+ paths: null
520
+ datasets: null
521
+ label_mask_paths: null
522
+ pad_direction: right
523
+ generate_attention_mask: false
524
+ num_workers: 0
525
+ drop_last: true
526
+ pin_memory: false
527
+ prefetch_factor: null
528
+ persistent_workers: false
529
+ timeout: 0
530
+ seed: null
531
+ instance_filter: null
532
+ mixture: null
533
+ sequence_length: 1792
534
+ shuffle: false
535
+ for_inference: true
536
+ split: validation
537
+ use_memory_cache: false
538
+ num_epochs: null
539
+ shuffle_buffer_size: 200
540
+ per_node_data_loader: null
541
+ device_eval_batch_size: null
542
+ subset_num_batches: 4
543
+ max_new_tokens: 16
544
+ mm_evaluator:
545
+ cider: ''
546
+ num_wandb_examples: 32
547
+ ptb_tokenizer: false
548
+ save_html: 0
549
+ save_predictions: null
550
+ named_entity_eval: false
551
+ save_tokens: false
552
+ vqa_eval: ''
553
+ n_to_log: 0
554
+ mme_eval: false
555
+ mmbench_eval: false
556
+ sugar_crepe_eval: false
557
+ pointing_eval: false
558
+ count_eval: false
559
+ point_count_eval: false
560
+ refexp_eval: false
561
+ pointing: false
562
+ android_eval: false
563
+ clock_eval: true
564
+ gpt_eval: null
565
+ save_dir: null
566
+ save_to_checkpoint_dir: false
567
+ eval_name: null
568
+ skip_if_metrics_cached: true
569
+ - label: android_control_ll
570
+ type: multi_modal
571
+ data:
572
+ multi_modal: true
573
+ mixture_or_task_name: android_control_ll
574
+ paths: null
575
+ datasets: null
576
+ label_mask_paths: null
577
+ pad_direction: right
578
+ generate_attention_mask: false
579
+ num_workers: 0
580
+ drop_last: false
581
+ pin_memory: false
582
+ prefetch_factor: null
583
+ persistent_workers: false
584
+ timeout: 0
585
+ seed: null
586
+ instance_filter: null
587
+ mixture: null
588
+ sequence_length: 1792
589
+ shuffle: false
590
+ for_inference: true
591
+ split: validation
592
+ use_memory_cache: false
593
+ num_epochs: 1
594
+ shuffle_buffer_size: 1000
595
+ per_node_data_loader: null
596
+ device_eval_batch_size: null
597
+ subset_num_batches: -1
598
+ max_new_tokens: 16
599
+ mm_evaluator:
600
+ cider: ''
601
+ num_wandb_examples: 32
602
+ ptb_tokenizer: false
603
+ save_html: 0
604
+ save_predictions: null
605
+ named_entity_eval: false
606
+ save_tokens: false
607
+ vqa_eval: ''
608
+ n_to_log: 0
609
+ mme_eval: false
610
+ mmbench_eval: false
611
+ sugar_crepe_eval: false
612
+ pointing_eval: false
613
+ count_eval: false
614
+ point_count_eval: false
615
+ refexp_eval: false
616
+ pointing: false
617
+ android_eval: true
618
+ clock_eval: false
619
+ gpt_eval: null
620
+ save_dir: null
621
+ save_to_checkpoint_dir: false
622
+ eval_name: null
623
+ skip_if_metrics_cached: true
624
+ - label: pointing_test
625
+ type: multi_modal
626
+ data:
627
+ multi_modal: true
628
+ mixture_or_task_name: pointing_test
629
+ paths: null
630
+ datasets: null
631
+ label_mask_paths: null
632
+ pad_direction: right
633
+ generate_attention_mask: false
634
+ num_workers: 0
635
+ drop_last: false
636
+ pin_memory: false
637
+ prefetch_factor: null
638
+ persistent_workers: false
639
+ timeout: 0
640
+ seed: null
641
+ instance_filter: null
642
+ mixture: null
643
+ sequence_length: 1792
644
+ shuffle: false
645
+ for_inference: true
646
+ split: test
647
+ use_memory_cache: false
648
+ num_epochs: 1
649
+ shuffle_buffer_size: 1000
650
+ per_node_data_loader: null
651
+ device_eval_batch_size: null
652
+ subset_num_batches: -1
653
+ max_new_tokens: 384
654
+ mm_evaluator:
655
+ cider: ''
656
+ num_wandb_examples: 32
657
+ ptb_tokenizer: false
658
+ save_html: 0
659
+ save_predictions: null
660
+ named_entity_eval: false
661
+ save_tokens: false
662
+ vqa_eval: ''
663
+ n_to_log: 0
664
+ mme_eval: false
665
+ mmbench_eval: false
666
+ sugar_crepe_eval: false
667
+ pointing_eval: false
668
+ count_eval: false
669
+ point_count_eval: false
670
+ refexp_eval: false
671
+ pointing: true
672
+ android_eval: false
673
+ clock_eval: false
674
+ gpt_eval: null
675
+ save_dir: null
676
+ save_to_checkpoint_dir: false
677
+ eval_name: null
678
+ skip_if_metrics_cached: true
679
+ - label: countbench_qa
680
+ type: multi_modal
681
+ data:
682
+ multi_modal: true
683
+ mixture_or_task_name: countbench_qa
684
+ paths: null
685
+ datasets: null
686
+ label_mask_paths: null
687
+ pad_direction: right
688
+ generate_attention_mask: false
689
+ num_workers: 0
690
+ drop_last: false
691
+ pin_memory: false
692
+ prefetch_factor: null
693
+ persistent_workers: false
694
+ timeout: 0
695
+ seed: null
696
+ instance_filter: null
697
+ mixture: null
698
+ sequence_length: 1792
699
+ shuffle: false
700
+ for_inference: true
701
+ split: huggingface
702
+ use_memory_cache: false
703
+ num_epochs: 1
704
+ shuffle_buffer_size: 1000
705
+ per_node_data_loader: null
706
+ device_eval_batch_size: null
707
+ subset_num_batches: -1
708
+ max_new_tokens: 384
709
+ mm_evaluator:
710
+ cider: ''
711
+ num_wandb_examples: 32
712
+ ptb_tokenizer: false
713
+ save_html: 0
714
+ save_predictions: null
715
+ named_entity_eval: false
716
+ save_tokens: false
717
+ vqa_eval: ''
718
+ n_to_log: 0
719
+ mme_eval: false
720
+ mmbench_eval: false
721
+ sugar_crepe_eval: false
722
+ pointing_eval: false
723
+ count_eval: false
724
+ point_count_eval: true
725
+ refexp_eval: false
726
+ pointing: false
727
+ android_eval: false
728
+ clock_eval: false
729
+ gpt_eval: null
730
+ save_dir: null
731
+ save_to_checkpoint_dir: false
732
+ eval_name: null
733
+ skip_if_metrics_cached: true
734
+ save_folder: /weka/oe-training-default/chrisc/cockatoo/models/uber-model-v11/7b-5510-3.2-synthetic
735
+ remote_save_folder: null
736
+ canceled_check_interval: 50
737
+ save_interval: 4000
738
+ save_interval_unsharded: 30000
739
+ save_interval_ephemeral: null
740
+ save_num_checkpoints_to_keep: 1
741
+ save_num_unsharded_checkpoints_to_keep: -1
742
+ save_overwrite: true
743
+ force_save_unsharded: false
744
+ no_pre_train_checkpoint: true
745
+ initial_model_checkpoint: /weka/oe-training-default/chrisc/cockatoo/models/dense-captioner-v22-qwen2/v2-lr2620/step22300-unsharded
746
+ load_model_config: null
747
+ load_path: null
748
+ load_path_sharded_checkpointer: null
749
+ reset_optimizer_state: false
750
+ reset_trainer_state: false
751
+ save_dataloader_state: false
752
+ reset_dataloader_state: false
753
+ sharded_checkpointer: torch_legacy
754
+ new_style_checkpoints: null
755
+ max_duration: 30000
756
+ global_train_batch_size: 256
757
+ device_train_batch_size: 2
758
+ device_train_microbatch_size: 2
759
+ device_eval_batch_size: 4
760
+ eval_subset_num_batches: 4
761
+ eval_on_load: false
762
+ device_inf_eval_batch_size: 4
763
+ inf_eval_subset_num_batches: -1
764
+ device_train_grad_accum: 1
765
+ max_grad_norm: 1.0
766
+ batch_divisor: global_batch
767
+ max_grad_norm_ratio: null
768
+ precision: amp_bf16
769
+ wandb:
770
+ project: cockatoo
771
+ entity: prior-ai2
772
+ group: uber-model-v11
773
+ name: 7b-5510-3.2-synthetic
774
+ tags:
775
+ - watching
776
+ log_artifacts: false
777
+ rank_zero_only: true
778
+ log_interval: 20
779
+ speed_monitor:
780
+ window_size: 20
781
+ gpu_flops_available: null
782
+ console_log_interval: 20
783
+ gen1_gc_interval: 1
784
+ compile: null
785
+ fsdp:
786
+ use_orig_params: true
787
+ sharding_strategy: FULL_SHARD
788
+ wrapping_strategy: by_block_and_size
789
+ precision: pure
790
+ hybrid_sharding_num_model_replicas: null
791
+ softmax_auxiliary_loss: true
792
+ softmax_auxiliary_loss_scale: 0.0001
793
+ time_limit: null
794
+ extra_steps_after_cancel: 10
795
+ early_stopping_factor: null
796
+ save_data_indices: true
797
+ python_profiling: false
798
+ torch_profiling: false
799
+ stop_at: 30000
800
+ stop_after: null
801
+ activation_checkpointing: one_in_two
802
+ fused_loss: null
803
+ tfds_dir: /weka/oe-training-default/mm-olmo/tensorflow_datasets
model/Molmo-7B-D-0924_expanded_vocab_128/model.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cce38c388938be800b54f6ab45188fcc91580605c3cdd397f4b483df6fd94860
3
+ size 32086178679