ykorkmaz commited on
Commit
affdc8f
·
verified ·
1 Parent(s): 5037062

Upload RFM model

Browse files
Files changed (1) hide show
  1. config.yaml +99 -0
config.yaml ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ dataloader_num_workers: 0
3
+ dataloader_pin_memory: false
4
+ dataset_preference_ratio: 0.7
5
+ dataset_type: default
6
+ eval_datasets:
7
+ - abraranwar/libero_rfm
8
+ - ykorkmaz/libero_failure_rfm
9
+ eval_subset_size: 500
10
+ eval_subsets:
11
+ - - libero256_10
12
+ - - libero_10_failure
13
+ force_reprocess: false
14
+ fps: 10
15
+ max_frames: 16
16
+ max_frames_for_preprocessing: 64
17
+ model_type: default
18
+ num_bins: 10
19
+ num_proc: 1
20
+ preference_ratio: 1.0
21
+ preference_strategy_ratio:
22
+ - 0.8
23
+ - 0.1
24
+ - 0.1
25
+ - 0.0
26
+ progress_ratio: 0.5
27
+ resized_height: 128
28
+ resized_width: 128
29
+ rewind_lengths: null
30
+ samples_per_trajectory: 1
31
+ seed: 42
32
+ shuffle: true
33
+ train_datasets:
34
+ - abraranwar/libero_rfm
35
+ train_subsets:
36
+ - - libero256_90
37
+ video_frame_sampling: uniform
38
+ debug: false
39
+ logging:
40
+ print_trainable_parameters: true
41
+ save_model: true
42
+ save_processor: true
43
+ use_wandb: true
44
+ wandb_entity: ykorkmaz
45
+ wandb_project: rfm
46
+ wandb_run_name: rfm_no_failure
47
+ mode: train
48
+ model:
49
+ base_model_id: Qwen/Qwen2.5-VL-3B-Instruct
50
+ torch_dtype: bfloat16
51
+ train_language_model: false
52
+ train_preference_head: true
53
+ train_progress_head: true
54
+ train_similarity_head: false
55
+ train_value_head: true
56
+ train_vision_encoder: true
57
+ trust_remote_code: true
58
+ peft:
59
+ bias: none
60
+ lora_alpha: 64
61
+ lora_dropout: 0.05
62
+ r: 32
63
+ target_modules:
64
+ - q_proj
65
+ - k_proj
66
+ - v_proj
67
+ - o_proj
68
+ - gate_proj
69
+ - up_proj
70
+ - down_proj
71
+ use_peft: false
72
+ training:
73
+ beta: 0.1
74
+ bf16: true
75
+ ddp_bucket_cap_mb: 25
76
+ ddp_find_unused_parameters: true
77
+ do_eval: true
78
+ eval_steps: 50
79
+ evaluation_strategy: steps
80
+ fp16: false
81
+ gradient_accumulation_steps: 1
82
+ gradient_checkpointing: true
83
+ learning_rate: 2.0e-05
84
+ logging_steps: 1
85
+ lr_scheduler_type: cosine
86
+ max_seq_length: 1024
87
+ max_steps: 5000
88
+ num_gpus: 2
89
+ num_train_epochs: -1
90
+ output_dir: /scr/ykorkmaz/rfm/logs/rfm_no_failure
91
+ per_device_eval_batch_size: 8
92
+ per_device_train_batch_size: 8
93
+ prediction_loss_only: true
94
+ remove_unused_columns: false
95
+ resume_from_checkpoint: null
96
+ save_steps: 200
97
+ save_strategy: steps
98
+ warmup_ratio: 0.1
99
+ warmup_steps: 0