andyjzhao commited on
Commit
33cb5c2
·
verified ·
1 Parent(s): afb0098

Upload hydra_cfg.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. hydra_cfg.yaml +214 -0
hydra_cfg.yaml ADDED
@@ -0,0 +1,214 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ RC_augmentation: false
2
+ _dataset_cfg_lookup:
3
+ gencode128k_basic:
4
+ hf_path: jzshared/gencode128k_basic
5
+ path: data/gencode128k_basic
6
+ type: refseq
7
+ gencode128k_debug:
8
+ hf_path: jzshared/gencode128k_debug
9
+ path: data/gencode128k_debug
10
+ type: refseq
11
+ gencode_human_12.8k:
12
+ hf_path: jzshared/gencode_human_12.8k
13
+ path: data/gencode_human_12.8k
14
+ type: refseq
15
+ gencode_human_128k:
16
+ hf_path: jzshared/gencode_human_128k
17
+ path: data/gencode_human_128k
18
+ type: refseq
19
+ hg38_128k:
20
+ hf_path: jzshared/hg38_cds_anchored_128000
21
+ path: data/hg38_cds_anchored_128000
22
+ type: refseq
23
+ hg38_12k:
24
+ hf_path: jzshared/hg38_12800
25
+ path: data/hg38_cds_anchored_len12800_mincds150_1000000samples
26
+ type: refseq
27
+ hg38_cds_4m:
28
+ hf_path: null
29
+ path: data/hg38_cds_dataset_4m_filtered
30
+ type: refseq
31
+ orca32m_cmp_seq:
32
+ eval_split: validation
33
+ hf_path: jzshared/orca32m_cmp
34
+ label_key: label_ut
35
+ mask_key: mask_ut
36
+ num_workers: 0
37
+ path: data/orca32m_cmp_seq
38
+ pin_memory: true
39
+ reference_id: hg38
40
+ sequence_format: string
41
+ sequence_key: sequence
42
+ shuffle: true
43
+ test_split: test
44
+ train_split: train
45
+ type: cmp_seq
46
+ alias: HNet_BPT989_12.8K-100B
47
+ alpha_exp: 1.0
48
+ alpha_max: 0.03
49
+ arch: hnet
50
+ batch_size: 16
51
+ bp_per_token: 989
52
+ cluster: mila
53
+ cmd: python src/scripts/train_genezip_v1.py exp=glm/stage1 data=gencode_human_12.8k
54
+ model=hnet/mamba_64m_2dc max_len=12800 batch_size=16 grad_acc_steps=2 max_train_steps=30600
55
+ eval_steps=500 num_valid_samples=3000 upload_to_hf=true wandb.project=DNAFM_v2 alpha_max=0.03
56
+ use_routing_floor=false strictness_max=0 region_info=promoter1_cds1_utr1_exon1_intron1_nig1_dig1
57
+ alias=HNet_BPT989_12.8K-100B bp_per_token=989 use_wandb=true hf_repo=jzshared/HNet_BPT989_12.8K-100B
58
+ config_path: null
59
+ data: gencode_human_12.8k
60
+ data_alias: ${.data}_${max_len}
61
+ dataset: ${_dataset_cfg_lookup[${data}]}
62
+ device: cuda
63
+ device_type: GPU
64
+ dirs:
65
+ data_cache: ${project_root}/data_cache/
66
+ data_storage: ${project_root}/data/
67
+ hydra: ${project_root}/temp/hydra/
68
+ output: ${project_root}/output/${data_alias}/${alias}/
69
+ temp: ${project_root}/temp/working_dir/${uid}/
70
+ wandb_cache: ${oc.env:WANDB_CACHE_DIR,${project_root}/temp/wandb_cache/}
71
+ epochs: 200
72
+ eval_batch_size: ${batch_size}
73
+ eval_steps: 500
74
+ grad_acc_steps: 2
75
+ hf_repo: jzshared/HNet_BPT989_12.8K-100B
76
+ hf_user: jzshared
77
+ is_distributed: true
78
+ local_rank: 0
79
+ log_every: 10
80
+ logging:
81
+ level: info
82
+ log_wandb_metric_to_stdout: true
83
+ lr: 0.001
84
+ master_port: '59371'
85
+ max_data_samples: null
86
+ max_eval_samples: ${num_valid_samples}
87
+ max_grad_norm: 2.0
88
+ max_len: 12800
89
+ max_length: ${max_len}
90
+ max_train_steps: 30600
91
+ min_routing_tokens: 8
92
+ mixed_precision: bf16
93
+ mode: Stage1
94
+ model:
95
+ arch: hnet
96
+ name: hnet_mamba_64m_2dc
97
+ model_alias: ${oc.select:model.name,UnknownModel}
98
+ model_cfg:
99
+ arch_layout:
100
+ - m2
101
+ - - m2
102
+ - - m15
103
+ - m2
104
+ - m2
105
+ attn_cfg:
106
+ num_heads:
107
+ - 8
108
+ - 8
109
+ - 12
110
+ rotary_emb_dim:
111
+ - 16
112
+ - 16
113
+ - 24
114
+ window_size:
115
+ - 511
116
+ - 511
117
+ - -1
118
+ d_intermediate:
119
+ - 0
120
+ - 0
121
+ - 2048
122
+ d_model:
123
+ - 512
124
+ - 512
125
+ - 768
126
+ min_routing_tokens: ${min_routing_tokens}
127
+ n_gpt: 1.0
128
+ r_hi: ${r_hi}
129
+ r_low: ${r_low}
130
+ r_warm_up_end: ${r_warm_up_end}
131
+ r_warm_up_start: ${r_warm_up_start}
132
+ ssm_cfg:
133
+ chunk_size: 256
134
+ d_conv: 4
135
+ d_state: 64
136
+ expand: 2
137
+ head_dim: 64
138
+ tie_embeddings: true
139
+ vocab_size: 12
140
+ name: hnet_base
141
+ num_test_samples: 0
142
+ num_train_samples: 0
143
+ num_valid_samples: 3000
144
+ pretrained_ckpt: null
145
+ private: false
146
+ project_root: ${hydra:runtime.cwd}
147
+ r_hi: 0.3
148
+ r_low: 0.0
149
+ r_warm_up_end: 750
150
+ r_warm_up_start: 200
151
+ rank: 0
152
+ reference_loss: null
153
+ region_info: promoter1_cds1_utr1_exon1_intron1_nig1_dig1
154
+ save_steps: 3000
155
+ seed: 0
156
+ source: ${dataset.type}
157
+ strictness_exp: 1.0
158
+ strictness_max: 0
159
+ tokenizer: fast
160
+ train_steps: 9999999
161
+ training:
162
+ adam_beta1: 0.9
163
+ adam_beta2: 0.95
164
+ bf16: true
165
+ dataloader_drop_last: true
166
+ dataloader_num_workers: 1
167
+ disable_tqdm: false
168
+ do_train: true
169
+ eval_steps: ${eval_steps}
170
+ eval_strategy: steps
171
+ gradient_accumulation_steps: ${grad_acc_steps}
172
+ gradient_checkpointing: false
173
+ group_by_length: false
174
+ hnet_initializer_range: 0.02
175
+ hnet_lr_multiplier: null
176
+ label_names:
177
+ - input_ids
178
+ learning_rate: ${lr}
179
+ logging_steps: ${log_every}
180
+ lr_scheduler_type: linear
181
+ max_grad_norm: ${max_grad_norm}
182
+ max_train_steps: ${max_train_steps}
183
+ num_train_epochs: ${epochs}
184
+ output_dir: ${dirs.output}
185
+ overrides: {}
186
+ per_device_eval_batch_size: ${eval_batch_size}
187
+ per_device_train_batch_size: ${batch_size}
188
+ remove_unused_columns: false
189
+ report_to: null
190
+ save_steps: ${save_steps}
191
+ save_strategy: steps
192
+ use_lr_multiplier: true
193
+ warmup_steps: 500
194
+ weight_decay: 0.1
195
+ training_alias: ${mode}_glm_s1_${region_info}_bp${bp_per_token}_aw${warmup_steps}_amax${alpha_max}_smax${strictness_max}_lr${lr}_e${epochs}_ms${max_train_steps}_maxlen${max_len}
196
+ uid: qsbiqtxx
197
+ upload_to_hf: true
198
+ use_routing_floor: false
199
+ use_wandb: true
200
+ valid_test_downsample: null
201
+ version: NA
202
+ wandb:
203
+ dir: ${dirs.wandb_cache}
204
+ entity: ${oc.select:env.vars.wandb_entity,${oc.env:WANDB_ENTITY,null}}
205
+ id: qsbiqtxx
206
+ mode: online
207
+ name: HNet_BPT989_12.8K-100B
208
+ project: DNAFM_v2
209
+ step_metric: null
210
+ tags:
211
+ - ${mode}
212
+ url: https://wandb.ai/liuxxiiiii/DNAFM_v2/runs/qsbiqtxx
213
+ warmup_steps: 0
214
+ world_size: 8