andyjzhao commited on
Commit
cdd07f9
·
verified ·
1 Parent(s): 9322e4c

Upload hydra_cfg.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. hydra_cfg.yaml +321 -0
hydra_cfg.yaml ADDED
@@ -0,0 +1,321 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ RC_augmentation: false
2
+ _dataset_cfg_lookup:
3
+ dlb_cmp_gm12878:
4
+ eval_split: validation
5
+ hf_path: jzshared/dlb_cmp_gm12878
6
+ label_key: label_ut
7
+ mask_key: mask_ut
8
+ num_workers: 0
9
+ path: data/dlb_cmp_gm12878
10
+ pin_memory: true
11
+ reference_id: hg38
12
+ sequence_format: string
13
+ sequence_key: sequence
14
+ shuffle: true
15
+ test_split: test
16
+ train_split: train
17
+ type: cmp_seq
18
+ dlb_cmp_h1hesc:
19
+ eval_split: validation
20
+ hf_path: jzshared/dlb_cmp_h1hesc
21
+ label_key: label_ut
22
+ mask_key: mask_ut
23
+ num_workers: 0
24
+ path: data/dlb_cmp_h1hesc
25
+ pin_memory: true
26
+ reference_id: hg38
27
+ sequence_format: string
28
+ sequence_key: sequence
29
+ shuffle: true
30
+ test_split: test
31
+ train_split: train
32
+ type: cmp_seq
33
+ dlb_cmp_hct116:
34
+ eval_split: validation
35
+ hf_path: jzshared/dlb_cmp_hct116
36
+ label_key: label_ut
37
+ mask_key: mask_ut
38
+ num_workers: 0
39
+ path: data/dlb_cmp_hct116
40
+ pin_memory: true
41
+ reference_id: hg38
42
+ sequence_format: string
43
+ sequence_key: sequence
44
+ shuffle: true
45
+ test_split: test
46
+ train_split: train
47
+ type: cmp_seq
48
+ dlb_cmp_hff:
49
+ eval_split: validation
50
+ hf_path: jzshared/dlb_cmp_hff
51
+ label_key: label_ut
52
+ mask_key: mask_ut
53
+ num_workers: 0
54
+ path: data/dlb_cmp_hff
55
+ pin_memory: true
56
+ reference_id: hg38
57
+ sequence_format: string
58
+ sequence_key: sequence
59
+ shuffle: true
60
+ test_split: test
61
+ train_split: train
62
+ type: cmp_seq
63
+ dlb_cmp_imr90:
64
+ eval_split: validation
65
+ hf_path: jzshared/dlb_cmp_imr90
66
+ label_key: label_ut
67
+ mask_key: mask_ut
68
+ num_workers: 0
69
+ path: data/dlb_cmp_imr90
70
+ pin_memory: true
71
+ reference_id: hg38
72
+ sequence_format: string
73
+ sequence_key: sequence
74
+ shuffle: true
75
+ test_split: test
76
+ train_split: train
77
+ type: cmp_seq
78
+ gencode128k_basic:
79
+ hf_path: jzshared/gencode128k_basic
80
+ path: data/gencode128k_basic
81
+ type: refseq
82
+ gencode128k_debug:
83
+ hf_path: jzshared/gencode128k_debug
84
+ path: data/gencode128k_debug
85
+ type: refseq
86
+ gencode_human_12.8k:
87
+ hf_path: jzshared/gencode_human_12.8k
88
+ path: data/gencode_human_12.8k
89
+ type: refseq
90
+ gencode_human_128k:
91
+ hf_path: jzshared/gencode_human_128k
92
+ path: data/gencode_human_128k
93
+ type: refseq
94
+ hg38_128k:
95
+ hf_path: jzshared/hg38_cds_anchored_128000
96
+ path: data/hg38_cds_anchored_128000
97
+ type: refseq
98
+ hg38_12k:
99
+ hf_path: jzshared/hg38_12800
100
+ path: data/hg38_cds_anchored_len12800_mincds150_1000000samples
101
+ type: refseq
102
+ hg38_cds_4m:
103
+ hf_path: null
104
+ path: data/hg38_cds_dataset_4m_filtered
105
+ type: refseq
106
+ orca32m_cmp_seq:
107
+ eval_split: validation
108
+ hf_path: jzshared/orca32m_cmp
109
+ label_key: label_ut
110
+ mask_key: mask_ut
111
+ num_workers: 0
112
+ path: data/orca32m_cmp_seq
113
+ pin_memory: true
114
+ reference_id: hg38
115
+ sequence_format: string
116
+ sequence_key: sequence
117
+ shuffle: true
118
+ test_split: test
119
+ train_split: train
120
+ type: cmp_seq
121
+ _unimportant_cfg:
122
+ fields:
123
+ - gpus
124
+ - debug
125
+ - wandb
126
+ - env
127
+ - uid
128
+ - local_rank
129
+ - is_distributed
130
+ - master_port
131
+ - device_type
132
+ - cluster
133
+ - world_size
134
+ - train_dataset
135
+ - eval_datasets
136
+ - user_cfg
137
+ - rank
138
+ - device
139
+ - hf_access_token
140
+ - hf_private
141
+ - hf_repo
142
+ - hf_user
143
+ - hf_token
144
+ - save_every
145
+ - eval_steps
146
+ - save_steps
147
+ - upload_to_hf
148
+ - logging
149
+ - log_every
150
+ - use_wandb
151
+ - project_root
152
+ - version
153
+ postfix:
154
+ - _path
155
+ - _file
156
+ - _dir
157
+ - _alias
158
+ - _prefix
159
+ prefix:
160
+ - _
161
+ alias: UNet_BPT256_12.8K
162
+ alpha_exp: 1.0
163
+ alpha_max: 0.03
164
+ arch: unet
165
+ batch_size: 32
166
+ bp_per_token: 256
167
+ cluster: mila
168
+ cmd: python src/scripts/UNet/train_unet.py exp=glm/stage1 data=gencode_human_12.8k
169
+ max_len=12800 batch_size=32 grad_acc_steps=1 max_train_steps=30600 eval_steps=500
170
+ num_valid_samples=3000 upload_to_hf=true wandb.project=DNAFM_v2 model=unet/unet_256bpt
171
+ bp_per_token=256 alias=UNet_BPT256_12.8K use_wandb=true hf_repo=jzshared/UNet_BPT256_12.8K
172
+ data: gencode_human_12.8k
173
+ data_alias: ${.data}_${max_len}
174
+ dataset: ${_dataset_cfg_lookup[${data}]}
175
+ device: cuda
176
+ device_type: GPU
177
+ dirs:
178
+ data_cache: ${project_root}/data_cache/
179
+ data_storage: ${project_root}/data/
180
+ hydra: ${project_root}/temp/hydra/
181
+ output: ${project_root}/output/${data_alias}/${alias}/
182
+ temp: ${project_root}/temp/working_dir/${uid}/
183
+ wandb_cache: ${oc.env:WANDB_CACHE_DIR,${project_root}/temp/wandb_cache/}
184
+ epochs: 200
185
+ eval_batch_size: ${batch_size}
186
+ eval_steps: 500
187
+ grad_acc_steps: 1
188
+ hf_private: false
189
+ hf_repo: jzshared/UNet_BPT256_12.8K
190
+ hf_user: jzshared
191
+ is_distributed: true
192
+ local_rank: 0
193
+ log_every: 10
194
+ logging:
195
+ level: info
196
+ log_wandb_metric_to_stdout: true
197
+ lr: 0.001
198
+ master_port: '42577'
199
+ max_data_samples: null
200
+ max_eval_samples: ${num_valid_samples}
201
+ max_grad_norm: 2.0
202
+ max_len: 12800
203
+ max_length: ${max_len}
204
+ max_train_steps: 30600
205
+ mixed_precision: bf16
206
+ mode: Formal
207
+ model:
208
+ arch: unet
209
+ name: unet_256bpt
210
+ model_alias: ${oc.select:model.name,UnknownModel}
211
+ model_cfg:
212
+ bottleneck:
213
+ arch: m
214
+ d_intermediate: 0
215
+ n_layers: 8
216
+ ssm_cfg:
217
+ chunk_size: 256
218
+ d_conv: 4
219
+ d_state: 64
220
+ expand: 2
221
+ head_dim: 64
222
+ channels:
223
+ - 256
224
+ - 256
225
+ - 384
226
+ - 512
227
+ - 512
228
+ - 768
229
+ - 768
230
+ - 768
231
+ - 768
232
+ conv_blocks_per_stage: 1
233
+ conv_kernel_size: 3
234
+ downsample_stages: 8
235
+ downsample_stride: 2
236
+ norm_epsilon: 1.0e-05
237
+ pad_to_multiple_of: 256
238
+ pad_token_id: 0
239
+ residual_in_fp32: true
240
+ tie_embeddings: true
241
+ upsample_mode: nearest
242
+ vocab_size: 12
243
+ name: unet_base
244
+ num_test_samples: 0
245
+ num_train_samples: 0
246
+ num_valid_samples: 3000
247
+ pretrained_ckpt: null
248
+ project_root: ${hydra:runtime.cwd}
249
+ rank: 0
250
+ reference_loss: null
251
+ region_info: promoter1_cds1_utr2_exon2_intron8_nig8_dig16
252
+ region_label_map:
253
+ cds: 1
254
+ dig: 6
255
+ exon: 3
256
+ intron: 4
257
+ nig: 5
258
+ promoter: 0
259
+ utr: 2
260
+ save_steps: 3000
261
+ seed: 0
262
+ source: ${dataset.type}
263
+ stop_steps: 0
264
+ strictness_exp: 1.0
265
+ strictness_max: 1.0
266
+ tokenizer: fast
267
+ train_steps: 9999999
268
+ training:
269
+ adam_beta1: 0.9
270
+ adam_beta2: 0.95
271
+ bf16: true
272
+ dataloader_drop_last: true
273
+ dataloader_num_workers: 1
274
+ disable_tqdm: false
275
+ do_train: true
276
+ eval_steps: ${eval_steps}
277
+ eval_strategy: steps
278
+ gradient_accumulation_steps: ${grad_acc_steps}
279
+ gradient_checkpointing: false
280
+ group_by_length: false
281
+ hnet_initializer_range: 0.02
282
+ hnet_lr_multiplier: null
283
+ label_names:
284
+ - input_ids
285
+ learning_rate: ${lr}
286
+ logging_steps: ${log_every}
287
+ lr_scheduler_type: linear
288
+ max_grad_norm: ${max_grad_norm}
289
+ max_train_steps: ${max_train_steps}
290
+ num_train_epochs: ${epochs}
291
+ output_dir: ${dirs.output}
292
+ overrides: {}
293
+ per_device_eval_batch_size: ${eval_batch_size}
294
+ per_device_train_batch_size: ${batch_size}
295
+ remove_unused_columns: false
296
+ report_to: null
297
+ save_steps: ${save_steps}
298
+ save_strategy: steps
299
+ use_lr_multiplier: true
300
+ warmup_steps: 500
301
+ weight_decay: 0.1
302
+ training_alias: ${mode}_glm_s1_${region_info}_bp${bp_per_token}_aw${warmup_steps}_amax${alpha_max}_smax${strictness_max}_lr${lr}_e${epochs}_ms${max_train_steps}_maxlen${max_len}
303
+ uid: wappb704
304
+ upload_to_hf: true
305
+ use_routing_ceiling: false
306
+ use_routing_floor: true
307
+ use_wandb: true
308
+ valid_test_downsample: null
309
+ version: NA
310
+ wandb:
311
+ dir: ${dirs.wandb_cache}
312
+ entity: ${oc.select:env.vars.wandb_entity,${oc.env:WANDB_ENTITY,null}}
313
+ id: wappb704
314
+ mode: online
315
+ name: UNet_BPT256_12.8K
316
+ project: DNAFM_v2
317
+ step_metric: null
318
+ tags: []
319
+ url: https://wandb.ai/jzshared/DNAFM_v2/runs/wappb704
320
+ warmup_steps: 0
321
+ world_size: 8