lucascamillomd commited on
Commit
2434f0d
·
verified ·
1 Parent(s): 05f9261

Upload config/maximum_lifespan.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. config/maximum_lifespan.yaml +148 -0
config/maximum_lifespan.yaml ADDED
@@ -0,0 +1,148 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ task_name: train
2
+ tags:
3
+ - finetune_mammalian_small_maximum_lifespan
4
+ - finetuning
5
+ - small
6
+ - mammalian
7
+ - maximum_lifespan
8
+ train: true
9
+ test: true
10
+ trainer_ckpt_path: null
11
+ model_ckpt_path: dependencies/models/small.ckpt
12
+ strict_load: false
13
+ seed: 42
14
+ data:
15
+ batch_size: 4
16
+ dna_llm: nucleotide-transformer-v2-500m-multi-species
17
+ max_length: 10000
18
+ sorting_strategy: sorted_chromosome
19
+ dna_context_len: 2001
20
+ num_workers: 8
21
+ pin_memory: false
22
+ _target_: cpgpt.data.cpgpt_datamodule.CpGPTDataModule
23
+ train_dir: ${paths.data_dir}/mammalian/processed/train_maximum_lifespan
24
+ val_dir: ${paths.data_dir}/mammalian/processed/val_maximum_lifespan
25
+ test_dir: ${paths.data_dir}/mammalian/processed/test_maximum_lifespan
26
+ dependencies_dir: ${paths.dependencies_dir}/mammalian
27
+ model:
28
+ optimizer:
29
+ _target_: schedulefree.AdamWScheduleFree
30
+ _partial_: true
31
+ lr: 0.0001
32
+ weight_decay: 0.01
33
+ betas:
34
+ - 0.9
35
+ - 0.95
36
+ scheduler:
37
+ _target_: torch.optim.lr_scheduler.ConstantLR
38
+ _partial_: true
39
+ factor: 1.0
40
+ total_iters: 1
41
+ net:
42
+ _target_: cpgpt.model.components.model.CpGPT
43
+ d_embedding: 128
44
+ d_hidden: 128
45
+ d_dna_embedding: 1024
46
+ n_attention_heads: 8
47
+ n_layers: 8
48
+ n_mlp_blocks: 3
49
+ dropout: 0.01
50
+ architecture: transformer
51
+ activation: swiglu
52
+ positional_encoding: rotary
53
+ sample_embedding_method: cls
54
+ use_power_norm: false
55
+ fft: false
56
+ use_condition_decoder: true
57
+ condition_size: 1
58
+ use_noise_decoder: false
59
+ mlp_block_bias: false
60
+ mlp_block_norm_type: rmsnorm
61
+ mlp_block_pre_norm: false
62
+ mlp_block_post_norm: false
63
+ transformer_block_bias: false
64
+ transformer_block_norm_type: rmsnorm
65
+ transformer_block_norm_first: true
66
+ transformer_block_dropout: 0.0
67
+ training:
68
+ generative_splits: 2
69
+ binarize_input: false
70
+ contrastive_threshold: 0.5
71
+ diffusion: false
72
+ reconstruct_mode: all
73
+ diffusion_params:
74
+ num_timesteps: 1000
75
+ loss_weights:
76
+ m_mae: 10.0
77
+ m_mae_unc: 1.0
78
+ betas_mae: 0.0
79
+ betas_kld: 0.0
80
+ betas_beta: 0.0
81
+ betas_wd: 1.0
82
+ contrastive: 1.0
83
+ sample_kld: 1.0
84
+ diffusion_mse: 0.0
85
+ condition_loss: 1.0
86
+ condition_decoder_loss: mae
87
+ _target_: cpgpt.model.cpgpt_module.CpGPTLitModule
88
+ compile: true
89
+ callbacks:
90
+ model_checkpoint:
91
+ _target_: lightning.pytorch.callbacks.ModelCheckpoint
92
+ dirpath: ${paths.output_dir}/checkpoints
93
+ filename: ${tags[0]}
94
+ monitor: val/condition_loss
95
+ verbose: false
96
+ save_last: true
97
+ save_top_k: 1
98
+ mode: min
99
+ auto_insert_metric_name: false
100
+ save_weights_only: false
101
+ every_n_train_steps: null
102
+ train_time_interval: null
103
+ every_n_epochs: null
104
+ save_on_train_epoch_end: null
105
+ model_summary:
106
+ _target_: lightning.pytorch.callbacks.RichModelSummary
107
+ max_depth: -1
108
+ rich_progress_bar:
109
+ _target_: lightning.pytorch.callbacks.RichProgressBar
110
+ logger:
111
+ wandb:
112
+ _target_: lightning.pytorch.loggers.wandb.WandbLogger
113
+ save_dir: ${paths.output_dir}
114
+ offline: false
115
+ id: null
116
+ anonymous: null
117
+ project: CpGPT
118
+ log_model: true
119
+ prefix: ''
120
+ entity: lucascamillo
121
+ group: ''
122
+ tags: ${tags}
123
+ job_type: ''
124
+ trainer:
125
+ _target_: lightning.pytorch.trainer.Trainer
126
+ default_root_dir: ${paths.output_dir}
127
+ min_steps: 1000
128
+ max_steps: 50000
129
+ accelerator: auto
130
+ devices: 1
131
+ precision: 16-mixed
132
+ val_check_interval: 1000
133
+ check_val_every_n_epoch: null
134
+ log_every_n_steps: 1
135
+ detect_anomaly: false
136
+ deterministic: false
137
+ accumulate_grad_batches: 1
138
+ paths:
139
+ root_dir: ${oc.env:PROJECT_ROOT}
140
+ data_dir: ${paths.root_dir}/data/
141
+ dependencies_dir: ${paths.root_dir}/dependencies/
142
+ log_dir: ${paths.root_dir}/logs/
143
+ output_dir: ${hydra:runtime.output_dir}
144
+ work_dir: ${hydra:runtime.cwd}
145
+ extras:
146
+ ignore_warnings: true
147
+ enforce_tags: true
148
+ print_config: true