cpgpt-models / config /age_cot.yaml
lucascamillomd's picture
Upload config/age_cot.yaml with huggingface_hub
7f0621d verified
task_name: train
tags:
- finetune_altumage_small_age
- finetuning
- small
- altumage
- age
train: true
test: true
trainer_ckpt_path: null
model_ckpt_path: dependencies/model/weights/small.ckpt
strict_load: false
seed: 42
data:
batch_size: 8
dna_llm: nucleotide-transformer-v2-500m-multi-species
max_length: 20000
sorting_strategy: sorted_chromosome
dna_context_len: 2001
num_workers: 8
pin_memory: false
_target_: cpgpt.data.cpgpt_datamodule.CpGPTDataModule
train_dir: ${paths.data_dir}/altumage/processed/train
val_dir: ${paths.data_dir}/altumage/processed/val
test_dir: ${paths.data_dir}/altumage/processed/test
dependencies_dir: ${paths.dependencies_dir}/human
model:
optimizer:
_target_: schedulefree.AdamWScheduleFree
_partial_: true
lr: 0.0001
weight_decay: 0.1
betas:
- 0.9
- 0.95
warmup_steps: ${trainer.min_steps}
scheduler:
_target_: torch.optim.lr_scheduler.ConstantLR
_partial_: true
factor: 1.0
total_iters: 1
net:
_target_: cpgpt.model.components.model.CpGPT
d_embedding: 128
d_hidden: 128
d_dna_embedding: 1024
n_attention_heads: 8
n_layers: 8
n_mlp_blocks: 3
dropout: 0.01
architecture: transformer
activation: swiglu
positional_encoding: rotary
sample_embedding_method: cls
use_power_norm: false
fft: false
use_condition_decoder: true
condition_size: 1
use_noise_decoder: false
mlp_block_bias: false
mlp_block_norm_type: rmsnorm
mlp_block_pre_norm: false
mlp_block_post_norm: false
transformer_block_bias: false
transformer_block_norm_type: rmsnorm
transformer_block_norm_first: true
transformer_block_dropout: 0.0
training:
generative_splits: 10
binarize_input: false
contrastive_threshold: 0.5
diffusion: false
reconstruct_mode: all
diffusion_params:
num_timesteps: 1000
loss_weights:
m_mae: 10.0
m_mae_unc: 1.0
betas_mae: 0.0
betas_kld: 0.0
betas_beta: 0.0
betas_wd: 1.0
contrastive: 1.0
sample_kld: 1.0
diffusion_mse: 0.0
condition_loss: 0.1
condition_decoder_loss: mae
_target_: cpgpt.model.cpgpt_module.CpGPTLitModule
compile: true
callbacks:
model_checkpoint:
_target_: lightning.pytorch.callbacks.ModelCheckpoint
dirpath: ${paths.output_dir}/checkpoints
filename: ${tags[0]}
monitor: val/condition_loss
verbose: false
save_last: true
save_top_k: 1
mode: min
auto_insert_metric_name: false
save_weights_only: false
every_n_train_steps: null
train_time_interval: null
every_n_epochs: null
save_on_train_epoch_end: null
model_summary:
_target_: lightning.pytorch.callbacks.RichModelSummary
max_depth: 2
rich_progress_bar:
_target_: lightning.pytorch.callbacks.RichProgressBar
logger:
wandb:
_target_: lightning.pytorch.loggers.wandb.WandbLogger
save_dir: ${paths.output_dir}
offline: false
id: null
anonymous: null
project: CpGPT
log_model: true
prefix: ''
entity: lucascamillo
group: ''
tags: ${tags}
job_type: ''
trainer:
_target_: lightning.pytorch.trainer.Trainer
default_root_dir: ${paths.output_dir}
min_steps: 2000
max_steps: 100000
accelerator: auto
devices: 1
precision: 16-mixed
val_check_interval: 1000
check_val_every_n_epoch: null
log_every_n_steps: 1
detect_anomaly: false
deterministic: false
accumulate_grad_batches: 1
paths:
root_dir: ${oc.env:PROJECT_ROOT}
data_dir: ${paths.root_dir}/data/
dependencies_dir: ${paths.root_dir}/dependencies/
log_dir: ${paths.root_dir}/logs/
output_dir: ${hydra:runtime.output_dir}
work_dir: ${hydra:runtime.cwd}
extras:
ignore_warnings: true
enforce_tags: true
print_config: true