LoFA_v0 / config.yaml
ymhao's picture
Upload config.yaml with huggingface_hub
eb12a54 verified
mixed_precision: bf16
stage1:
_target_: lofa.hyper_modulator_mask.TransformerV2
context_dim: 4096
d_model: 768
n_heads: 12
n_blocks: 8
patch_size: 96
num_type: 8
lora_rank: 32
num_depth: 30
factorized: false
qk_norm: true
use_cls: true
decode_cls: false
tau: 0.02
temperature: 0.1
out_act: softplus
stage2:
_target_: lofa.hyper_modulator_transformer.TransformerV2_stage2
context_dim: 4096
d_model: 768
n_heads: 12
n_blocks: 8
patch_size: 96
num_type: 8
lora_rank: 32
num_depth: 30
factorized: false
qk_norm: true
use_cls: true
decode_cls: false