mixed_precision: bf16 stage1: _target_: lofa.hyper_modulator_mask.TransformerV2 context_dim: 4096 d_model: 768 n_heads: 12 n_blocks: 8 patch_size: 96 num_type: 8 lora_rank: 32 num_depth: 30 factorized: false qk_norm: true use_cls: true decode_cls: false tau: 0.02 temperature: 0.1 out_act: softplus stage2: _target_: lofa.hyper_modulator_transformer.TransformerV2_stage2 context_dim: 4096 d_model: 768 n_heads: 12 n_blocks: 8 patch_size: 96 num_type: 8 lora_rank: 32 num_depth: 30 factorized: false qk_norm: true use_cls: true decode_cls: false