040 / transformer_ema /config.json
open1986's picture
v001
3f62f9e verified
raw
history blame contribute delete
725 Bytes
{
"_class_name": "SapiensImageGenTransformer2DModel",
"_diffusers_version": "0.32.0.dev0",
"_name_or_path": "./sapiens_imagegen-0_5b",
"attention_bias": false,
"attention_head_dim": 32,
"caption_channels": 2304,
"cross_attention_dim": 2240,
"cross_attention_head_dim": 112,
"decay": 0.995,
"dropout": 0.0,
"in_channels": 32,
"inv_gamma": 1.0,
"min_decay": 0.0,
"mlp_ratio": 2.5,
"norm_elementwise_affine": false,
"norm_eps": 1e-06,
"num_attention_heads": 70,
"num_cross_attention_heads": 20,
"num_layers": 20,
"optimization_step": 10000,
"out_channels": 32,
"patch_size": 1,
"power": 0.6666666666666666,
"sample_size": 32,
"update_after_step": 0,
"use_ema_warmup": false
}