speedrun / config.json
lhallee's picture
Upload PLM
4a27083 verified
{
"add_att_soft_cap": false,
"architectures": [
"PLM"
],
"attention_soft_cap": 64.0,
"dtype": "bfloat16",
"expansion_ratio": 2.0,
"hidden_size": 768,
"masked_diffusion": true,
"mlm": true,
"num_attention_heads": 6,
"num_hidden_layers": 24,
"sliding_window_size": 2048,
"soft_logit_cap": 32.0,
"tie_embeddings": false,
"token_dropout": true,
"transformers_version": "4.57.6",
"unet": true,
"vocab_size": 33
}