{ "add_att_soft_cap": false, "architectures": [ "PLM" ], "attention_soft_cap": 64.0, "dtype": "bfloat16", "expansion_ratio": 2.0, "hidden_size": 768, "masked_diffusion": true, "mlm": true, "num_attention_heads": 6, "num_hidden_layers": 24, "sliding_window_size": 2048, "soft_logit_cap": 32.0, "tie_embeddings": false, "token_dropout": true, "transformers_version": "4.57.6", "unet": true, "vocab_size": 33 }