| { | |
| "add_att_soft_cap": false, | |
| "architectures": [ | |
| "PLM" | |
| ], | |
| "attention_soft_cap": 64.0, | |
| "dtype": "bfloat16", | |
| "expansion_ratio": 2.0, | |
| "hidden_size": 768, | |
| "masked_diffusion": true, | |
| "mlm": true, | |
| "num_attention_heads": 6, | |
| "num_hidden_layers": 24, | |
| "sliding_window_size": 2048, | |
| "soft_logit_cap": 32.0, | |
| "tie_embeddings": false, | |
| "token_dropout": true, | |
| "transformers_version": "4.57.6", | |
| "unet": true, | |
| "vocab_size": 33 | |
| } | |