diffusionGPT / config.json
JorgeVanco's picture
Upload folder using huggingface_hub
8c2cc2d verified
{
"architectures": [
"ModernBertForMaskedLM"
],
"attention_bias": false,
"attention_dropout": 0.0,
"bos_token_id": 50256,
"classifier_activation": "gelu",
"classifier_bias": false,
"classifier_dropout": 0.0,
"classifier_pooling": "mean",
"cls_token_id": 50281,
"custom_pipelines": {
"text-diffusion": {
"impl": "pipeline.TextDiffusionPipeline",
"pt": [
"AutoModelForMaskedLM"
],
"tf": []
}
},
"decoder_bias": true,
"deterministic_flash_attn": false,
"dtype": "float32",
"embedding_dropout": 0.0,
"eos_token_id": 50259,
"global_attn_every_n_layers": 3,
"global_rope_theta": 160000.0,
"gradient_checkpointing": false,
"hidden_activation": "gelu",
"hidden_size": 1280,
"initializer_cutoff_factor": 2.0,
"initializer_range": 0.02,
"intermediate_size": 5120,
"layer_norm_eps": 1e-05,
"local_attention": 128,
"local_rope_theta": 10000.0,
"mask_token_id": 50258,
"max_position_embeddings": 8192,
"mlp_bias": false,
"mlp_dropout": 0.0,
"model_type": "modernbert",
"norm_bias": false,
"norm_eps": 1e-05,
"num_attention_heads": 10,
"num_hidden_layers": 20,
"pad_token_id": 50257,
"position_embedding_type": "absolute",
"repad_logits_with_grad": false,
"sep_token_id": 50282,
"seq_length": 2048,
"sparse_pred_ignore_index": -100,
"sparse_prediction": false,
"transformers_version": "4.56.2",
"use_cache": false,
"vocab_size": 50263
}