EnricoFermi commited on
Commit
590e7ee
·
verified ·
1 Parent(s): 80206c0

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +2 -2
config.json CHANGED
@@ -56,7 +56,7 @@
56
  "mamba_ssm_dtype": "float32",
57
  "max_position_embeddings": 262144,
58
  "mlp_only_layers": [],
59
- "model_type": "qwen3_5",
60
  "mtp_num_hidden_layers": 1,
61
  "mtp_use_dedicated_embeddings": false,
62
  "num_attention_heads": 16,
@@ -80,4 +80,4 @@
80
  "transformers_version": "5.3.0",
81
  "use_cache": true,
82
  "vocab_size": 248320
83
- }
 
56
  "mamba_ssm_dtype": "float32",
57
  "max_position_embeddings": 262144,
58
  "mlp_only_layers": [],
59
+ "model_type": "qwen3_5_text",
60
  "mtp_num_hidden_layers": 1,
61
  "mtp_use_dedicated_embeddings": false,
62
  "num_attention_heads": 16,
 
80
  "transformers_version": "5.3.0",
81
  "use_cache": true,
82
  "vocab_size": 248320
83
+ }