moe_expert_static / config.json
alexchen4ai's picture
Upload model
7c89214 verified
raw
history blame contribute delete
388 Bytes
{
"_name_or_path": "/workspace/qnn-cores/examples/moe_exp/modelfiles/moe_expert_split",
"architectures": [
"Qwen3MoeModelStatic"
],
"dtype": "float32",
"hidden_act": "silu",
"hidden_size": 2048,
"initializer_range": 0.02,
"model_type": "qwen3_moe",
"moe_intermediate_size": 768,
"num_experts": 128,
"torch_dtype": "float32",
"transformers_version": "4.47.0"
}