File size: 351 Bytes
3df123f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
{
"variant": "gptoss_phase1",
"model_type": "adamba-moe",
"architecture": "HybridMoEGPT (Attention + MoE + Mamba)",
"base_model": "gpt-oss-20b",
"parameters": "21.9B",
"n_embd": 2880,
"features": [
"mamba_integration",
"moe_32experts"
],
"n_layers": 36,
"vocab_size": 201088,
"num_experts": 32,
"experts_per_token": 4
} |