File size: 351 Bytes
3df123f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
{
  "variant": "gptoss_phase1",
  "model_type": "adamba-moe",
  "architecture": "HybridMoEGPT (Attention + MoE + Mamba)",
  "base_model": "gpt-oss-20b",
  "parameters": "21.9B",
  "n_embd": 2880,
  "features": [
    "mamba_integration",
    "moe_32experts"
  ],
  "n_layers": 36,
  "vocab_size": 201088,
  "num_experts": 32,
  "experts_per_token": 4
}