File size: 368 Bytes
5101044 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 | {
"head_dim": 128,
"hidden_size": 5376,
"intermediate_size": 21504,
"model_type": "gemma3_text",
"num_attention_heads": 32,
"num_hidden_layers": 62,
"num_key_value_heads": 16,
"query_pre_attn_scalar": 168,
"rope_scaling": {
"factor": 8.0,
"rope_type": "linear"
},
"sliding_window": 1024,
"architectures": [
"Gemma3ForCausalLM"
]
} |