File size: 274 Bytes
4b2c87d
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
{
  "architectures": [
    "Gemma3ForConditionalGeneration"
  ],
  "model_type": "gemma3",
  "torch_dtype": "torch.bfloat16",
  "hidden_size": 1344,
  "num_hidden_layers": 8,
  "num_attention_heads": 2,
  "intermediate_size": 5376,
  "head_dim": 672,
  "vocab_size": 65552
}