File size: 1,410 Bytes
0989d5f
 
 
 
2cb3712
 
0989d5f
2cb3712
 
0989d5f
2cb3712
 
0989d5f
e65b9bd
 
 
 
 
b4eeae1
e65b9bd
2cb3712
e65b9bd
0989d5f
 
 
2cb3712
0989d5f
e65b9bd
 
2cb3712
e65b9bd
2cb3712
0989d5f
2cb3712
bf493e6
0989d5f
e65b9bd
 
b4eeae1
e65b9bd
2cb3712
e65b9bd
0989d5f
 
e65b9bd
0989d5f
2cb3712
 
 
 
 
 
0989d5f
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
{
  "architectures": [
    "PaliGemmaForConditionalGeneration"
  ],
  "bos_token_id": 2,
  "eos_token_id": 1,
  "hidden_size": 2048,
  "ignore_index": -100,
  "image_token_index": 257152,
  "model_type": "paligemma",
  "pad_token_id": 0,
  "projection_dim": 2048,
  "text_config": {
    "attention_bias": false,
    "attention_dropout": 0.0,
    "head_dim": 256,
    "hidden_act": "gelu_pytorch_tanh",
    "hidden_activation": null,
    "hidden_size": 16,
    "initializer_range": 0.02,
    "intermediate_size": 16384,
    "max_position_embeddings": 8192,
    "model_type": "gemma",
    "num_attention_heads": 4,
    "num_hidden_layers": 2,
    "num_image_tokens": 256,
    "num_key_value_heads": 2,
    "rms_norm_eps": 1e-06,
    "rope_theta": 10000.0,
    "torch_dtype": "float32",
    "use_cache": true,
    "vocab_size": 257216
  },
  "torch_dtype": "bfloat16",
  "transformers_version": "4.56.0.dev0",
  "vision_config": {
    "attention_dropout": 0.0,
    "hidden_act": "gelu_pytorch_tanh",
    "hidden_size": 16,
    "image_size": 224,
    "intermediate_size": 4304,
    "layer_norm_eps": 1e-06,
    "model_type": "siglip_vision_model",
    "num_attention_heads": 4,
    "num_channels": 3,
    "num_hidden_layers": 2,
    "num_image_tokens": 256,
    "num_key_value_heads": 2,
    "patch_size": 14,
    "projection_dim": 2048,
    "projector_hidden_act": "gelu_fast",
    "vision_use_head": false
  }
}