File size: 1,438 Bytes
0989d5f
 
 
 
2cb3712
2c7bec0
2cb3712
0989d5f
2cb3712
 
0989d5f
2cb3712
 
0989d5f
e65b9bd
 
8a1afeb
e65b9bd
 
f69c7f9
b4eeae1
e65b9bd
2cb3712
8a1afeb
e65b9bd
0989d5f
 
 
2cb3712
0989d5f
e65b9bd
 
 
2cb3712
0989d5f
f69c7f9
0989d5f
e65b9bd
07fd1ad
e65b9bd
b4eeae1
e65b9bd
2cb3712
e65b9bd
0989d5f
 
e65b9bd
0989d5f
2cb3712
 
 
 
 
 
0989d5f
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
{
  "architectures": [
    "PaliGemmaForConditionalGeneration"
  ],
  "bos_token_id": 2,
  "dtype": "float32",
  "eos_token_id": 1,
  "hidden_size": 2048,
  "ignore_index": -100,
  "image_token_index": 257152,
  "model_type": "paligemma",
  "pad_token_id": 0,
  "projection_dim": 2048,
  "text_config": {
    "attention_bias": false,
    "attention_dropout": 0.0,
    "dtype": "float32",
    "head_dim": 256,
    "hidden_act": "gelu_pytorch_tanh",
    "hidden_activation": null,
    "hidden_size": 16,
    "initializer_range": 0.02,
    "intermediate_size": 16384,
    "layer_types": null,
    "max_position_embeddings": 8192,
    "model_type": "gemma",
    "num_attention_heads": 4,
    "num_hidden_layers": 2,
    "num_image_tokens": 256,
    "num_key_value_heads": 2,
    "rms_norm_eps": 1e-06,
    "rope_theta": 10000.0,
    "use_cache": true,
    "vocab_size": 257216
  },
  "transformers_version": "4.57.3",
  "vision_config": {
    "attention_dropout": 0.0,
    "embed_dim": 64,
    "hidden_act": "gelu_pytorch_tanh",
    "hidden_size": 16,
    "image_size": 224,
    "intermediate_size": 4304,
    "layer_norm_eps": 1e-06,
    "model_type": "siglip_vision_model",
    "num_attention_heads": 4,
    "num_channels": 3,
    "num_hidden_layers": 2,
    "num_image_tokens": 256,
    "num_key_value_heads": 2,
    "patch_size": 14,
    "projection_dim": 2048,
    "projector_hidden_act": "gelu_fast",
    "vision_use_head": false
  }
}