File size: 1,444 Bytes
0989d5f 2cb3712 8a1afeb 2cb3712 0989d5f 2cb3712 0989d5f 2cb3712 0989d5f e65b9bd 8a1afeb e65b9bd b4eeae1 e65b9bd 2cb3712 8a1afeb e65b9bd 0989d5f 2cb3712 0989d5f e65b9bd 2cb3712 0989d5f 8a1afeb 0989d5f e65b9bd 8a1afeb e65b9bd b4eeae1 e65b9bd 2cb3712 e65b9bd 0989d5f e65b9bd 0989d5f 2cb3712 0989d5f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
{
"architectures": [
"PaliGemmaForConditionalGeneration"
],
"bos_token_id": 2,
"dtype": "bfloat16",
"eos_token_id": 1,
"hidden_size": 2048,
"ignore_index": -100,
"image_token_index": 257152,
"model_type": "paligemma",
"pad_token_id": 0,
"projection_dim": 2048,
"text_config": {
"attention_bias": false,
"attention_dropout": 0.0,
"dtype": "float32",
"head_dim": 256,
"hidden_act": "gelu_pytorch_tanh",
"hidden_activation": null,
"hidden_size": 16,
"initializer_range": 0.02,
"intermediate_size": 16384,
"layer_types": null,
"max_position_embeddings": 8192,
"model_type": "gemma",
"num_attention_heads": 4,
"num_hidden_layers": 2,
"num_image_tokens": 256,
"num_key_value_heads": 2,
"rms_norm_eps": 1e-06,
"rope_theta": 10000.0,
"use_cache": true,
"vocab_size": 257216
},
"transformers_version": "4.57.0.dev0",
"vision_config": {
"attention_dropout": 0.0,
"embed_dim": 32,
"hidden_act": "gelu_pytorch_tanh",
"hidden_size": 16,
"image_size": 224,
"intermediate_size": 4304,
"layer_norm_eps": 1e-06,
"model_type": "siglip_vision_model",
"num_attention_heads": 4,
"num_channels": 3,
"num_hidden_layers": 2,
"num_image_tokens": 256,
"num_key_value_heads": 2,
"patch_size": 14,
"projection_dim": 2048,
"projector_hidden_act": "gelu_fast",
"vision_use_head": false
}
}
|