File size: 1,410 Bytes
0989d5f 2cb3712 0989d5f 2cb3712 0989d5f 2cb3712 0989d5f e65b9bd b4eeae1 e65b9bd 2cb3712 e65b9bd 0989d5f 2cb3712 0989d5f e65b9bd 2cb3712 e65b9bd 2cb3712 0989d5f 2cb3712 bf493e6 0989d5f e65b9bd b4eeae1 e65b9bd 2cb3712 e65b9bd 0989d5f e65b9bd 0989d5f 2cb3712 0989d5f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 |
{
"architectures": [
"PaliGemmaForConditionalGeneration"
],
"bos_token_id": 2,
"eos_token_id": 1,
"hidden_size": 2048,
"ignore_index": -100,
"image_token_index": 257152,
"model_type": "paligemma",
"pad_token_id": 0,
"projection_dim": 2048,
"text_config": {
"attention_bias": false,
"attention_dropout": 0.0,
"head_dim": 256,
"hidden_act": "gelu_pytorch_tanh",
"hidden_activation": null,
"hidden_size": 16,
"initializer_range": 0.02,
"intermediate_size": 16384,
"max_position_embeddings": 8192,
"model_type": "gemma",
"num_attention_heads": 4,
"num_hidden_layers": 2,
"num_image_tokens": 256,
"num_key_value_heads": 2,
"rms_norm_eps": 1e-06,
"rope_theta": 10000.0,
"torch_dtype": "float32",
"use_cache": true,
"vocab_size": 257216
},
"torch_dtype": "bfloat16",
"transformers_version": "4.56.0.dev0",
"vision_config": {
"attention_dropout": 0.0,
"hidden_act": "gelu_pytorch_tanh",
"hidden_size": 16,
"image_size": 224,
"intermediate_size": 4304,
"layer_norm_eps": 1e-06,
"model_type": "siglip_vision_model",
"num_attention_heads": 4,
"num_channels": 3,
"num_hidden_layers": 2,
"num_image_tokens": 256,
"num_key_value_heads": 2,
"patch_size": 14,
"projection_dim": 2048,
"projector_hidden_act": "gelu_fast",
"vision_use_head": false
}
}
|