File size: 1,438 Bytes
0989d5f 2cb3712 2c7bec0 2cb3712 0989d5f 2cb3712 0989d5f 2cb3712 0989d5f e65b9bd 8a1afeb e65b9bd f69c7f9 b4eeae1 e65b9bd 2cb3712 8a1afeb e65b9bd 0989d5f 2cb3712 0989d5f e65b9bd 2cb3712 0989d5f f69c7f9 0989d5f e65b9bd 07fd1ad e65b9bd b4eeae1 e65b9bd 2cb3712 e65b9bd 0989d5f e65b9bd 0989d5f 2cb3712 0989d5f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 |
{
"architectures": [
"PaliGemmaForConditionalGeneration"
],
"bos_token_id": 2,
"dtype": "float32",
"eos_token_id": 1,
"hidden_size": 2048,
"ignore_index": -100,
"image_token_index": 257152,
"model_type": "paligemma",
"pad_token_id": 0,
"projection_dim": 2048,
"text_config": {
"attention_bias": false,
"attention_dropout": 0.0,
"dtype": "float32",
"head_dim": 256,
"hidden_act": "gelu_pytorch_tanh",
"hidden_activation": null,
"hidden_size": 16,
"initializer_range": 0.02,
"intermediate_size": 16384,
"layer_types": null,
"max_position_embeddings": 8192,
"model_type": "gemma",
"num_attention_heads": 4,
"num_hidden_layers": 2,
"num_image_tokens": 256,
"num_key_value_heads": 2,
"rms_norm_eps": 1e-06,
"rope_theta": 10000.0,
"use_cache": true,
"vocab_size": 257216
},
"transformers_version": "4.57.3",
"vision_config": {
"attention_dropout": 0.0,
"embed_dim": 64,
"hidden_act": "gelu_pytorch_tanh",
"hidden_size": 16,
"image_size": 224,
"intermediate_size": 4304,
"layer_norm_eps": 1e-06,
"model_type": "siglip_vision_model",
"num_attention_heads": 4,
"num_channels": 3,
"num_hidden_layers": 2,
"num_image_tokens": 256,
"num_key_value_heads": 2,
"patch_size": 14,
"projection_dim": 2048,
"projector_hidden_act": "gelu_fast",
"vision_use_head": false
}
}
|