Upload folder using huggingface_hub
Browse files- config.json +4 -24
- model.safetensors +2 -2
- processor_config.json +7 -11
config.json
CHANGED
|
@@ -4,6 +4,7 @@
|
|
| 4 |
],
|
| 5 |
"audio_config": {
|
| 6 |
"_name_or_path": "",
|
|
|
|
| 7 |
"chunk_size_feed_forward": 0,
|
| 8 |
"conf_attention_chunk_size": 12,
|
| 9 |
"conf_attention_context_left": 13,
|
|
@@ -65,7 +66,6 @@
|
|
| 65 |
"sscp_conv_time_pad_bottom": null,
|
| 66 |
"sscp_conv_time_pad_top": null,
|
| 67 |
"streaming": false,
|
| 68 |
-
"tokenizer_class": null,
|
| 69 |
"use_clipped_linears": true
|
| 70 |
},
|
| 71 |
"audio_token_id": 258881,
|
|
@@ -194,24 +194,6 @@
|
|
| 194 |
"hidden_activation": "gelu_pytorch_tanh",
|
| 195 |
"hidden_size": 768,
|
| 196 |
"intermediate_size": 3072,
|
| 197 |
-
"layer_types": [
|
| 198 |
-
"full_attention",
|
| 199 |
-
"full_attention",
|
| 200 |
-
"full_attention",
|
| 201 |
-
"full_attention",
|
| 202 |
-
"full_attention",
|
| 203 |
-
"full_attention",
|
| 204 |
-
"full_attention",
|
| 205 |
-
"full_attention",
|
| 206 |
-
"full_attention",
|
| 207 |
-
"full_attention",
|
| 208 |
-
"full_attention",
|
| 209 |
-
"full_attention",
|
| 210 |
-
"full_attention",
|
| 211 |
-
"full_attention",
|
| 212 |
-
"full_attention",
|
| 213 |
-
"full_attention"
|
| 214 |
-
],
|
| 215 |
"max_position_embeddings": 131072,
|
| 216 |
"model_type": "gemma4_vision",
|
| 217 |
"num_attention_heads": 12,
|
|
@@ -222,12 +204,10 @@
|
|
| 222 |
"position_embedding_size": 10240,
|
| 223 |
"rms_norm_eps": 1e-06,
|
| 224 |
"rope_parameters": {
|
| 225 |
-
"
|
| 226 |
-
|
| 227 |
-
"rope_type": "default"
|
| 228 |
-
}
|
| 229 |
},
|
| 230 |
-
"
|
| 231 |
"use_clipped_linears": true
|
| 232 |
},
|
| 233 |
"vision_soft_tokens_per_image": 280
|
|
|
|
| 4 |
],
|
| 5 |
"audio_config": {
|
| 6 |
"_name_or_path": "",
|
| 7 |
+
"architectures": null,
|
| 8 |
"chunk_size_feed_forward": 0,
|
| 9 |
"conf_attention_chunk_size": 12,
|
| 10 |
"conf_attention_context_left": 13,
|
|
|
|
| 66 |
"sscp_conv_time_pad_bottom": null,
|
| 67 |
"sscp_conv_time_pad_top": null,
|
| 68 |
"streaming": false,
|
|
|
|
| 69 |
"use_clipped_linears": true
|
| 70 |
},
|
| 71 |
"audio_token_id": 258881,
|
|
|
|
| 194 |
"hidden_activation": "gelu_pytorch_tanh",
|
| 195 |
"hidden_size": 768,
|
| 196 |
"intermediate_size": 3072,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 197 |
"max_position_embeddings": 131072,
|
| 198 |
"model_type": "gemma4_vision",
|
| 199 |
"num_attention_heads": 12,
|
|
|
|
| 204 |
"position_embedding_size": 10240,
|
| 205 |
"rms_norm_eps": 1e-06,
|
| 206 |
"rope_parameters": {
|
| 207 |
+
"rope_theta": 100.0,
|
| 208 |
+
"rope_type": "default"
|
|
|
|
|
|
|
| 209 |
},
|
| 210 |
+
"standardize": false,
|
| 211 |
"use_clipped_linears": true
|
| 212 |
},
|
| 213 |
"vision_soft_tokens_per_image": 280
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d9397575f5910802b4570ca20bfd84164e295878e59706014bad874bdfa6cbef
|
| 3 |
+
size 15992604452
|
processor_config.json
CHANGED
|
@@ -28,26 +28,22 @@
|
|
| 28 |
"do_rescale": true,
|
| 29 |
"do_resize": true,
|
| 30 |
"image_mean": [
|
| 31 |
-
0.
|
| 32 |
-
0.
|
| 33 |
-
0.
|
| 34 |
],
|
| 35 |
"image_processor_type": "Gemma4ImageProcessor",
|
| 36 |
"image_seq_length": 280,
|
| 37 |
"image_std": [
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
],
|
| 42 |
"max_soft_tokens": 280,
|
| 43 |
"patch_size": 16,
|
| 44 |
"pooling_kernel_size": 3,
|
| 45 |
"resample": 3,
|
| 46 |
-
"rescale_factor": 0.00392156862745098
|
| 47 |
-
"size": {
|
| 48 |
-
"height": 224,
|
| 49 |
-
"width": 224
|
| 50 |
-
}
|
| 51 |
},
|
| 52 |
"image_seq_length": 280,
|
| 53 |
"processor_class": "Gemma4Processor"
|
|
|
|
| 28 |
"do_rescale": true,
|
| 29 |
"do_resize": true,
|
| 30 |
"image_mean": [
|
| 31 |
+
0.0,
|
| 32 |
+
0.0,
|
| 33 |
+
0.0
|
| 34 |
],
|
| 35 |
"image_processor_type": "Gemma4ImageProcessor",
|
| 36 |
"image_seq_length": 280,
|
| 37 |
"image_std": [
|
| 38 |
+
1.0,
|
| 39 |
+
1.0,
|
| 40 |
+
1.0
|
| 41 |
],
|
| 42 |
"max_soft_tokens": 280,
|
| 43 |
"patch_size": 16,
|
| 44 |
"pooling_kernel_size": 3,
|
| 45 |
"resample": 3,
|
| 46 |
+
"rescale_factor": 0.00392156862745098
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
},
|
| 48 |
"image_seq_length": 280,
|
| 49 |
"processor_class": "Gemma4Processor"
|