RyanMullins commited on
Commit
11d45b1
·
verified ·
1 Parent(s): 8bacfbc

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -3,22 +3,26 @@
3
  "Gemma4ForConditionalGeneration"
4
  ],
5
  "audio_config": null,
6
- "audio_ms_per_token": 40,
7
  "audio_token_id": 258881,
 
8
  "boa_token_id": 256000,
 
9
  "boi_token_id": 255999,
 
10
  "dtype": "bfloat16",
11
  "eoa_token_id": 258883,
 
12
  "eoi_token_id": 258882,
 
13
  "eos_token_id": [
14
  1,
15
  106
16
  ],
17
  "image_token_id": 258880,
 
18
  "initializer_range": 0.02,
19
  "model_type": "gemma4",
20
  "text_config": {
21
- "_sliding_window_pattern": 6,
22
  "attention_bias": false,
23
  "attention_dropout": 0.0,
24
  "attention_k_eq_v": true,
@@ -33,7 +37,7 @@
33
  "head_dim": 256,
34
  "hidden_activation": "gelu_pytorch_tanh",
35
  "hidden_size": 5376,
36
- "hidden_size_per_layer_input": null,
37
  "initializer_range": 0.02,
38
  "intermediate_size": 21504,
39
  "layer_types": [
@@ -173,7 +177,7 @@
173
  "full_attention"
174
  ],
175
  "max_position_embeddings": 131072,
176
- "model_type": "",
177
  "num_attention_heads": 16,
178
  "num_hidden_layers": 27,
179
  "num_key_value_heads": 16,
 
3
  "Gemma4ForConditionalGeneration"
4
  ],
5
  "audio_config": null,
 
6
  "audio_token_id": 258881,
7
+ "audio_token_index": 258881,
8
  "boa_token_id": 256000,
9
+ "boa_token_index": 256000,
10
  "boi_token_id": 255999,
11
+ "boi_token_index": 255999,
12
  "dtype": "bfloat16",
13
  "eoa_token_id": 258883,
14
+ "eoa_token_index": 258883,
15
  "eoi_token_id": 258882,
16
+ "eoi_token_index": 258882,
17
  "eos_token_id": [
18
  1,
19
  106
20
  ],
21
  "image_token_id": 258880,
22
+ "image_token_index": 258880,
23
  "initializer_range": 0.02,
24
  "model_type": "gemma4",
25
  "text_config": {
 
26
  "attention_bias": false,
27
  "attention_dropout": 0.0,
28
  "attention_k_eq_v": true,
 
37
  "head_dim": 256,
38
  "hidden_activation": "gelu_pytorch_tanh",
39
  "hidden_size": 5376,
40
+ "hidden_size_per_layer_input": 0,
41
  "initializer_range": 0.02,
42
  "intermediate_size": 21504,
43
  "layer_types": [
 
177
  "full_attention"
178
  ],
179
  "max_position_embeddings": 131072,
180
+ "model_type": "gemma4_vision",
181
  "num_attention_heads": 16,
182
  "num_hidden_layers": 27,
183
  "num_key_value_heads": 16,
generation_config.json CHANGED
@@ -8,8 +8,5 @@
8
  ],
9
  "pad_token_id": 0,
10
  "temperature": 1.0,
11
- "do_sample": true,
12
- "top_k": 64,
13
- "top_p": 0.95,
14
  "transformers_version": "5.3.0.dev0"
15
  }
 
8
  ],
9
  "pad_token_id": 0,
10
  "temperature": 1.0,
 
 
 
11
  "transformers_version": "5.3.0.dev0"
12
  }
processor_config.json CHANGED
@@ -23,7 +23,6 @@
23
  "sampling_rate": 16000
24
  },
25
  "image_processor": {
26
- "data_format": "channels_first",
27
  "do_convert_rgb": true,
28
  "do_normalize": false,
29
  "do_rescale": true,
@@ -33,7 +32,7 @@
33
  0.5,
34
  0.5
35
  ],
36
- "image_processor_type": "Gemma4ImageProcessorFast",
37
  "image_seq_length": 280,
38
  "image_std": [
39
  0.5,
 
23
  "sampling_rate": 16000
24
  },
25
  "image_processor": {
 
26
  "do_convert_rgb": true,
27
  "do_normalize": false,
28
  "do_rescale": true,
 
32
  0.5,
33
  0.5
34
  ],
35
+ "image_processor_type": "Gemma4ImageProcessor",
36
  "image_seq_length": 280,
37
  "image_std": [
38
  0.5,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6f9cb1153f49bc7e0c148d0c1766017aed9512f45eb33afeeb71565d29c938b
3
- size 32169319
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84cce305d1e3df8c85320d42a3373a11771adfb03e2dec7bd44f5b408b165175
3
+ size 32169443