RyanMullins commited on
Commit
626c007
·
verified ·
1 Parent(s): c1c4cf1

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -3,18 +3,22 @@
3
  "Gemma4ForConditionalGeneration"
4
  ],
5
  "audio_config": null,
6
- "audio_ms_per_token": 40,
7
  "audio_token_id": 258881,
 
8
  "boa_token_id": 256000,
 
9
  "boi_token_id": 255999,
 
10
  "dtype": "bfloat16",
11
  "eoa_token_id": 258883,
 
12
  "eoi_token_id": 258882,
 
13
  "image_token_id": 258880,
 
14
  "initializer_range": 0.02,
15
  "model_type": "gemma4",
16
  "text_config": {
17
- "_sliding_window_pattern": 6,
18
  "attention_bias": false,
19
  "attention_dropout": 0.0,
20
  "attention_k_eq_v": true,
@@ -29,7 +33,7 @@
29
  "head_dim": 256,
30
  "hidden_activation": "gelu_pytorch_tanh",
31
  "hidden_size": 5376,
32
- "hidden_size_per_layer_input": null,
33
  "initializer_range": 0.02,
34
  "intermediate_size": 21504,
35
  "layer_types": [
@@ -169,7 +173,7 @@
169
  "full_attention"
170
  ],
171
  "max_position_embeddings": 131072,
172
- "model_type": "",
173
  "num_attention_heads": 16,
174
  "num_hidden_layers": 27,
175
  "num_key_value_heads": 16,
 
3
  "Gemma4ForConditionalGeneration"
4
  ],
5
  "audio_config": null,
 
6
  "audio_token_id": 258881,
7
+ "audio_token_index": 258881,
8
  "boa_token_id": 256000,
9
+ "boa_token_index": 256000,
10
  "boi_token_id": 255999,
11
+ "boi_token_index": 255999,
12
  "dtype": "bfloat16",
13
  "eoa_token_id": 258883,
14
+ "eoa_token_index": 258883,
15
  "eoi_token_id": 258882,
16
+ "eoi_token_index": 258882,
17
  "image_token_id": 258880,
18
+ "image_token_index": 258880,
19
  "initializer_range": 0.02,
20
  "model_type": "gemma4",
21
  "text_config": {
 
22
  "attention_bias": false,
23
  "attention_dropout": 0.0,
24
  "attention_k_eq_v": true,
 
33
  "head_dim": 256,
34
  "hidden_activation": "gelu_pytorch_tanh",
35
  "hidden_size": 5376,
36
+ "hidden_size_per_layer_input": 0,
37
  "initializer_range": 0.02,
38
  "intermediate_size": 21504,
39
  "layer_types": [
 
173
  "full_attention"
174
  ],
175
  "max_position_embeddings": 131072,
176
+ "model_type": "gemma4_vision",
177
  "num_attention_heads": 16,
178
  "num_hidden_layers": 27,
179
  "num_key_value_heads": 16,
generation_config.json CHANGED
@@ -4,8 +4,5 @@
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
6
  "temperature": 1.0,
7
- "do_sample": true,
8
- "top_k": 64,
9
- "top_p": 0.95,
10
  "transformers_version": "5.3.0.dev0"
11
  }
 
4
  "eos_token_id": 1,
5
  "pad_token_id": 0,
6
  "temperature": 1.0,
 
 
 
7
  "transformers_version": "5.3.0.dev0"
8
  }
processor_config.json CHANGED
@@ -23,7 +23,6 @@
23
  "sampling_rate": 16000
24
  },
25
  "image_processor": {
26
- "data_format": "channels_first",
27
  "do_convert_rgb": true,
28
  "do_normalize": false,
29
  "do_rescale": true,
@@ -33,7 +32,7 @@
33
  0.5,
34
  0.5
35
  ],
36
- "image_processor_type": "Gemma4ImageProcessorFast",
37
  "image_seq_length": 280,
38
  "image_std": [
39
  0.5,
 
23
  "sampling_rate": 16000
24
  },
25
  "image_processor": {
 
26
  "do_convert_rgb": true,
27
  "do_normalize": false,
28
  "do_rescale": true,
 
32
  0.5,
33
  0.5
34
  ],
35
+ "image_processor_type": "Gemma4ImageProcessor",
36
  "image_seq_length": 280,
37
  "image_std": [
38
  0.5,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6f9cb1153f49bc7e0c148d0c1766017aed9512f45eb33afeeb71565d29c938b
3
- size 32169319
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84cce305d1e3df8c85320d42a3373a11771adfb03e2dec7bd44f5b408b165175
3
+ size 32169443