lthn commited on
Commit
bf40d0a
·
verified ·
1 Parent(s): ea0e9c0

Upload config.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +49 -11
config.json CHANGED
@@ -17,16 +17,6 @@
17
  "image_token_id": 258880,
18
  "initializer_range": 0.02,
19
  "model_type": "gemma4",
20
- "quantization": {
21
- "group_size": 64,
22
- "bits": 4,
23
- "mode": "affine"
24
- },
25
- "quantization_config": {
26
- "group_size": 64,
27
- "bits": 4,
28
- "mode": "affine"
29
- },
30
  "text_config": {
31
  "attention_bias": false,
32
  "attention_dropout": 0.0,
@@ -139,5 +129,53 @@
139
  "tie_word_embeddings": true,
140
  "transformers_version": "5.5.0.dev0",
141
  "video_token_id": 258884,
142
- "vision_soft_tokens_per_image": 280
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
143
  }
 
17
  "image_token_id": 258880,
18
  "initializer_range": 0.02,
19
  "model_type": "gemma4",
 
 
 
 
 
 
 
 
 
 
20
  "text_config": {
21
  "attention_bias": false,
22
  "attention_dropout": 0.0,
 
129
  "tie_word_embeddings": true,
130
  "transformers_version": "5.5.0.dev0",
131
  "video_token_id": 258884,
132
+ "vision_config": {
133
+ "_name_or_path": "",
134
+ "architectures": null,
135
+ "attention_bias": false,
136
+ "attention_dropout": 0.0,
137
+ "chunk_size_feed_forward": 0,
138
+ "default_output_length": 280,
139
+ "dtype": "bfloat16",
140
+ "global_head_dim": 72,
141
+ "head_dim": 72,
142
+ "hidden_activation": "gelu_pytorch_tanh",
143
+ "hidden_size": 1152,
144
+ "id2label": {
145
+ "0": "LABEL_0",
146
+ "1": "LABEL_1"
147
+ },
148
+ "initializer_range": 0.02,
149
+ "intermediate_size": 4304,
150
+ "is_encoder_decoder": false,
151
+ "label2id": {
152
+ "LABEL_0": 0,
153
+ "LABEL_1": 1
154
+ },
155
+ "max_position_embeddings": 131072,
156
+ "model_type": "gemma4_vision",
157
+ "num_attention_heads": 16,
158
+ "num_hidden_layers": 27,
159
+ "num_key_value_heads": 16,
160
+ "output_attentions": false,
161
+ "output_hidden_states": false,
162
+ "patch_size": 16,
163
+ "pooling_kernel_size": 3,
164
+ "position_embedding_size": 10240,
165
+ "problem_type": null,
166
+ "return_dict": true,
167
+ "rms_norm_eps": 1e-06,
168
+ "rope_parameters": {
169
+ "rope_theta": 100.0,
170
+ "rope_type": "default"
171
+ },
172
+ "standardize": true,
173
+ "use_clipped_linears": false
174
+ },
175
+ "vision_soft_tokens_per_image": 280,
176
+ "quantization_config": {
177
+ "bits": 4,
178
+ "group_size": 64,
179
+ "mode": "affine"
180
+ }
181
  }