rippertnt commited on
Commit
29bbf81
·
verified ·
1 Parent(s): 0fbb721

Upload 10 files

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<image_soft_token>": 262144
3
+ }
config.json ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Gemma3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attn_logit_softcapping": null,
8
+ "bos_token_id": 2,
9
+ "cache_implementation": "hybrid",
10
+ "eos_token_id": 1,
11
+ "final_logit_softcapping": null,
12
+ "head_dim": 256,
13
+ "hidden_activation": "gelu_pytorch_tanh",
14
+ "hidden_size": 2560,
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 10240,
17
+ "max_position_embeddings": 131072,
18
+ "model_type": "gemma3_text",
19
+ "num_attention_heads": 8,
20
+ "num_hidden_layers": 34,
21
+ "num_key_value_heads": 4,
22
+ "pad_token_id": 0,
23
+ "quantization_config": {
24
+ "bits": 4,
25
+ "checkpoint_format": "gptq",
26
+ "desc_act": true,
27
+ "group_size": 128,
28
+ "lm_head": false,
29
+ "meta": {
30
+ "damp_auto_increment": 0.0025,
31
+ "damp_percent": 0.01,
32
+ "mse": 0.0,
33
+ "quantizer": [
34
+ "gptqmodel:2.1.0"
35
+ ],
36
+ "static_groups": false,
37
+ "true_sequential": true,
38
+ "uri": "https://github.com/modelcloud/gptqmodel"
39
+ },
40
+ "pack_dtype": "int32",
41
+ "quant_method": "gptq",
42
+ "sym": true
43
+ },
44
+ "query_pre_attn_scalar": 256,
45
+ "rms_norm_eps": 1e-06,
46
+ "rope_local_base_freq": 10000.0,
47
+ "rope_scaling": {
48
+ "factor": 8.0,
49
+ "rope_type": "linear"
50
+ },
51
+ "rope_theta": 1000000.0,
52
+ "sliding_window": 1024,
53
+ "sliding_window_pattern": 6,
54
+ "torch_dtype": "bfloat16",
55
+ "transformers_version": "4.51.0.dev0",
56
+ "use_cache": true,
57
+ "vocab_size": 262208
58
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 2,
4
+ "cache_implementation": "hybrid",
5
+ "eos_token_id": [
6
+ 1,
7
+ 106
8
+ ],
9
+ "pad_token_id": 0,
10
+ "transformers_version": "4.51.0.dev0"
11
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb090ed2366e1f66be19be369c9970ddf046dd4d71abf5f7873ef7dac22dcadd
3
+ size 3013773592
quant_log.csv ADDED
@@ -0,0 +1,239 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ layer,module,loss,damp,time
2
+ 0,self_attn.k_proj,198.69464,0.01000,1.094
3
+ 0,self_attn.v_proj,195.76051,0.01000,0.776
4
+ 0,self_attn.q_proj,243.30057,0.01000,0.792
5
+ 0,self_attn.o_proj,36.38606,0.01000,0.626
6
+ 0,mlp.up_proj,13.06959,0.01000,0.789
7
+ 0,mlp.gate_proj,13.87642,0.01000,0.792
8
+ 0,mlp.down_proj,0.57033,0.01000,3.270
9
+ 1,self_attn.k_proj,11.76147,0.01000,0.767
10
+ 1,self_attn.v_proj,11.56775,0.01000,0.765
11
+ 1,self_attn.q_proj,3.06964,0.01000,0.768
12
+ 1,self_attn.o_proj,1.91879,0.01000,0.607
13
+ 1,mlp.up_proj,16.77610,0.01000,0.773
14
+ 1,mlp.gate_proj,18.79979,0.01000,0.773
15
+ 1,mlp.down_proj,0.10125,0.01000,3.253
16
+ 2,self_attn.k_proj,13.69234,0.01000,0.760
17
+ 2,self_attn.v_proj,13.41772,0.01000,0.757
18
+ 2,self_attn.q_proj,6.14136,0.01000,0.770
19
+ 2,self_attn.o_proj,2.78490,0.01000,0.609
20
+ 2,mlp.up_proj,18.90626,0.01000,0.777
21
+ 2,mlp.gate_proj,19.42534,0.01000,0.777
22
+ 2,mlp.down_proj,0.20656,0.01000,3.240
23
+ 3,self_attn.k_proj,12.26875,0.01000,0.771
24
+ 3,self_attn.v_proj,11.49991,0.01000,0.767
25
+ 3,self_attn.q_proj,3.87051,0.01000,0.768
26
+ 3,self_attn.o_proj,6.08537,0.01000,0.611
27
+ 3,mlp.up_proj,27.04681,0.01000,0.781
28
+ 3,mlp.gate_proj,27.28091,0.01000,0.783
29
+ 3,mlp.down_proj,0.71027,0.01000,3.261
30
+ 4,self_attn.k_proj,13.06411,0.01000,0.761
31
+ 4,self_attn.v_proj,10.97907,0.01000,0.767
32
+ 4,self_attn.q_proj,8.22816,0.01000,0.767
33
+ 4,self_attn.o_proj,2.14936,0.01000,0.615
34
+ 4,mlp.up_proj,13.32326,0.01000,0.779
35
+ 4,mlp.gate_proj,13.68868,0.01000,0.777
36
+ 4,mlp.down_proj,0.13622,0.01000,3.269
37
+ 5,self_attn.k_proj,10.77550,0.01000,0.769
38
+ 5,self_attn.v_proj,5.25996,0.01000,0.766
39
+ 5,self_attn.q_proj,1.15328,0.01000,0.770
40
+ 5,self_attn.o_proj,0.36760,0.01000,0.620
41
+ 5,mlp.up_proj,14.03705,0.01000,0.783
42
+ 5,mlp.gate_proj,14.03240,0.01000,0.777
43
+ 5,mlp.down_proj,0.09100,0.01000,3.232
44
+ 6,self_attn.k_proj,7.06065,0.01000,0.757
45
+ 6,self_attn.v_proj,5.88580,0.01000,0.757
46
+ 6,self_attn.q_proj,2.05951,0.01000,0.763
47
+ 6,self_attn.o_proj,0.54376,0.01000,0.607
48
+ 6,mlp.up_proj,13.85581,0.01000,0.775
49
+ 6,mlp.gate_proj,13.96360,0.01000,0.774
50
+ 6,mlp.down_proj,0.20618,0.01000,3.237
51
+ 7,self_attn.k_proj,3.64678,0.01000,0.758
52
+ 7,self_attn.v_proj,3.37626,0.01000,0.758
53
+ 7,self_attn.q_proj,6.33875,0.01000,0.761
54
+ 7,self_attn.o_proj,1.19201,0.01000,0.608
55
+ 7,mlp.up_proj,9.08091,0.01000,0.777
56
+ 7,mlp.gate_proj,9.16898,0.01000,0.773
57
+ 7,mlp.down_proj,0.12787,0.01000,3.235
58
+ 8,self_attn.k_proj,2.76998,0.01000,0.763
59
+ 8,self_attn.v_proj,2.57230,0.01000,0.757
60
+ 8,self_attn.q_proj,3.70224,0.01000,0.758
61
+ 8,self_attn.o_proj,0.43583,0.01000,0.608
62
+ 8,mlp.up_proj,7.05781,0.01000,0.771
63
+ 8,mlp.gate_proj,7.07004,0.01000,0.769
64
+ 8,mlp.down_proj,0.12891,0.01000,3.226
65
+ 9,self_attn.k_proj,5.24076,0.01000,0.760
66
+ 9,self_attn.v_proj,4.81714,0.01000,0.765
67
+ 9,self_attn.q_proj,1.28618,0.01000,0.761
68
+ 9,self_attn.o_proj,0.43630,0.01000,0.605
69
+ 9,mlp.up_proj,4.60663,0.01000,0.775
70
+ 9,mlp.gate_proj,4.60346,0.01000,0.769
71
+ 9,mlp.down_proj,0.07520,0.01000,3.231
72
+ 10,self_attn.k_proj,1.50184,0.01000,0.758
73
+ 10,self_attn.v_proj,1.45593,0.01000,0.762
74
+ 10,self_attn.q_proj,1.89232,0.01000,0.757
75
+ 10,self_attn.o_proj,0.19516,0.01000,0.605
76
+ 10,mlp.up_proj,4.09116,0.01000,0.770
77
+ 10,mlp.gate_proj,4.13186,0.01000,0.768
78
+ 10,mlp.down_proj,0.06140,0.01000,3.245
79
+ 11,self_attn.k_proj,3.61625,0.01000,0.762
80
+ 11,self_attn.v_proj,2.01000,0.01000,0.757
81
+ 11,self_attn.q_proj,0.53930,0.01000,0.766
82
+ 11,self_attn.o_proj,0.53488,0.01000,0.605
83
+ 11,mlp.up_proj,5.95990,0.01000,0.775
84
+ 11,mlp.gate_proj,6.01873,0.01000,0.771
85
+ 11,mlp.down_proj,1.49569,0.01000,3.277
86
+ 12,self_attn.k_proj,2.09444,0.01000,0.772
87
+ 12,self_attn.v_proj,1.75990,0.01000,0.764
88
+ 12,self_attn.q_proj,0.98691,0.01000,0.772
89
+ 12,self_attn.o_proj,0.32763,0.01000,0.611
90
+ 12,mlp.up_proj,2.53169,0.01000,0.784
91
+ 12,mlp.gate_proj,2.49588,0.01000,0.783
92
+ 12,mlp.down_proj,0.02076,0.01000,3.223
93
+ 13,self_attn.k_proj,3.88221,0.01000,0.765
94
+ 13,self_attn.v_proj,3.63538,0.01000,0.755
95
+ 13,self_attn.q_proj,0.56800,0.01000,0.764
96
+ 13,self_attn.o_proj,0.46188,0.01000,0.604
97
+ 13,mlp.up_proj,2.62719,0.01000,0.773
98
+ 13,mlp.gate_proj,2.61783,0.01000,0.773
99
+ 13,mlp.down_proj,0.01987,0.01000,3.300
100
+ 14,self_attn.k_proj,1.91174,0.01000,0.778
101
+ 14,self_attn.v_proj,1.67628,0.01000,0.773
102
+ 14,self_attn.q_proj,0.47230,0.01000,0.768
103
+ 14,self_attn.o_proj,0.22776,0.01000,0.621
104
+ 14,mlp.up_proj,2.01387,0.01000,0.778
105
+ 14,mlp.gate_proj,1.99085,0.01000,0.781
106
+ 14,mlp.down_proj,0.00910,0.01000,3.318
107
+ 15,self_attn.k_proj,1.71465,0.01000,0.776
108
+ 15,self_attn.v_proj,1.51512,0.01000,0.775
109
+ 15,self_attn.q_proj,0.28649,0.01000,0.784
110
+ 15,self_attn.o_proj,0.16413,0.01000,0.622
111
+ 15,mlp.up_proj,4.75803,0.01000,0.795
112
+ 15,mlp.gate_proj,4.92565,0.01000,0.793
113
+ 15,mlp.down_proj,0.02030,0.01000,3.306
114
+ 16,self_attn.k_proj,1.73286,0.01000,0.779
115
+ 16,self_attn.v_proj,1.52700,0.01000,0.776
116
+ 16,self_attn.q_proj,0.91500,0.01000,0.777
117
+ 16,self_attn.o_proj,0.11427,0.01000,0.619
118
+ 16,mlp.up_proj,3.70094,0.01000,0.783
119
+ 16,mlp.gate_proj,3.71431,0.01000,0.775
120
+ 16,mlp.down_proj,0.04519,0.01000,3.253
121
+ 17,self_attn.k_proj,0.43938,0.01000,0.762
122
+ 17,self_attn.v_proj,0.35879,0.01000,0.763
123
+ 17,self_attn.q_proj,0.65869,0.01000,0.765
124
+ 17,self_attn.o_proj,0.04037,0.01000,0.611
125
+ 17,mlp.up_proj,3.58469,0.01000,0.777
126
+ 17,mlp.gate_proj,3.66276,0.01000,0.770
127
+ 17,mlp.down_proj,0.02796,0.01000,3.257
128
+ 18,self_attn.k_proj,3.42578,0.01000,0.767
129
+ 18,self_attn.v_proj,3.09306,0.01000,0.761
130
+ 18,self_attn.q_proj,2.25221,0.01000,0.778
131
+ 18,self_attn.o_proj,0.24882,0.01000,0.610
132
+ 18,mlp.up_proj,4.09926,0.01000,0.779
133
+ 18,mlp.gate_proj,4.20446,0.01000,0.773
134
+ 18,mlp.down_proj,0.03055,0.01000,3.236
135
+ 19,self_attn.k_proj,2.29446,0.01000,0.761
136
+ 19,self_attn.v_proj,1.46921,0.01000,0.762
137
+ 19,self_attn.q_proj,2.12650,0.01000,0.761
138
+ 19,self_attn.o_proj,0.10793,0.01000,0.607
139
+ 19,mlp.up_proj,3.87133,0.01000,0.774
140
+ 19,mlp.gate_proj,4.02278,0.01000,0.771
141
+ 19,mlp.down_proj,0.02028,0.01000,3.260
142
+ 20,self_attn.k_proj,6.02629,0.01000,0.766
143
+ 20,self_attn.v_proj,5.54468,0.01000,0.758
144
+ 20,self_attn.q_proj,1.53195,0.01000,0.763
145
+ 20,self_attn.o_proj,0.51680,0.01000,0.612
146
+ 20,mlp.up_proj,3.97480,0.01000,0.775
147
+ 20,mlp.gate_proj,4.07580,0.01000,0.778
148
+ 20,mlp.down_proj,0.02892,0.01000,3.246
149
+ 21,self_attn.k_proj,5.52127,0.01000,0.768
150
+ 21,self_attn.v_proj,5.55817,0.01000,0.766
151
+ 21,self_attn.q_proj,2.16732,0.01000,0.779
152
+ 21,self_attn.o_proj,0.71680,0.01000,0.611
153
+ 21,mlp.up_proj,5.24819,0.01000,0.785
154
+ 21,mlp.gate_proj,5.41273,0.01000,0.785
155
+ 21,mlp.down_proj,0.04581,0.01000,3.246
156
+ 22,self_attn.k_proj,5.04561,0.01000,0.764
157
+ 22,self_attn.v_proj,4.48255,0.01000,0.761
158
+ 22,self_attn.q_proj,2.35098,0.01000,0.762
159
+ 22,self_attn.o_proj,0.48656,0.01000,0.606
160
+ 22,mlp.up_proj,6.14334,0.01000,0.778
161
+ 22,mlp.gate_proj,6.36109,0.01000,0.772
162
+ 22,mlp.down_proj,0.08612,0.01000,3.247
163
+ 23,self_attn.k_proj,4.16688,0.01000,0.767
164
+ 23,self_attn.v_proj,4.31470,0.01000,0.766
165
+ 23,self_attn.q_proj,0.69882,0.01000,0.763
166
+ 23,self_attn.o_proj,0.37936,0.01000,0.613
167
+ 23,mlp.up_proj,8.11542,0.01000,0.780
168
+ 23,mlp.gate_proj,8.28101,0.01000,0.775
169
+ 23,mlp.down_proj,0.09309,0.01000,3.265
170
+ 24,self_attn.k_proj,8.86272,0.01000,0.761
171
+ 24,self_attn.v_proj,8.50640,0.01000,0.760
172
+ 24,self_attn.q_proj,6.60116,0.01000,0.762
173
+ 24,self_attn.o_proj,0.78601,0.01000,0.612
174
+ 24,mlp.up_proj,8.48724,0.01000,0.778
175
+ 24,mlp.gate_proj,8.67515,0.01000,0.772
176
+ 24,mlp.down_proj,0.05045,0.01000,3.243
177
+ 25,self_attn.k_proj,20.85596,0.01000,0.763
178
+ 25,self_attn.v_proj,20.31687,0.01000,0.760
179
+ 25,self_attn.q_proj,4.40460,0.01000,0.765
180
+ 25,self_attn.o_proj,0.96945,0.01000,0.619
181
+ 25,mlp.up_proj,8.51875,0.01000,0.793
182
+ 25,mlp.gate_proj,8.63196,0.01000,0.790
183
+ 25,mlp.down_proj,0.06100,0.01000,3.293
184
+ 26,self_attn.k_proj,3.44908,0.01000,0.776
185
+ 26,self_attn.v_proj,3.70904,0.01000,0.776
186
+ 26,self_attn.q_proj,5.38957,0.01000,0.776
187
+ 26,self_attn.o_proj,0.09136,0.01000,0.617
188
+ 26,mlp.up_proj,7.22075,0.01000,0.787
189
+ 26,mlp.gate_proj,7.31307,0.01000,0.788
190
+ 26,mlp.down_proj,0.03320,0.01000,3.299
191
+ 27,self_attn.k_proj,3.16697,0.01000,0.778
192
+ 27,self_attn.v_proj,2.71902,0.01000,0.772
193
+ 27,self_attn.q_proj,6.13490,0.01000,0.776
194
+ 27,self_attn.o_proj,0.42601,0.01000,0.617
195
+ 27,mlp.up_proj,6.95249,0.01000,0.790
196
+ 27,mlp.gate_proj,7.11151,0.01000,0.789
197
+ 27,mlp.down_proj,0.10148,0.01000,3.296
198
+ 28,self_attn.k_proj,5.35555,0.01000,0.774
199
+ 28,self_attn.v_proj,5.28846,0.01000,0.774
200
+ 28,self_attn.q_proj,7.92698,0.01000,0.780
201
+ 28,self_attn.o_proj,0.21795,0.01000,0.616
202
+ 28,mlp.up_proj,4.31994,0.01000,0.790
203
+ 28,mlp.gate_proj,4.40171,0.01000,0.777
204
+ 28,mlp.down_proj,0.03607,0.01000,3.242
205
+ 29,self_attn.k_proj,1.51074,0.01000,0.767
206
+ 29,self_attn.v_proj,2.13929,0.01000,0.761
207
+ 29,self_attn.q_proj,1.33246,0.01000,0.762
208
+ 29,self_attn.o_proj,0.01362,0.01000,0.621
209
+ 29,mlp.up_proj,2.46330,0.01000,0.779
210
+ 29,mlp.gate_proj,2.51679,0.01000,0.771
211
+ 29,mlp.down_proj,0.01289,0.01000,3.285
212
+ 30,self_attn.k_proj,6.86594,0.01000,0.774
213
+ 30,self_attn.v_proj,7.52277,0.01000,0.772
214
+ 30,self_attn.q_proj,0.92430,0.01000,0.779
215
+ 30,self_attn.o_proj,0.15717,0.01000,0.618
216
+ 30,mlp.up_proj,5.04547,0.01000,0.785
217
+ 30,mlp.gate_proj,5.15212,0.01000,0.782
218
+ 30,mlp.down_proj,0.04664,0.01000,3.287
219
+ 31,self_attn.k_proj,25.90375,0.01000,0.775
220
+ 31,self_attn.v_proj,20.06681,0.01000,0.774
221
+ 31,self_attn.q_proj,7.96819,0.01000,0.775
222
+ 31,self_attn.o_proj,3.03762,0.01000,0.620
223
+ 31,mlp.up_proj,8.97430,0.01000,0.789
224
+ 31,mlp.gate_proj,9.35650,0.01000,0.788
225
+ 31,mlp.down_proj,0.20598,0.01000,3.297
226
+ 32,self_attn.k_proj,2.33364,0.01000,0.776
227
+ 32,self_attn.v_proj,2.47755,0.01000,0.776
228
+ 32,self_attn.q_proj,3.65308,0.01000,0.773
229
+ 32,self_attn.o_proj,0.12603,0.01000,0.618
230
+ 32,mlp.up_proj,7.92819,0.01000,0.788
231
+ 32,mlp.gate_proj,8.39347,0.01000,0.783
232
+ 32,mlp.down_proj,0.07191,0.01000,3.293
233
+ 33,self_attn.k_proj,3.03554,0.01000,0.774
234
+ 33,self_attn.v_proj,3.00391,0.01000,0.773
235
+ 33,self_attn.q_proj,4.58165,0.01000,0.774
236
+ 33,self_attn.o_proj,0.56861,0.01000,0.623
237
+ 33,mlp.up_proj,13.49779,0.01000,0.789
238
+ 33,mlp.gate_proj,13.66948,0.01000,0.787
239
+ 33,mlp.down_proj,0.34180,0.01000,3.298
quantize_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 4,
3
+ "group_size": 128,
4
+ "desc_act": true,
5
+ "sym": true,
6
+ "lm_head": false,
7
+ "quant_method": "gptq",
8
+ "checkpoint_format": "gptq",
9
+ "pack_dtype": "int32",
10
+ "meta": {
11
+ "quantizer": [
12
+ "gptqmodel:2.1.0"
13
+ ],
14
+ "uri": "https://github.com/modelcloud/gptqmodel",
15
+ "damp_percent": 0.01,
16
+ "damp_auto_increment": 0.0025,
17
+ "static_groups": false,
18
+ "true_sequential": true,
19
+ "mse": 0.0
20
+ }
21
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "boi_token": "<start_of_image>",
3
+ "bos_token": {
4
+ "content": "<bos>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ "eoi_token": "<end_of_image>",
11
+ "eos_token": {
12
+ "content": "<end_of_turn>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false
17
+ },
18
+ "image_token": "<image_soft_token>",
19
+ "pad_token": "<pad>",
20
+ "unk_token": {
21
+ "content": "<unk>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false
26
+ }
27
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4667f2089529e8e7657cfb6d1c19910ae71ff5f28aa7ab2ff2763330affad795
3
+ size 33384568
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1299c11d7cf632ef3b4e11937501358ada021bbdf7c47638d13c0ee982f2e79c
3
+ size 4689074
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff