prince-canuma commited on
Commit
629084c
·
verified ·
1 Parent(s): 77a459a

Add files using upload-large-folder tool

Browse files
README.md ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: mlx
3
+ tags:
4
+ - falcon-h1
5
+ - edge
6
+ - mlx
7
+ license: other
8
+ license_name: falcon-llm-license
9
+ license_link: https://falconllm.tii.ae/falcon-terms-and-conditions.html
10
+ base_model: tiiuae/Falcon-H1-Tiny-R-0.6B
11
+ pipeline_tag: text-generation
12
+ ---
13
+
14
+ # mlx-community/Falcon-H1-Tiny-R-0.6B-4bit
15
+
16
+ This model [mlx-community/Falcon-H1-Tiny-R-0.6B-4bit](https://huggingface.co/mlx-community/Falcon-H1-Tiny-R-0.6B-4bit) was
17
+ converted to MLX format from [tiiuae/Falcon-H1-Tiny-R-0.6B](https://huggingface.co/tiiuae/Falcon-H1-Tiny-R-0.6B)
18
+ using mlx-lm version **0.30.5**.
19
+
20
+ ## Use with mlx
21
+
22
+ ```bash
23
+ pip install mlx-lm
24
+ ```
25
+
26
+ ```python
27
+ from mlx_lm import load, generate
28
+
29
+ model, tokenizer = load("mlx-community/Falcon-H1-Tiny-R-0.6B-4bit")
30
+
31
+ prompt = "hello"
32
+
33
+ if tokenizer.chat_template is not None:
34
+ messages = [{"role": "user", "content": prompt}]
35
+ prompt = tokenizer.apply_chat_template(
36
+ messages, add_generation_prompt=True, return_dict=False,
37
+ )
38
+
39
+ response = generate(model, tokenizer, prompt=prompt, verbose=True)
40
+ ```
chat_template.jinja ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %} {{bos_token}}<|system|>
2
+ {%- if messages[0]['role'] == 'system' %}
3
+ {{ messages[0]['content'] }}
4
+ {%- set remaining_messages = messages[1:] %}
5
+ {%- else %}
6
+ {%- set remaining_messages = messages %}
7
+ {%- endif %}
8
+ {{ 'You are a Falcon assistant skilled in function calling. You are helpful, respectful, and concise.
9
+
10
+ # Tools
11
+
12
+ You have access to the following functions. You MUST use them to answer questions when needed. For each function call, you MUST return a JSON object inside <tool_call></tool_call> tags.
13
+
14
+ <tools>' + tools|tojson(indent=2) + '</tools>
15
+
16
+ # Output Format
17
+
18
+ Your response MUST follow this format when making function calls:
19
+ <tool_call>
20
+ [
21
+ {"name": "function_name", "arguments": {"arg1": "value1", "arg2": "value2"}},
22
+ {"name": "another_function", "arguments": {"arg": "value"}}
23
+ ]
24
+ </tool_call>
25
+ If no function calls are needed, respond normally without the tool_call tags.' }}
26
+ {%- for message in remaining_messages %}
27
+ {%- if message['role'] == 'user' %}
28
+ <|im_start|>user
29
+ {{ message['content'] }}<|im_end|>
30
+ {%- elif message['role'] == 'assistant' %}
31
+ {%- if message.content %}
32
+ <|im_start|>assistant
33
+ {{ message['content'] }}
34
+ <|im_end|>
35
+ {%- endif %}
36
+ {%- if message.tool_calls %}
37
+ <tool_call>
38
+ {{ message.tool_calls|tojson(indent=2) }}
39
+ </tool_call>
40
+ {%- endif %}
41
+ {%- elif message['role'] == 'tool' %}
42
+ <|im_start|>assistant
43
+ <tool_response>
44
+ {{ message['content'] }}
45
+ </tool_response><|im_end|>
46
+ {%- endif %}
47
+ {%- endfor %}
48
+ {{ '<|im_start|>assistant
49
+ ' if add_generation_prompt }}
50
+ {%- else %} {{bos_token}}{% for message in messages %} {{ '<|im_start|>' + message['role'] + '
51
+ ' + message['content'] + '<|im_end|>
52
+ ' }} {% endfor %} {% if add_generation_prompt %}{{ '<|im_start|>assistant
53
+ ' }}{% endif %} {%- endif %}
config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "FalconH1ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attention_in_multiplier": 1.0,
8
+ "attention_out_multiplier": 1.0,
9
+ "attn_layer_indices": null,
10
+ "bos_token_id": 1,
11
+ "dtype": "bfloat16",
12
+ "embedding_multiplier": 1.0,
13
+ "eos_token_id": [
14
+ 11,
15
+ 228
16
+ ],
17
+ "expansion_factor": 2.0,
18
+ "head_dim": 64,
19
+ "hidden_act": "silu",
20
+ "hidden_size": 1024,
21
+ "initializer_range": 0.02,
22
+ "intermediate_size": 2048,
23
+ "key_multiplier": 0.03125,
24
+ "lm_head_multiplier": 1.0,
25
+ "mamba_chunk_size": 128,
26
+ "mamba_conv_bias": true,
27
+ "mamba_d_conv": 4,
28
+ "mamba_d_head": 64,
29
+ "mamba_d_ssm": 1536,
30
+ "mamba_d_state": 128,
31
+ "mamba_expand": 2,
32
+ "mamba_n_groups": 1,
33
+ "mamba_n_heads": 24,
34
+ "mamba_norm_before_gate": false,
35
+ "mamba_proj_bias": false,
36
+ "mamba_rms_norm": false,
37
+ "mamba_use_mlp": true,
38
+ "max_position_embeddings": 262144,
39
+ "mlp_bias": false,
40
+ "mlp_expansion_factor": 8,
41
+ "mlp_multipliers": [
42
+ 1.0,
43
+ 1.0
44
+ ],
45
+ "model_type": "falcon_h1",
46
+ "num_attention_heads": 8,
47
+ "num_hidden_layers": 44,
48
+ "num_key_value_heads": 2,
49
+ "num_logits_to_keep": 1,
50
+ "pad_token_id": 0,
51
+ "projectors_bias": false,
52
+ "quantization": {
53
+ "group_size": 64,
54
+ "bits": 4,
55
+ "mode": "affine"
56
+ },
57
+ "quantization_config": {
58
+ "group_size": 64,
59
+ "bits": 4,
60
+ "mode": "affine"
61
+ },
62
+ "rms_norm_eps": 1e-05,
63
+ "rope_scaling": null,
64
+ "rope_theta": 100000000000.0,
65
+ "sliding_window": null,
66
+ "ssm_in_multiplier": 1.0,
67
+ "ssm_multipliers": [
68
+ 1.0,
69
+ 1.0,
70
+ 1.0,
71
+ 1.0,
72
+ 0.03125
73
+ ],
74
+ "ssm_out_multiplier": 1.0,
75
+ "tie_word_embeddings": false,
76
+ "time_step_floor": 0.0001,
77
+ "time_step_max": 0.1,
78
+ "time_step_min": 0.001,
79
+ "time_step_rank": "auto",
80
+ "transformers_version": "4.57.0",
81
+ "use_cache": true,
82
+ "vocab_size": 32768
83
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": [
5
+ 11,
6
+ 228
7
+ ],
8
+ "pad_token_id": 0,
9
+ "transformers_version": "4.57.0"
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:582e883fefa82d2f03c69605b22dc9112f97fd0e83fe811e888bf7a9a694c0e3
3
+ size 350929024
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,343 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|end_of_text|>",
6
+ "extra_special_tokens": [
7
+ "<|pad|>",
8
+ ">>ABSTRACT<<",
9
+ ">>INTRODUCTION<<",
10
+ ">>SUMMARY<<",
11
+ ">>COMMENT<<",
12
+ ">>ANSWER<<",
13
+ ">>QUESTION<<",
14
+ ">>DOMAIN<<",
15
+ ">>PREFIX<<",
16
+ ">>SUFFIX<<",
17
+ ">>MIDDLE<<",
18
+ "<|finetune_right_pad_id|>",
19
+ "<|start_header_id|>",
20
+ "<|end_header_id|>",
21
+ "<|eom_id|>",
22
+ "<|eot_id|>",
23
+ "<|begin_of_text|>",
24
+ ">>TITLE<<",
25
+ "<tool_response>",
26
+ "</tool_response>",
27
+ "<tool_call>",
28
+ "</tool_call>",
29
+ "<schema>",
30
+ "</schema>",
31
+ "<scratch_pad>",
32
+ "</scratch_pad>",
33
+ "<thinking>",
34
+ "</thinking>",
35
+ "<explanation>",
36
+ "</explanation>",
37
+ "<file_sep>",
38
+ "<repo_name>",
39
+ "<|im_end|>",
40
+ "<|im_start|>",
41
+ "<|system|>",
42
+ "<|file_sep|>",
43
+ "<|repo_name|>",
44
+ "<|repo_tree|>",
45
+ ">>UNUSED_221<<",
46
+ ">>UNUSED_222<<",
47
+ ">>UNUSED_223<<",
48
+ ">>UNUSED_224<<",
49
+ ">>UNUSED_225<<",
50
+ ">>UNUSED_226<<",
51
+ ">>UNUSED_227<<",
52
+ ">>UNUSED_228<<",
53
+ ">>UNUSED_229<<",
54
+ ">>UNUSED_230<<",
55
+ ">>UNUSED_231<<",
56
+ ">>UNUSED_232<<",
57
+ ">>UNUSED_233<<",
58
+ ">>UNUSED_234<<",
59
+ ">>UNUSED_235<<",
60
+ ">>UNUSED_236<<",
61
+ ">>UNUSED_237<<",
62
+ ">>UNUSED_238<<",
63
+ ">>UNUSED_239<<",
64
+ ">>UNUSED_240<<",
65
+ ">>UNUSED_241<<",
66
+ ">>UNUSED_242<<",
67
+ ">>UNUSED_243<<",
68
+ ">>UNUSED_244<<",
69
+ ">>UNUSED_245<<",
70
+ ">>UNUSED_246<<",
71
+ ">>UNUSED_247<<",
72
+ ">>UNUSED_248<<",
73
+ ">>UNUSED_249<<",
74
+ ">>UNUSED_250<<",
75
+ ">>UNUSED_251<<",
76
+ ">>UNUSED_252<<",
77
+ ">>UNUSED_253<<",
78
+ ">>UNUSED_254<<",
79
+ ">>UNUSED_255<<",
80
+ ">>UNUSED_256<<",
81
+ ">>UNUSED_257<<",
82
+ ">>UNUSED_258<<",
83
+ ">>UNUSED_259<<",
84
+ ">>UNUSED_260<<",
85
+ ">>UNUSED_261<<",
86
+ ">>UNUSED_262<<",
87
+ ">>UNUSED_263<<",
88
+ ">>UNUSED_264<<",
89
+ ">>UNUSED_265<<",
90
+ ">>UNUSED_266<<",
91
+ ">>UNUSED_267<<",
92
+ ">>UNUSED_268<<",
93
+ ">>UNUSED_269<<",
94
+ ">>UNUSED_270<<",
95
+ ">>UNUSED_271<<",
96
+ ">>UNUSED_272<<",
97
+ ">>UNUSED_273<<",
98
+ ">>UNUSED_274<<",
99
+ ">>UNUSED_275<<",
100
+ ">>UNUSED_276<<",
101
+ ">>UNUSED_277<<",
102
+ ">>UNUSED_278<<",
103
+ ">>UNUSED_279<<",
104
+ ">>UNUSED_280<<",
105
+ ">>UNUSED_281<<",
106
+ ">>UNUSED_282<<",
107
+ ">>UNUSED_283<<",
108
+ ">>UNUSED_284<<",
109
+ ">>UNUSED_285<<",
110
+ ">>UNUSED_286<<",
111
+ ">>UNUSED_287<<",
112
+ ">>UNUSED_288<<",
113
+ ">>UNUSED_289<<",
114
+ ">>UNUSED_290<<",
115
+ ">>UNUSED_291<<",
116
+ ">>UNUSED_292<<",
117
+ ">>UNUSED_293<<",
118
+ ">>UNUSED_294<<",
119
+ ">>UNUSED_295<<",
120
+ ">>UNUSED_296<<",
121
+ ">>UNUSED_297<<",
122
+ ">>UNUSED_298<<",
123
+ ">>UNUSED_299<<",
124
+ ">>UNUSED_300<<",
125
+ ">>UNUSED_301<<",
126
+ ">>UNUSED_302<<",
127
+ ">>UNUSED_303<<",
128
+ ">>UNUSED_304<<",
129
+ ">>UNUSED_305<<",
130
+ ">>UNUSED_306<<",
131
+ ">>UNUSED_307<<",
132
+ ">>UNUSED_308<<",
133
+ ">>UNUSED_309<<",
134
+ ">>UNUSED_310<<",
135
+ ">>UNUSED_311<<",
136
+ ">>UNUSED_312<<",
137
+ ">>UNUSED_313<<",
138
+ ">>UNUSED_314<<",
139
+ ">>UNUSED_315<<",
140
+ ">>UNUSED_316<<",
141
+ ">>UNUSED_317<<",
142
+ ">>UNUSED_318<<",
143
+ ">>UNUSED_319<<",
144
+ ">>UNUSED_320<<",
145
+ ">>UNUSED_321<<",
146
+ ">>UNUSED_322<<",
147
+ ">>UNUSED_323<<",
148
+ ">>UNUSED_324<<",
149
+ ">>UNUSED_325<<",
150
+ ">>UNUSED_326<<",
151
+ ">>UNUSED_327<<",
152
+ ">>UNUSED_328<<",
153
+ ">>UNUSED_329<<",
154
+ ">>UNUSED_330<<",
155
+ ">>UNUSED_331<<",
156
+ ">>UNUSED_332<<",
157
+ ">>UNUSED_333<<",
158
+ ">>UNUSED_334<<",
159
+ ">>UNUSED_335<<",
160
+ ">>UNUSED_336<<",
161
+ ">>UNUSED_337<<",
162
+ ">>UNUSED_338<<",
163
+ ">>UNUSED_339<<",
164
+ ">>UNUSED_340<<",
165
+ ">>UNUSED_341<<",
166
+ ">>UNUSED_342<<",
167
+ ">>UNUSED_343<<",
168
+ ">>UNUSED_344<<",
169
+ ">>UNUSED_345<<",
170
+ ">>UNUSED_346<<",
171
+ ">>UNUSED_347<<",
172
+ ">>UNUSED_348<<",
173
+ ">>UNUSED_349<<",
174
+ ">>UNUSED_350<<",
175
+ ">>UNUSED_351<<",
176
+ ">>UNUSED_352<<",
177
+ ">>UNUSED_353<<",
178
+ ">>UNUSED_354<<",
179
+ ">>UNUSED_355<<",
180
+ ">>UNUSED_356<<",
181
+ ">>UNUSED_357<<",
182
+ ">>UNUSED_358<<",
183
+ ">>UNUSED_359<<",
184
+ ">>UNUSED_360<<",
185
+ ">>UNUSED_361<<",
186
+ ">>UNUSED_362<<",
187
+ ">>UNUSED_363<<",
188
+ ">>UNUSED_364<<",
189
+ ">>UNUSED_365<<",
190
+ ">>UNUSED_366<<",
191
+ ">>UNUSED_367<<",
192
+ ">>UNUSED_368<<",
193
+ ">>UNUSED_369<<",
194
+ ">>UNUSED_370<<",
195
+ ">>UNUSED_371<<",
196
+ ">>UNUSED_372<<",
197
+ ">>UNUSED_373<<",
198
+ ">>UNUSED_374<<",
199
+ ">>UNUSED_375<<",
200
+ ">>UNUSED_376<<",
201
+ ">>UNUSED_377<<",
202
+ ">>UNUSED_378<<",
203
+ ">>UNUSED_379<<",
204
+ ">>UNUSED_380<<",
205
+ ">>UNUSED_381<<",
206
+ ">>UNUSED_382<<",
207
+ ">>UNUSED_383<<",
208
+ ">>UNUSED_384<<",
209
+ ">>UNUSED_385<<",
210
+ ">>UNUSED_386<<",
211
+ ">>UNUSED_387<<",
212
+ ">>UNUSED_388<<",
213
+ ">>UNUSED_389<<",
214
+ ">>UNUSED_390<<",
215
+ ">>UNUSED_391<<",
216
+ ">>UNUSED_392<<",
217
+ ">>UNUSED_393<<",
218
+ ">>UNUSED_394<<",
219
+ ">>UNUSED_395<<",
220
+ ">>UNUSED_396<<",
221
+ ">>UNUSED_397<<",
222
+ ">>UNUSED_398<<",
223
+ ">>UNUSED_399<<",
224
+ ">>UNUSED_400<<",
225
+ ">>UNUSED_401<<",
226
+ ">>UNUSED_402<<",
227
+ ">>UNUSED_403<<",
228
+ ">>UNUSED_404<<",
229
+ ">>UNUSED_405<<",
230
+ ">>UNUSED_406<<",
231
+ ">>UNUSED_407<<",
232
+ ">>UNUSED_408<<",
233
+ ">>UNUSED_409<<",
234
+ ">>UNUSED_410<<",
235
+ ">>UNUSED_411<<",
236
+ ">>UNUSED_412<<",
237
+ ">>UNUSED_413<<",
238
+ ">>UNUSED_414<<",
239
+ ">>UNUSED_415<<",
240
+ ">>UNUSED_416<<",
241
+ ">>UNUSED_417<<",
242
+ ">>UNUSED_418<<",
243
+ ">>UNUSED_419<<",
244
+ ">>UNUSED_420<<",
245
+ ">>UNUSED_421<<",
246
+ ">>UNUSED_422<<",
247
+ ">>UNUSED_423<<",
248
+ ">>UNUSED_424<<",
249
+ ">>UNUSED_425<<",
250
+ ">>UNUSED_426<<",
251
+ ">>UNUSED_427<<",
252
+ ">>UNUSED_428<<",
253
+ ">>UNUSED_429<<",
254
+ ">>UNUSED_430<<",
255
+ ">>UNUSED_431<<",
256
+ ">>UNUSED_432<<",
257
+ ">>UNUSED_433<<",
258
+ ">>UNUSED_434<<",
259
+ ">>UNUSED_435<<",
260
+ ">>UNUSED_436<<",
261
+ ">>UNUSED_437<<",
262
+ ">>UNUSED_438<<",
263
+ ">>UNUSED_439<<",
264
+ ">>UNUSED_440<<",
265
+ ">>UNUSED_441<<",
266
+ ">>UNUSED_442<<",
267
+ ">>UNUSED_443<<",
268
+ ">>UNUSED_444<<",
269
+ ">>UNUSED_445<<",
270
+ ">>UNUSED_446<<",
271
+ ">>UNUSED_447<<",
272
+ ">>UNUSED_448<<",
273
+ ">>UNUSED_449<<",
274
+ ">>UNUSED_450<<",
275
+ ">>UNUSED_451<<",
276
+ ">>UNUSED_452<<",
277
+ ">>UNUSED_453<<",
278
+ ">>UNUSED_454<<",
279
+ ">>UNUSED_455<<",
280
+ ">>UNUSED_456<<",
281
+ ">>UNUSED_457<<",
282
+ ">>UNUSED_458<<",
283
+ ">>UNUSED_459<<",
284
+ ">>UNUSED_460<<",
285
+ ">>UNUSED_461<<",
286
+ ">>UNUSED_462<<",
287
+ ">>UNUSED_463<<",
288
+ ">>UNUSED_464<<",
289
+ ">>UNUSED_465<<",
290
+ ">>UNUSED_466<<",
291
+ ">>UNUSED_467<<",
292
+ ">>UNUSED_468<<",
293
+ ">>UNUSED_469<<",
294
+ ">>UNUSED_470<<",
295
+ ">>UNUSED_471<<",
296
+ ">>UNUSED_472<<",
297
+ ">>UNUSED_473<<",
298
+ ">>UNUSED_474<<",
299
+ ">>UNUSED_475<<",
300
+ ">>UNUSED_476<<",
301
+ ">>UNUSED_477<<",
302
+ ">>UNUSED_478<<",
303
+ ">>UNUSED_479<<",
304
+ ">>UNUSED_480<<",
305
+ ">>UNUSED_481<<",
306
+ ">>UNUSED_482<<",
307
+ ">>UNUSED_483<<",
308
+ ">>UNUSED_484<<",
309
+ ">>UNUSED_485<<",
310
+ ">>UNUSED_486<<",
311
+ ">>UNUSED_487<<",
312
+ ">>UNUSED_488<<",
313
+ ">>UNUSED_489<<",
314
+ ">>UNUSED_490<<",
315
+ ">>UNUSED_491<<",
316
+ ">>UNUSED_492<<",
317
+ ">>UNUSED_493<<",
318
+ ">>UNUSED_494<<",
319
+ ">>UNUSED_495<<",
320
+ ">>UNUSED_496<<",
321
+ ">>UNUSED_497<<",
322
+ ">>UNUSED_498<<",
323
+ ">>UNUSED_499<<",
324
+ ">>UNUSED_500<<",
325
+ ">>UNUSED_501<<",
326
+ ">>UNUSED_502<<",
327
+ ">>UNUSED_503<<",
328
+ ">>UNUSED_504<<",
329
+ ">>UNUSED_505<<",
330
+ ">>UNUSED_506<<",
331
+ ">>UNUSED_507<<",
332
+ ">>UNUSED_508<<",
333
+ ">>UNUSED_509<<",
334
+ ">>UNUSED_510<<",
335
+ ">>UNUSED_511<<"
336
+ ],
337
+ "is_local": true,
338
+ "model_max_length": 1000000000000000019884624838656,
339
+ "model_specific_special_tokens": {},
340
+ "pad_token": "<|pad|>",
341
+ "pad_token_id": 0,
342
+ "tokenizer_class": "TokenizersBackend"
343
+ }