electroglyph commited on
Commit
207b8d6
·
verified ·
1 Parent(s): 1cc68f4

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ benchmark.png filter=lfs diff=lfs merge=lfs -text
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,770 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - unsloth
4
+ - sentence-transformers
5
+ - sentence-similarity
6
+ - feature-extraction
7
+ - dense
8
+ - generated_from_trainer
9
+ - dataset_size:761918
10
+ - loss:MultipleNegativesRankingLoss
11
+ base_model: Alibaba-NLP/gte-modernbert-base
12
+ pipeline_tag: sentence-similarity
13
+ library_name: sentence-transformers
14
+ ---
15
+
16
+ # SentenceTransformer
17
+
18
+ This model was finetuned with [Unsloth](https://github.com/unslothai/unsloth).
19
+
20
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
21
+ based on Alibaba-NLP/gte-modernbert-base
22
+
23
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) on the json dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
24
+
25
+ This model is finetuned specifically for fiction retrieval. It's been trained on sci-fi, fantasy, mystery, and other fiction genres.
26
+
27
+ Dataset size: 800k rows based on 100% manually cleaned data.
28
+
29
+ This model surpasses Qwen3 4B embedding model on my test set (40k examples with hard negatives) by 0.5%.
30
+
31
+ Model accuracy increased from 90.8% to 95.7% on the test set.
32
+
33
+ Some MTEB benchmarks saw some pretty big losses, they're detailed below.
34
+
35
+ I did plenty of runs with this model from rank 1 on up, but all of them damaged other benchmarks similarly.
36
+
37
+ ## Benchmark comparison, base model vs. this one
38
+
39
+ There are plenty of benchmark regressions with this model, but a couple gains.
40
+
41
+ ![benchmark](./benchmark.png)
42
+
43
+
44
+ ## Model Details
45
+
46
+ ### Model Description
47
+ - **Model Type:** Sentence Transformer
48
+ - **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision e7f32e3c00f91d699e8c43b53106206bcc72bb22 -->
49
+ - **Maximum Sequence Length:** 512 tokens
50
+ - **Output Dimensionality:** 768 dimensions
51
+ - **Similarity Function:** Cosine Similarity
52
+ - **Training Dataset:**
53
+ - json
54
+
55
+ ### Model Sources
56
+
57
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
58
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
59
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
60
+
61
+ ### Full Model Architecture
62
+
63
+ ```
64
+ SentenceTransformer(
65
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'PeftModelForFeatureExtraction'})
66
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
67
+ )
68
+ ```
69
+
70
+ ## Usage
71
+
72
+ ### Direct Usage (Sentence Transformers)
73
+
74
+ First install the Sentence Transformers library:
75
+
76
+ ```bash
77
+ pip install -U sentence-transformers
78
+ ```
79
+
80
+ Then you can load this model and run inference.
81
+ ```python
82
+ from sentence_transformers import SentenceTransformer
83
+
84
+ # Download from the 🤗 Hub
85
+ model = SentenceTransformer("sentence_transformers_model_id")
86
+ # Run inference
87
+ sentences = [
88
+ 'A being associated with malevolent influence must exhibit tangible signs of corruption, while a person disregarded by others despite a legitimate claim to priority quietly calculates a missed academic obligation, realizing that certain values surpass academic perfection.',
89
+ "Surely a demon-caller must show some outward manifestation of that kind of evil. Norman pushed into the first available seat. His injured hand should've entitled him to one the moment he got on the bus but not one of his selfish, self-centered fellow students would get up although he'd glared at all and sundry. Still sulking, he fished his calculator out of his shirt pocket, and began to work out the time he'd need to spend downtown. He was, at that very moment, missing an analytical geometry class. It was the first class he'd ever skipped. His parents would have fits. He didn't care. As much as he'd hoarded every A and A plus---he had a complete record of every mark he'd ever received---he'd realized in the last couple of days that some things were more important.",
90
+ 'It was one of the things he liked best about this part of the city, the fact that it never really slept, and it was why he had his home as close to it as he could get. Two blocks past Yonge, he turned into a circular drive and followed the curve around to the door of his building. In his time, he had lived in castles of every description, a fair number of very private country estates, and even a crypt or two when times were bad, but it had been centuries since he\'d had a home that suited him as well as the condominium he\'d bought in the heart of Toronto. "Good evening, Mr. Fitzroy." "Evening, Greg. Anything happening?" The security guard smiled and reached for the door release.',
91
+ ]
92
+ embeddings = model.encode(sentences)
93
+ print(embeddings.shape)
94
+ # [3, 768]
95
+
96
+ # Get the similarity scores for the embeddings
97
+ similarities = model.similarity(embeddings, embeddings)
98
+ print(similarities)
99
+ # tensor([[ 1.0000, 0.4737, -0.0624],
100
+ # [ 0.4737, 1.0000, 0.0190],
101
+ # [-0.0624, 0.0190, 1.0000]])
102
+ ```
103
+
104
+
105
+ ## Training Details
106
+
107
+ * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
108
+ ```json
109
+ {
110
+ "scale": 20.0,
111
+ "similarity_fct": "cos_sim",
112
+ "gather_across_devices": false
113
+ }
114
+ ```
115
+
116
+ ### Training Hyperparameters
117
+ #### Non-Default Hyperparameters
118
+
119
+ - `per_device_train_batch_size`: 500
120
+ - `learning_rate`: 6e-06
121
+ - `weight_decay`: 0.01
122
+ - `lr_scheduler_type`: cosine_with_restarts
123
+ - `warmup_steps`: 100
124
+ - `bf16`: True
125
+
126
+ #### All Hyperparameters
127
+ <details><summary>Click to expand</summary>
128
+
129
+ - `overwrite_output_dir`: False
130
+ - `do_predict`: False
131
+ - `eval_strategy`: no
132
+ - `prediction_loss_only`: True
133
+ - `per_device_train_batch_size`: 500
134
+ - `per_device_eval_batch_size`: 8
135
+ - `per_gpu_train_batch_size`: None
136
+ - `per_gpu_eval_batch_size`: None
137
+ - `gradient_accumulation_steps`: 1
138
+ - `eval_accumulation_steps`: None
139
+ - `torch_empty_cache_steps`: None
140
+ - `learning_rate`: 6e-06
141
+ - `weight_decay`: 0.01
142
+ - `adam_beta1`: 0.9
143
+ - `adam_beta2`: 0.999
144
+ - `adam_epsilon`: 1e-08
145
+ - `max_grad_norm`: 1.0
146
+ - `num_train_epochs`: 3
147
+ - `max_steps`: -1
148
+ - `lr_scheduler_type`: cosine_with_restarts
149
+ - `lr_scheduler_kwargs`: {}
150
+ - `warmup_ratio`: 0.0
151
+ - `warmup_steps`: 100
152
+ - `log_level`: passive
153
+ - `log_level_replica`: warning
154
+ - `log_on_each_node`: True
155
+ - `logging_nan_inf_filter`: True
156
+ - `save_safetensors`: True
157
+ - `save_on_each_node`: False
158
+ - `save_only_model`: False
159
+ - `restore_callback_states_from_checkpoint`: False
160
+ - `no_cuda`: False
161
+ - `use_cpu`: False
162
+ - `use_mps_device`: False
163
+ - `seed`: 42
164
+ - `data_seed`: None
165
+ - `jit_mode_eval`: False
166
+ - `bf16`: True
167
+ - `fp16`: False
168
+ - `fp16_opt_level`: O1
169
+ - `half_precision_backend`: auto
170
+ - `bf16_full_eval`: False
171
+ - `fp16_full_eval`: False
172
+ - `tf32`: None
173
+ - `local_rank`: 0
174
+ - `ddp_backend`: None
175
+ - `tpu_num_cores`: None
176
+ - `tpu_metrics_debug`: False
177
+ - `debug`: []
178
+ - `dataloader_drop_last`: False
179
+ - `dataloader_num_workers`: 0
180
+ - `dataloader_prefetch_factor`: None
181
+ - `past_index`: -1
182
+ - `disable_tqdm`: False
183
+ - `remove_unused_columns`: True
184
+ - `label_names`: None
185
+ - `load_best_model_at_end`: False
186
+ - `ignore_data_skip`: False
187
+ - `fsdp`: []
188
+ - `fsdp_min_num_params`: 0
189
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
190
+ - `fsdp_transformer_layer_cls_to_wrap`: None
191
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
192
+ - `parallelism_config`: None
193
+ - `deepspeed`: None
194
+ - `label_smoothing_factor`: 0.0
195
+ - `optim`: adamw_torch_fused
196
+ - `optim_args`: None
197
+ - `adafactor`: False
198
+ - `group_by_length`: False
199
+ - `length_column_name`: length
200
+ - `project`: huggingface
201
+ - `trackio_space_id`: trackio
202
+ - `ddp_find_unused_parameters`: None
203
+ - `ddp_bucket_cap_mb`: None
204
+ - `ddp_broadcast_buffers`: False
205
+ - `dataloader_pin_memory`: True
206
+ - `dataloader_persistent_workers`: False
207
+ - `skip_memory_metrics`: True
208
+ - `use_legacy_prediction_loop`: False
209
+ - `push_to_hub`: False
210
+ - `resume_from_checkpoint`: None
211
+ - `hub_model_id`: None
212
+ - `hub_strategy`: every_save
213
+ - `hub_private_repo`: None
214
+ - `hub_always_push`: False
215
+ - `hub_revision`: None
216
+ - `gradient_checkpointing`: False
217
+ - `gradient_checkpointing_kwargs`: None
218
+ - `include_inputs_for_metrics`: False
219
+ - `include_for_metrics`: []
220
+ - `eval_do_concat_batches`: True
221
+ - `fp16_backend`: auto
222
+ - `push_to_hub_model_id`: None
223
+ - `push_to_hub_organization`: None
224
+ - `mp_parameters`:
225
+ - `auto_find_batch_size`: False
226
+ - `full_determinism`: False
227
+ - `torchdynamo`: None
228
+ - `ray_scope`: last
229
+ - `ddp_timeout`: 1800
230
+ - `torch_compile`: False
231
+ - `torch_compile_backend`: None
232
+ - `torch_compile_mode`: None
233
+ - `include_tokens_per_second`: False
234
+ - `include_num_input_tokens_seen`: no
235
+ - `neftune_noise_alpha`: None
236
+ - `optim_target_modules`: None
237
+ - `batch_eval_metrics`: False
238
+ - `eval_on_start`: False
239
+ - `use_liger_kernel`: False
240
+ - `liger_kernel_config`: None
241
+ - `eval_use_gather_object`: False
242
+ - `average_tokens_across_devices`: True
243
+ - `prompts`: None
244
+ - `batch_sampler`: batch_sampler
245
+ - `multi_dataset_batch_sampler`: proportional
246
+ - `router_mapping`: {}
247
+ - `learning_rate_mapping`: {}
248
+
249
+ </details>
250
+
251
+ ### Training Logs
252
+ <details><summary>Click to expand</summary>
253
+
254
+ | Epoch | Step | Training Loss |
255
+ |:------:|:----:|:-------------:|
256
+ | 0.0066 | 10 | 5.0675 |
257
+ | 0.0131 | 20 | 5.0698 |
258
+ | 0.0197 | 30 | 5.0747 |
259
+ | 0.0262 | 40 | 5.1686 |
260
+ | 0.0328 | 50 | 5.0452 |
261
+ | 0.0394 | 60 | 5.0103 |
262
+ | 0.0459 | 70 | 5.0011 |
263
+ | 0.0525 | 80 | 5.0711 |
264
+ | 0.0591 | 90 | 5.0241 |
265
+ | 0.0656 | 100 | 4.8915 |
266
+ | 0.0722 | 110 | 5.031 |
267
+ | 0.0787 | 120 | 4.9899 |
268
+ | 0.0853 | 130 | 4.835 |
269
+ | 0.0919 | 140 | 4.9621 |
270
+ | 0.0984 | 150 | 4.8343 |
271
+ | 0.1050 | 160 | 4.9046 |
272
+ | 0.1115 | 170 | 4.7145 |
273
+ | 0.1181 | 180 | 4.799 |
274
+ | 0.1247 | 190 | 4.6537 |
275
+ | 0.1312 | 200 | 4.619 |
276
+ | 0.1378 | 210 | 4.5875 |
277
+ | 0.1444 | 220 | 4.5914 |
278
+ | 0.1509 | 230 | 4.5139 |
279
+ | 0.1575 | 240 | 4.4542 |
280
+ | 0.1640 | 250 | 4.3606 |
281
+ | 0.1706 | 260 | 4.3778 |
282
+ | 0.1772 | 270 | 4.3027 |
283
+ | 0.1837 | 280 | 4.2375 |
284
+ | 0.1903 | 290 | 4.1997 |
285
+ | 0.1969 | 300 | 4.1519 |
286
+ | 0.2034 | 310 | 4.0318 |
287
+ | 0.2100 | 320 | 3.9845 |
288
+ | 0.2165 | 330 | 3.9522 |
289
+ | 0.2231 | 340 | 3.8775 |
290
+ | 0.2297 | 350 | 3.852 |
291
+ | 0.2362 | 360 | 3.7913 |
292
+ | 0.2428 | 370 | 3.718 |
293
+ | 0.2493 | 380 | 3.6734 |
294
+ | 0.2559 | 390 | 3.5953 |
295
+ | 0.2625 | 400 | 3.5469 |
296
+ | 0.2690 | 410 | 3.5365 |
297
+ | 0.2756 | 420 | 3.3974 |
298
+ | 0.2822 | 430 | 3.3869 |
299
+ | 0.2887 | 440 | 3.3836 |
300
+ | 0.2953 | 450 | 3.3066 |
301
+ | 0.3018 | 460 | 3.2457 |
302
+ | 0.3084 | 470 | 3.1955 |
303
+ | 0.3150 | 480 | 3.1353 |
304
+ | 0.3215 | 490 | 3.0846 |
305
+ | 0.3281 | 500 | 3.0547 |
306
+ | 0.3346 | 510 | 2.925 |
307
+ | 0.3412 | 520 | 2.894 |
308
+ | 0.3478 | 530 | 2.827 |
309
+ | 0.3543 | 540 | 2.7675 |
310
+ | 0.3609 | 550 | 2.7378 |
311
+ | 0.3675 | 560 | 2.7029 |
312
+ | 0.3740 | 570 | 2.6941 |
313
+ | 0.3806 | 580 | 2.6109 |
314
+ | 0.3871 | 590 | 2.559 |
315
+ | 0.3937 | 600 | 2.5034 |
316
+ | 0.4003 | 610 | 2.468 |
317
+ | 0.4068 | 620 | 2.3859 |
318
+ | 0.4134 | 630 | 2.3422 |
319
+ | 0.4199 | 640 | 2.3226 |
320
+ | 0.4265 | 650 | 2.2064 |
321
+ | 0.4331 | 660 | 2.1447 |
322
+ | 0.4396 | 670 | 2.1366 |
323
+ | 0.4462 | 680 | 2.0387 |
324
+ | 0.4528 | 690 | 2.0422 |
325
+ | 0.4593 | 700 | 2.0372 |
326
+ | 0.4659 | 710 | 1.943 |
327
+ | 0.4724 | 720 | 1.9419 |
328
+ | 0.4790 | 730 | 1.8448 |
329
+ | 0.4856 | 740 | 1.8012 |
330
+ | 0.4921 | 750 | 1.7876 |
331
+ | 0.4987 | 760 | 1.7801 |
332
+ | 0.5052 | 770 | 1.7242 |
333
+ | 0.5118 | 780 | 1.675 |
334
+ | 0.5184 | 790 | 1.6643 |
335
+ | 0.5249 | 800 | 1.5671 |
336
+ | 0.5315 | 810 | 1.5788 |
337
+ | 0.5381 | 820 | 1.5244 |
338
+ | 0.5446 | 830 | 1.4709 |
339
+ | 0.5512 | 840 | 1.4927 |
340
+ | 0.5577 | 850 | 1.4416 |
341
+ | 0.5643 | 860 | 1.3888 |
342
+ | 0.5709 | 870 | 1.3499 |
343
+ | 0.5774 | 880 | 1.3941 |
344
+ | 0.5840 | 890 | 1.3458 |
345
+ | 0.5906 | 900 | 1.272 |
346
+ | 0.5971 | 910 | 1.3102 |
347
+ | 0.6037 | 920 | 1.2406 |
348
+ | 0.6102 | 930 | 1.2744 |
349
+ | 0.6168 | 940 | 1.193 |
350
+ | 0.6234 | 950 | 1.1719 |
351
+ | 0.6299 | 960 | 1.1651 |
352
+ | 0.6365 | 970 | 1.1368 |
353
+ | 0.6430 | 980 | 1.1108 |
354
+ | 0.6496 | 990 | 1.078 |
355
+ | 0.6562 | 1000 | 1.0485 |
356
+ | 0.6627 | 1010 | 1.077 |
357
+ | 0.6693 | 1020 | 1.0684 |
358
+ | 0.6759 | 1030 | 1.0077 |
359
+ | 0.6824 | 1040 | 1.0239 |
360
+ | 0.6890 | 1050 | 1.0173 |
361
+ | 0.6955 | 1060 | 0.9564 |
362
+ | 0.7021 | 1070 | 0.9656 |
363
+ | 0.7087 | 1080 | 0.931 |
364
+ | 0.7152 | 1090 | 0.9259 |
365
+ | 0.7218 | 1100 | 0.9225 |
366
+ | 0.7283 | 1110 | 0.9269 |
367
+ | 0.7349 | 1120 | 0.9215 |
368
+ | 0.7415 | 1130 | 0.9313 |
369
+ | 0.7480 | 1140 | 0.8937 |
370
+ | 0.7546 | 1150 | 0.8362 |
371
+ | 0.7612 | 1160 | 0.8669 |
372
+ | 0.7677 | 1170 | 0.8452 |
373
+ | 0.7743 | 1180 | 0.8577 |
374
+ | 0.7808 | 1190 | 0.8658 |
375
+ | 0.7874 | 1200 | 0.8164 |
376
+ | 0.7940 | 1210 | 0.7835 |
377
+ | 0.8005 | 1220 | 0.8487 |
378
+ | 0.8071 | 1230 | 0.8133 |
379
+ | 0.8136 | 1240 | 0.7967 |
380
+ | 0.8202 | 1250 | 0.7979 |
381
+ | 0.8268 | 1260 | 0.7847 |
382
+ | 0.8333 | 1270 | 0.7839 |
383
+ | 0.8399 | 1280 | 0.7853 |
384
+ | 0.8465 | 1290 | 0.8109 |
385
+ | 0.8530 | 1300 | 0.795 |
386
+ | 0.8596 | 1310 | 0.7547 |
387
+ | 0.8661 | 1320 | 0.7476 |
388
+ | 0.8727 | 1330 | 0.7205 |
389
+ | 0.8793 | 1340 | 0.7456 |
390
+ | 0.8858 | 1350 | 0.7237 |
391
+ | 0.8924 | 1360 | 0.7035 |
392
+ | 0.8990 | 1370 | 0.7214 |
393
+ | 0.9055 | 1380 | 0.7065 |
394
+ | 0.9121 | 1390 | 0.6964 |
395
+ | 0.9186 | 1400 | 0.7121 |
396
+ | 0.9252 | 1410 | 0.6969 |
397
+ | 0.9318 | 1420 | 0.6862 |
398
+ | 0.9383 | 1430 | 0.685 |
399
+ | 0.9449 | 1440 | 0.7056 |
400
+ | 0.9514 | 1450 | 0.6949 |
401
+ | 0.9580 | 1460 | 0.6807 |
402
+ | 0.9646 | 1470 | 0.6497 |
403
+ | 0.9711 | 1480 | 0.666 |
404
+ | 0.9777 | 1490 | 0.6498 |
405
+ | 0.9843 | 1500 | 0.6875 |
406
+ | 0.9908 | 1510 | 0.656 |
407
+ | 0.9974 | 1520 | 0.6613 |
408
+ | 1.0039 | 1530 | 0.6246 |
409
+ | 1.0105 | 1540 | 0.655 |
410
+ | 1.0171 | 1550 | 0.657 |
411
+ | 1.0236 | 1560 | 0.6602 |
412
+ | 1.0302 | 1570 | 0.6445 |
413
+ | 1.0367 | 1580 | 0.6322 |
414
+ | 1.0433 | 1590 | 0.6166 |
415
+ | 1.0499 | 1600 | 0.6297 |
416
+ | 1.0564 | 1610 | 0.6116 |
417
+ | 1.0630 | 1620 | 0.6136 |
418
+ | 1.0696 | 1630 | 0.5916 |
419
+ | 1.0761 | 1640 | 0.5943 |
420
+ | 1.0827 | 1650 | 0.5993 |
421
+ | 1.0892 | 1660 | 0.6135 |
422
+ | 1.0958 | 1670 | 0.6004 |
423
+ | 1.1024 | 1680 | 0.6194 |
424
+ | 1.1089 | 1690 | 0.6129 |
425
+ | 1.1155 | 1700 | 0.6048 |
426
+ | 1.1220 | 1710 | 0.5834 |
427
+ | 1.1286 | 1720 | 0.5922 |
428
+ | 1.1352 | 1730 | 0.5785 |
429
+ | 1.1417 | 1740 | 0.5823 |
430
+ | 1.1483 | 1750 | 0.5734 |
431
+ | 1.1549 | 1760 | 0.5811 |
432
+ | 1.1614 | 1770 | 0.5692 |
433
+ | 1.1680 | 1780 | 0.5707 |
434
+ | 1.1745 | 1790 | 0.5548 |
435
+ | 1.1811 | 1800 | 0.5642 |
436
+ | 1.1877 | 1810 | 0.5833 |
437
+ | 1.1942 | 1820 | 0.5718 |
438
+ | 1.2008 | 1830 | 0.566 |
439
+ | 1.2073 | 1840 | 0.5551 |
440
+ | 1.2139 | 1850 | 0.5514 |
441
+ | 1.2205 | 1860 | 0.5401 |
442
+ | 1.2270 | 1870 | 0.5588 |
443
+ | 1.2336 | 1880 | 0.5573 |
444
+ | 1.2402 | 1890 | 0.5503 |
445
+ | 1.2467 | 1900 | 0.5557 |
446
+ | 1.2533 | 1910 | 0.5424 |
447
+ | 1.2598 | 1920 | 0.5466 |
448
+ | 1.2664 | 1930 | 0.5508 |
449
+ | 1.2730 | 1940 | 0.5467 |
450
+ | 1.2795 | 1950 | 0.5132 |
451
+ | 1.2861 | 1960 | 0.5291 |
452
+ | 1.2927 | 1970 | 0.5243 |
453
+ | 1.2992 | 1980 | 0.5275 |
454
+ | 1.3058 | 1990 | 0.5483 |
455
+ | 1.3123 | 2000 | 0.522 |
456
+ | 1.3189 | 2010 | 0.5147 |
457
+ | 1.3255 | 2020 | 0.5062 |
458
+ | 1.3320 | 2030 | 0.5119 |
459
+ | 1.3386 | 2040 | 0.5184 |
460
+ | 1.3451 | 2050 | 0.5039 |
461
+ | 1.3517 | 2060 | 0.5283 |
462
+ | 1.3583 | 2070 | 0.5046 |
463
+ | 1.3648 | 2080 | 0.5343 |
464
+ | 1.3714 | 2090 | 0.4888 |
465
+ | 1.3780 | 2100 | 0.5241 |
466
+ | 1.3845 | 2110 | 0.5108 |
467
+ | 1.3911 | 2120 | 0.516 |
468
+ | 1.3976 | 2130 | 0.5101 |
469
+ | 1.4042 | 2140 | 0.5246 |
470
+ | 1.4108 | 2150 | 0.5045 |
471
+ | 1.4173 | 2160 | 0.5161 |
472
+ | 1.4239 | 2170 | 0.4788 |
473
+ | 1.4304 | 2180 | 0.5106 |
474
+ | 1.4370 | 2190 | 0.476 |
475
+ | 1.4436 | 2200 | 0.4804 |
476
+ | 1.4501 | 2210 | 0.4874 |
477
+ | 1.4567 | 2220 | 0.4802 |
478
+ | 1.4633 | 2230 | 0.5203 |
479
+ | 1.4698 | 2240 | 0.4944 |
480
+ | 1.4764 | 2250 | 0.4876 |
481
+ | 1.4829 | 2260 | 0.4912 |
482
+ | 1.4895 | 2270 | 0.476 |
483
+ | 1.4961 | 2280 | 0.4859 |
484
+ | 1.5026 | 2290 | 0.4505 |
485
+ | 1.5092 | 2300 | 0.4949 |
486
+ | 1.5157 | 2310 | 0.4947 |
487
+ | 1.5223 | 2320 | 0.4726 |
488
+ | 1.5289 | 2330 | 0.4549 |
489
+ | 1.5354 | 2340 | 0.4434 |
490
+ | 1.5420 | 2350 | 0.4546 |
491
+ | 1.5486 | 2360 | 0.4513 |
492
+ | 1.5551 | 2370 | 0.4672 |
493
+ | 1.5617 | 2380 | 0.4639 |
494
+ | 1.5682 | 2390 | 0.4575 |
495
+ | 1.5748 | 2400 | 0.4719 |
496
+ | 1.5814 | 2410 | 0.469 |
497
+ | 1.5879 | 2420 | 0.4521 |
498
+ | 1.5945 | 2430 | 0.4529 |
499
+ | 1.6010 | 2440 | 0.4834 |
500
+ | 1.6076 | 2450 | 0.4621 |
501
+ | 1.6142 | 2460 | 0.463 |
502
+ | 1.6207 | 2470 | 0.4837 |
503
+ | 1.6273 | 2480 | 0.4687 |
504
+ | 1.6339 | 2490 | 0.463 |
505
+ | 1.6404 | 2500 | 0.4518 |
506
+ | 1.6470 | 2510 | 0.4356 |
507
+ | 1.6535 | 2520 | 0.4459 |
508
+ | 1.6601 | 2530 | 0.4746 |
509
+ | 1.6667 | 2540 | 0.4518 |
510
+ | 1.6732 | 2550 | 0.4363 |
511
+ | 1.6798 | 2560 | 0.4604 |
512
+ | 1.6864 | 2570 | 0.4481 |
513
+ | 1.6929 | 2580 | 0.4525 |
514
+ | 1.6995 | 2590 | 0.4421 |
515
+ | 1.7060 | 2600 | 0.4265 |
516
+ | 1.7126 | 2610 | 0.4366 |
517
+ | 1.7192 | 2620 | 0.4444 |
518
+ | 1.7257 | 2630 | 0.4632 |
519
+ | 1.7323 | 2640 | 0.4587 |
520
+ | 1.7388 | 2650 | 0.4421 |
521
+ | 1.7454 | 2660 | 0.444 |
522
+ | 1.7520 | 2670 | 0.432 |
523
+ | 1.7585 | 2680 | 0.4491 |
524
+ | 1.7651 | 2690 | 0.4375 |
525
+ | 1.7717 | 2700 | 0.4425 |
526
+ | 1.7782 | 2710 | 0.4448 |
527
+ | 1.7848 | 2720 | 0.4526 |
528
+ | 1.7913 | 2730 | 0.4277 |
529
+ | 1.7979 | 2740 | 0.4399 |
530
+ | 1.8045 | 2750 | 0.4287 |
531
+ | 1.8110 | 2760 | 0.4185 |
532
+ | 1.8176 | 2770 | 0.4332 |
533
+ | 1.8241 | 2780 | 0.4272 |
534
+ | 1.8307 | 2790 | 0.4358 |
535
+ | 1.8373 | 2800 | 0.4337 |
536
+ | 1.8438 | 2810 | 0.4259 |
537
+ | 1.8504 | 2820 | 0.4497 |
538
+ | 1.8570 | 2830 | 0.4191 |
539
+ | 1.8635 | 2840 | 0.4286 |
540
+ | 1.8701 | 2850 | 0.4492 |
541
+ | 1.8766 | 2860 | 0.4306 |
542
+ | 1.8832 | 2870 | 0.4337 |
543
+ | 1.8898 | 2880 | 0.4314 |
544
+ | 1.8963 | 2890 | 0.448 |
545
+ | 1.9029 | 2900 | 0.4277 |
546
+ | 1.9094 | 2910 | 0.424 |
547
+ | 1.9160 | 2920 | 0.428 |
548
+ | 1.9226 | 2930 | 0.4297 |
549
+ | 1.9291 | 2940 | 0.4316 |
550
+ | 1.9357 | 2950 | 0.4436 |
551
+ | 1.9423 | 2960 | 0.4071 |
552
+ | 1.9488 | 2970 | 0.4108 |
553
+ | 1.9554 | 2980 | 0.4073 |
554
+ | 1.9619 | 2990 | 0.4289 |
555
+ | 1.9685 | 3000 | 0.4496 |
556
+ | 1.9751 | 3010 | 0.4236 |
557
+ | 1.9816 | 3020 | 0.4381 |
558
+ | 1.9882 | 3030 | 0.4109 |
559
+ | 1.9948 | 3040 | 0.4152 |
560
+ | 2.0013 | 3050 | 0.4137 |
561
+ | 2.0079 | 3060 | 0.3986 |
562
+ | 2.0144 | 3070 | 0.3965 |
563
+ | 2.0210 | 3080 | 0.4313 |
564
+ | 2.0276 | 3090 | 0.4331 |
565
+ | 2.0341 | 3100 | 0.4285 |
566
+ | 2.0407 | 3110 | 0.4441 |
567
+ | 2.0472 | 3120 | 0.4243 |
568
+ | 2.0538 | 3130 | 0.4157 |
569
+ | 2.0604 | 3140 | 0.4086 |
570
+ | 2.0669 | 3150 | 0.4282 |
571
+ | 2.0735 | 3160 | 0.4152 |
572
+ | 2.0801 | 3170 | 0.4172 |
573
+ | 2.0866 | 3180 | 0.4288 |
574
+ | 2.0932 | 3190 | 0.4062 |
575
+ | 2.0997 | 3200 | 0.4116 |
576
+ | 2.1063 | 3210 | 0.4197 |
577
+ | 2.1129 | 3220 | 0.4071 |
578
+ | 2.1194 | 3230 | 0.4073 |
579
+ | 2.1260 | 3240 | 0.4109 |
580
+ | 2.1325 | 3250 | 0.4383 |
581
+ | 2.1391 | 3260 | 0.4108 |
582
+ | 2.1457 | 3270 | 0.4011 |
583
+ | 2.1522 | 3280 | 0.4035 |
584
+ | 2.1588 | 3290 | 0.4307 |
585
+ | 2.1654 | 3300 | 0.4208 |
586
+ | 2.1719 | 3310 | 0.4041 |
587
+ | 2.1785 | 3320 | 0.3979 |
588
+ | 2.1850 | 3330 | 0.4002 |
589
+ | 2.1916 | 3340 | 0.4056 |
590
+ | 2.1982 | 3350 | 0.4198 |
591
+ | 2.2047 | 3360 | 0.4036 |
592
+ | 2.2113 | 3370 | 0.4353 |
593
+ | 2.2178 | 3380 | 0.4239 |
594
+ | 2.2244 | 3390 | 0.4004 |
595
+ | 2.2310 | 3400 | 0.4006 |
596
+ | 2.2375 | 3410 | 0.4101 |
597
+ | 2.2441 | 3420 | 0.4269 |
598
+ | 2.2507 | 3430 | 0.4261 |
599
+ | 2.2572 | 3440 | 0.4173 |
600
+ | 2.2638 | 3450 | 0.4402 |
601
+ | 2.2703 | 3460 | 0.4145 |
602
+ | 2.2769 | 3470 | 0.4161 |
603
+ | 2.2835 | 3480 | 0.4261 |
604
+ | 2.2900 | 3490 | 0.4079 |
605
+ | 2.2966 | 3500 | 0.3978 |
606
+ | 2.3031 | 3510 | 0.3916 |
607
+ | 2.3097 | 3520 | 0.4072 |
608
+ | 2.3163 | 3530 | 0.3966 |
609
+ | 2.3228 | 3540 | 0.3942 |
610
+ | 2.3294 | 3550 | 0.4018 |
611
+ | 2.3360 | 3560 | 0.3935 |
612
+ | 2.3425 | 3570 | 0.3886 |
613
+ | 2.3491 | 3580 | 0.4145 |
614
+ | 2.3556 | 3590 | 0.4053 |
615
+ | 2.3622 | 3600 | 0.3955 |
616
+ | 2.3688 | 3610 | 0.3994 |
617
+ | 2.3753 | 3620 | 0.4014 |
618
+ | 2.3819 | 3630 | 0.4036 |
619
+ | 2.3885 | 3640 | 0.407 |
620
+ | 2.3950 | 3650 | 0.3799 |
621
+ | 2.4016 | 3660 | 0.4 |
622
+ | 2.4081 | 3670 | 0.3964 |
623
+ | 2.4147 | 3680 | 0.4073 |
624
+ | 2.4213 | 3690 | 0.3909 |
625
+ | 2.4278 | 3700 | 0.4277 |
626
+ | 2.4344 | 3710 | 0.3893 |
627
+ | 2.4409 | 3720 | 0.3876 |
628
+ | 2.4475 | 3730 | 0.3905 |
629
+ | 2.4541 | 3740 | 0.4181 |
630
+ | 2.4606 | 3750 | 0.4039 |
631
+ | 2.4672 | 3760 | 0.4087 |
632
+ | 2.4738 | 3770 | 0.4034 |
633
+ | 2.4803 | 3780 | 0.4166 |
634
+ | 2.4869 | 3790 | 0.3807 |
635
+ | 2.4934 | 3800 | 0.4094 |
636
+ | 2.5 | 3810 | 0.4192 |
637
+ | 2.5066 | 3820 | 0.4167 |
638
+ | 2.5131 | 3830 | 0.3972 |
639
+ | 2.5197 | 3840 | 0.379 |
640
+ | 2.5262 | 3850 | 0.4003 |
641
+ | 2.5328 | 3860 | 0.3856 |
642
+ | 2.5394 | 3870 | 0.389 |
643
+ | 2.5459 | 3880 | 0.4066 |
644
+ | 2.5525 | 3890 | 0.3891 |
645
+ | 2.5591 | 3900 | 0.3839 |
646
+ | 2.5656 | 3910 | 0.4188 |
647
+ | 2.5722 | 3920 | 0.3821 |
648
+ | 2.5787 | 3930 | 0.4134 |
649
+ | 2.5853 | 3940 | 0.4149 |
650
+ | 2.5919 | 3950 | 0.4123 |
651
+ | 2.5984 | 3960 | 0.3925 |
652
+ | 2.6050 | 3970 | 0.4189 |
653
+ | 2.6115 | 3980 | 0.4144 |
654
+ | 2.6181 | 3990 | 0.4001 |
655
+ | 2.6247 | 4000 | 0.3972 |
656
+ | 2.6312 | 4010 | 0.3868 |
657
+ | 2.6378 | 4020 | 0.3963 |
658
+ | 2.6444 | 4030 | 0.4155 |
659
+ | 2.6509 | 4040 | 0.4055 |
660
+ | 2.6575 | 4050 | 0.3961 |
661
+ | 2.6640 | 4060 | 0.4101 |
662
+ | 2.6706 | 4070 | 0.396 |
663
+ | 2.6772 | 4080 | 0.3872 |
664
+ | 2.6837 | 4090 | 0.386 |
665
+ | 2.6903 | 4100 | 0.3717 |
666
+ | 2.6969 | 4110 | 0.397 |
667
+ | 2.7034 | 4120 | 0.4023 |
668
+ | 2.7100 | 4130 | 0.4019 |
669
+ | 2.7165 | 4140 | 0.4095 |
670
+ | 2.7231 | 4150 | 0.4092 |
671
+ | 2.7297 | 4160 | 0.4066 |
672
+ | 2.7362 | 4170 | 0.396 |
673
+ | 2.7428 | 4180 | 0.3928 |
674
+ | 2.7493 | 4190 | 0.393 |
675
+ | 2.7559 | 4200 | 0.3986 |
676
+ | 2.7625 | 4210 | 0.3779 |
677
+ | 2.7690 | 4220 | 0.3917 |
678
+ | 2.7756 | 4230 | 0.3849 |
679
+ | 2.7822 | 4240 | 0.3947 |
680
+ | 2.7887 | 4250 | 0.4006 |
681
+ | 2.7953 | 4260 | 0.4004 |
682
+ | 2.8018 | 4270 | 0.3827 |
683
+ | 2.8084 | 4280 | 0.3976 |
684
+ | 2.8150 | 4290 | 0.3877 |
685
+ | 2.8215 | 4300 | 0.3898 |
686
+ | 2.8281 | 4310 | 0.4024 |
687
+ | 2.8346 | 4320 | 0.3987 |
688
+ | 2.8412 | 4330 | 0.3911 |
689
+ | 2.8478 | 4340 | 0.3928 |
690
+ | 2.8543 | 4350 | 0.3822 |
691
+ | 2.8609 | 4360 | 0.3747 |
692
+ | 2.8675 | 4370 | 0.3974 |
693
+ | 2.8740 | 4380 | 0.3851 |
694
+ | 2.8806 | 4390 | 0.3983 |
695
+ | 2.8871 | 4400 | 0.417 |
696
+ | 2.8937 | 4410 | 0.4063 |
697
+ | 2.9003 | 4420 | 0.4019 |
698
+ | 2.9068 | 4430 | 0.4062 |
699
+ | 2.9134 | 4440 | 0.3901 |
700
+ | 2.9199 | 4450 | 0.3877 |
701
+ | 2.9265 | 4460 | 0.3725 |
702
+ | 2.9331 | 4470 | 0.3931 |
703
+ | 2.9396 | 4480 | 0.3822 |
704
+ | 2.9462 | 4490 | 0.3925 |
705
+ | 2.9528 | 4500 | 0.4192 |
706
+ | 2.9593 | 4510 | 0.3881 |
707
+ | 2.9659 | 4520 | 0.4004 |
708
+ | 2.9724 | 4530 | 0.4037 |
709
+ | 2.9790 | 4540 | 0.3993 |
710
+ | 2.9856 | 4550 | 0.4212 |
711
+ | 2.9921 | 4560 | 0.3766 |
712
+ | 2.9987 | 4570 | 0.3995 |
713
+
714
+ </details>
715
+
716
+ ### Framework Versions
717
+ - Python: 3.13.9
718
+ - Sentence Transformers: 5.2.0
719
+ - Transformers: 4.57.3
720
+ - PyTorch: 2.9.1+cu128
721
+ - Accelerate: 1.12.0
722
+ - Datasets: 4.3.0
723
+ - Tokenizers: 0.22.1
724
+
725
+ ## Citation
726
+
727
+ ### BibTeX
728
+
729
+ #### Sentence Transformers
730
+ ```bibtex
731
+ @inproceedings{reimers-2019-sentence-bert,
732
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
733
+ author = "Reimers, Nils and Gurevych, Iryna",
734
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
735
+ month = "11",
736
+ year = "2019",
737
+ publisher = "Association for Computational Linguistics",
738
+ url = "https://arxiv.org/abs/1908.10084",
739
+ }
740
+ ```
741
+
742
+ #### MultipleNegativesRankingLoss
743
+ ```bibtex
744
+ @misc{henderson2017efficient,
745
+ title={Efficient Natural Language Response Suggestion for Smart Reply},
746
+ author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
747
+ year={2017},
748
+ eprint={1705.00652},
749
+ archivePrefix={arXiv},
750
+ primaryClass={cs.CL}
751
+ }
752
+ ```
753
+
754
+ <!--
755
+ ## Glossary
756
+
757
+ *Clearly define terms in order to be accessible across audiences.*
758
+ -->
759
+
760
+ <!--
761
+ ## Model Card Authors
762
+
763
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
764
+ -->
765
+
766
+ <!--
767
+ ## Model Card Contact
768
+
769
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
770
+ -->
benchmark.png ADDED

Git LFS Details

  • SHA256: 5c025b5691db5524760e312eb8ffc4246a1fe84889a85563785efc94d005b5b0
  • Pointer size: 131 Bytes
  • Size of remote file: 404 kB
config.json ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "ModernBertModel"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 50281,
8
+ "classifier_activation": "gelu",
9
+ "classifier_bias": false,
10
+ "classifier_dropout": 0.0,
11
+ "classifier_pooling": "mean",
12
+ "cls_token_id": 50281,
13
+ "decoder_bias": true,
14
+ "deterministic_flash_attn": false,
15
+ "torch_dtype": "bfloat16",
16
+ "embedding_dropout": 0.0,
17
+ "eos_token_id": 50282,
18
+ "global_attn_every_n_layers": 3,
19
+ "global_rope_theta": 160000.0,
20
+ "gradient_checkpointing": false,
21
+ "hidden_activation": "gelu",
22
+ "hidden_size": 768,
23
+ "initializer_cutoff_factor": 2.0,
24
+ "initializer_range": 0.02,
25
+ "intermediate_size": 1152,
26
+ "layer_norm_eps": 1e-05,
27
+ "local_attention": 128,
28
+ "local_rope_theta": 10000.0,
29
+ "max_position_embeddings": 8192,
30
+ "mlp_bias": false,
31
+ "mlp_dropout": 0.0,
32
+ "model_type": "modernbert",
33
+ "norm_bias": false,
34
+ "norm_eps": 1e-05,
35
+ "num_attention_heads": 12,
36
+ "num_hidden_layers": 22,
37
+ "pad_token_id": 50283,
38
+ "position_embedding_type": "absolute",
39
+ "repad_logits_with_grad": false,
40
+ "sep_token_id": 50282,
41
+ "sparse_pred_ignore_index": -100,
42
+ "sparse_prediction": false,
43
+ "tokenizer_class": "PreTrainedTokenizerFast",
44
+ "transformers_version": "4.57.3",
45
+ "unsloth_version": "2025.12.9",
46
+ "vocab_size": 50368
47
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "SentenceTransformer",
3
+ "__version__": {
4
+ "sentence_transformers": "5.2.0",
5
+ "transformers": "4.57.3",
6
+ "pytorch": "2.9.1+cu128"
7
+ },
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91bdc73389563141ff956874466c38f4e0579d1d56f18d3ab653bc874a290bf1
3
+ size 298041568
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": true,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,946 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "|||IP_ADDRESS|||",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": false
10
+ },
11
+ "1": {
12
+ "content": "<|padding|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "50254": {
20
+ "content": " ",
21
+ "lstrip": false,
22
+ "normalized": true,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": false
26
+ },
27
+ "50255": {
28
+ "content": " ",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": false
34
+ },
35
+ "50256": {
36
+ "content": " ",
37
+ "lstrip": false,
38
+ "normalized": true,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": false
42
+ },
43
+ "50257": {
44
+ "content": " ",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": false
50
+ },
51
+ "50258": {
52
+ "content": " ",
53
+ "lstrip": false,
54
+ "normalized": true,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": false
58
+ },
59
+ "50259": {
60
+ "content": " ",
61
+ "lstrip": false,
62
+ "normalized": true,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": false
66
+ },
67
+ "50260": {
68
+ "content": " ",
69
+ "lstrip": false,
70
+ "normalized": true,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": false
74
+ },
75
+ "50261": {
76
+ "content": " ",
77
+ "lstrip": false,
78
+ "normalized": true,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": false
82
+ },
83
+ "50262": {
84
+ "content": " ",
85
+ "lstrip": false,
86
+ "normalized": true,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": false
90
+ },
91
+ "50263": {
92
+ "content": " ",
93
+ "lstrip": false,
94
+ "normalized": true,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": false
98
+ },
99
+ "50264": {
100
+ "content": " ",
101
+ "lstrip": false,
102
+ "normalized": true,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": false
106
+ },
107
+ "50265": {
108
+ "content": " ",
109
+ "lstrip": false,
110
+ "normalized": true,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": false
114
+ },
115
+ "50266": {
116
+ "content": " ",
117
+ "lstrip": false,
118
+ "normalized": true,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": false
122
+ },
123
+ "50267": {
124
+ "content": " ",
125
+ "lstrip": false,
126
+ "normalized": true,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": false
130
+ },
131
+ "50268": {
132
+ "content": " ",
133
+ "lstrip": false,
134
+ "normalized": true,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": false
138
+ },
139
+ "50269": {
140
+ "content": " ",
141
+ "lstrip": false,
142
+ "normalized": true,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": false
146
+ },
147
+ "50270": {
148
+ "content": " ",
149
+ "lstrip": false,
150
+ "normalized": true,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": false
154
+ },
155
+ "50271": {
156
+ "content": " ",
157
+ "lstrip": false,
158
+ "normalized": true,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": false
162
+ },
163
+ "50272": {
164
+ "content": " ",
165
+ "lstrip": false,
166
+ "normalized": true,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": false
170
+ },
171
+ "50273": {
172
+ "content": " ",
173
+ "lstrip": false,
174
+ "normalized": true,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": false
178
+ },
179
+ "50274": {
180
+ "content": " ",
181
+ "lstrip": false,
182
+ "normalized": true,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": false
186
+ },
187
+ "50275": {
188
+ "content": " ",
189
+ "lstrip": false,
190
+ "normalized": true,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": false
194
+ },
195
+ "50276": {
196
+ "content": " ",
197
+ "lstrip": false,
198
+ "normalized": true,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": false
202
+ },
203
+ "50277": {
204
+ "content": "|||EMAIL_ADDRESS|||",
205
+ "lstrip": false,
206
+ "normalized": true,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": false
210
+ },
211
+ "50278": {
212
+ "content": "|||PHONE_NUMBER|||",
213
+ "lstrip": false,
214
+ "normalized": true,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": false
218
+ },
219
+ "50279": {
220
+ "content": "<|endoftext|>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "50280": {
228
+ "content": "[UNK]",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "50281": {
236
+ "content": "[CLS]",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "50282": {
244
+ "content": "[SEP]",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "50283": {
252
+ "content": "[PAD]",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "50284": {
260
+ "content": "[MASK]",
261
+ "lstrip": true,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "50285": {
268
+ "content": "[unused0]",
269
+ "lstrip": false,
270
+ "normalized": true,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": false
274
+ },
275
+ "50286": {
276
+ "content": "[unused1]",
277
+ "lstrip": false,
278
+ "normalized": true,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": false
282
+ },
283
+ "50287": {
284
+ "content": "[unused2]",
285
+ "lstrip": false,
286
+ "normalized": true,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": false
290
+ },
291
+ "50288": {
292
+ "content": "[unused3]",
293
+ "lstrip": false,
294
+ "normalized": true,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": false
298
+ },
299
+ "50289": {
300
+ "content": "[unused4]",
301
+ "lstrip": false,
302
+ "normalized": true,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": false
306
+ },
307
+ "50290": {
308
+ "content": "[unused5]",
309
+ "lstrip": false,
310
+ "normalized": true,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": false
314
+ },
315
+ "50291": {
316
+ "content": "[unused6]",
317
+ "lstrip": false,
318
+ "normalized": true,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": false
322
+ },
323
+ "50292": {
324
+ "content": "[unused7]",
325
+ "lstrip": false,
326
+ "normalized": true,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": false
330
+ },
331
+ "50293": {
332
+ "content": "[unused8]",
333
+ "lstrip": false,
334
+ "normalized": true,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": false
338
+ },
339
+ "50294": {
340
+ "content": "[unused9]",
341
+ "lstrip": false,
342
+ "normalized": true,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": false
346
+ },
347
+ "50295": {
348
+ "content": "[unused10]",
349
+ "lstrip": false,
350
+ "normalized": true,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": false
354
+ },
355
+ "50296": {
356
+ "content": "[unused11]",
357
+ "lstrip": false,
358
+ "normalized": true,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": false
362
+ },
363
+ "50297": {
364
+ "content": "[unused12]",
365
+ "lstrip": false,
366
+ "normalized": true,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": false
370
+ },
371
+ "50298": {
372
+ "content": "[unused13]",
373
+ "lstrip": false,
374
+ "normalized": true,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": false
378
+ },
379
+ "50299": {
380
+ "content": "[unused14]",
381
+ "lstrip": false,
382
+ "normalized": true,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": false
386
+ },
387
+ "50300": {
388
+ "content": "[unused15]",
389
+ "lstrip": false,
390
+ "normalized": true,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": false
394
+ },
395
+ "50301": {
396
+ "content": "[unused16]",
397
+ "lstrip": false,
398
+ "normalized": true,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": false
402
+ },
403
+ "50302": {
404
+ "content": "[unused17]",
405
+ "lstrip": false,
406
+ "normalized": true,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": false
410
+ },
411
+ "50303": {
412
+ "content": "[unused18]",
413
+ "lstrip": false,
414
+ "normalized": true,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": false
418
+ },
419
+ "50304": {
420
+ "content": "[unused19]",
421
+ "lstrip": false,
422
+ "normalized": true,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": false
426
+ },
427
+ "50305": {
428
+ "content": "[unused20]",
429
+ "lstrip": false,
430
+ "normalized": true,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": false
434
+ },
435
+ "50306": {
436
+ "content": "[unused21]",
437
+ "lstrip": false,
438
+ "normalized": true,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": false
442
+ },
443
+ "50307": {
444
+ "content": "[unused22]",
445
+ "lstrip": false,
446
+ "normalized": true,
447
+ "rstrip": false,
448
+ "single_word": false,
449
+ "special": false
450
+ },
451
+ "50308": {
452
+ "content": "[unused23]",
453
+ "lstrip": false,
454
+ "normalized": true,
455
+ "rstrip": false,
456
+ "single_word": false,
457
+ "special": false
458
+ },
459
+ "50309": {
460
+ "content": "[unused24]",
461
+ "lstrip": false,
462
+ "normalized": true,
463
+ "rstrip": false,
464
+ "single_word": false,
465
+ "special": false
466
+ },
467
+ "50310": {
468
+ "content": "[unused25]",
469
+ "lstrip": false,
470
+ "normalized": true,
471
+ "rstrip": false,
472
+ "single_word": false,
473
+ "special": false
474
+ },
475
+ "50311": {
476
+ "content": "[unused26]",
477
+ "lstrip": false,
478
+ "normalized": true,
479
+ "rstrip": false,
480
+ "single_word": false,
481
+ "special": false
482
+ },
483
+ "50312": {
484
+ "content": "[unused27]",
485
+ "lstrip": false,
486
+ "normalized": true,
487
+ "rstrip": false,
488
+ "single_word": false,
489
+ "special": false
490
+ },
491
+ "50313": {
492
+ "content": "[unused28]",
493
+ "lstrip": false,
494
+ "normalized": true,
495
+ "rstrip": false,
496
+ "single_word": false,
497
+ "special": false
498
+ },
499
+ "50314": {
500
+ "content": "[unused29]",
501
+ "lstrip": false,
502
+ "normalized": true,
503
+ "rstrip": false,
504
+ "single_word": false,
505
+ "special": false
506
+ },
507
+ "50315": {
508
+ "content": "[unused30]",
509
+ "lstrip": false,
510
+ "normalized": true,
511
+ "rstrip": false,
512
+ "single_word": false,
513
+ "special": false
514
+ },
515
+ "50316": {
516
+ "content": "[unused31]",
517
+ "lstrip": false,
518
+ "normalized": true,
519
+ "rstrip": false,
520
+ "single_word": false,
521
+ "special": false
522
+ },
523
+ "50317": {
524
+ "content": "[unused32]",
525
+ "lstrip": false,
526
+ "normalized": true,
527
+ "rstrip": false,
528
+ "single_word": false,
529
+ "special": false
530
+ },
531
+ "50318": {
532
+ "content": "[unused33]",
533
+ "lstrip": false,
534
+ "normalized": true,
535
+ "rstrip": false,
536
+ "single_word": false,
537
+ "special": false
538
+ },
539
+ "50319": {
540
+ "content": "[unused34]",
541
+ "lstrip": false,
542
+ "normalized": true,
543
+ "rstrip": false,
544
+ "single_word": false,
545
+ "special": false
546
+ },
547
+ "50320": {
548
+ "content": "[unused35]",
549
+ "lstrip": false,
550
+ "normalized": true,
551
+ "rstrip": false,
552
+ "single_word": false,
553
+ "special": false
554
+ },
555
+ "50321": {
556
+ "content": "[unused36]",
557
+ "lstrip": false,
558
+ "normalized": true,
559
+ "rstrip": false,
560
+ "single_word": false,
561
+ "special": false
562
+ },
563
+ "50322": {
564
+ "content": "[unused37]",
565
+ "lstrip": false,
566
+ "normalized": true,
567
+ "rstrip": false,
568
+ "single_word": false,
569
+ "special": false
570
+ },
571
+ "50323": {
572
+ "content": "[unused38]",
573
+ "lstrip": false,
574
+ "normalized": true,
575
+ "rstrip": false,
576
+ "single_word": false,
577
+ "special": false
578
+ },
579
+ "50324": {
580
+ "content": "[unused39]",
581
+ "lstrip": false,
582
+ "normalized": true,
583
+ "rstrip": false,
584
+ "single_word": false,
585
+ "special": false
586
+ },
587
+ "50325": {
588
+ "content": "[unused40]",
589
+ "lstrip": false,
590
+ "normalized": true,
591
+ "rstrip": false,
592
+ "single_word": false,
593
+ "special": false
594
+ },
595
+ "50326": {
596
+ "content": "[unused41]",
597
+ "lstrip": false,
598
+ "normalized": true,
599
+ "rstrip": false,
600
+ "single_word": false,
601
+ "special": false
602
+ },
603
+ "50327": {
604
+ "content": "[unused42]",
605
+ "lstrip": false,
606
+ "normalized": true,
607
+ "rstrip": false,
608
+ "single_word": false,
609
+ "special": false
610
+ },
611
+ "50328": {
612
+ "content": "[unused43]",
613
+ "lstrip": false,
614
+ "normalized": true,
615
+ "rstrip": false,
616
+ "single_word": false,
617
+ "special": false
618
+ },
619
+ "50329": {
620
+ "content": "[unused44]",
621
+ "lstrip": false,
622
+ "normalized": true,
623
+ "rstrip": false,
624
+ "single_word": false,
625
+ "special": false
626
+ },
627
+ "50330": {
628
+ "content": "[unused45]",
629
+ "lstrip": false,
630
+ "normalized": true,
631
+ "rstrip": false,
632
+ "single_word": false,
633
+ "special": false
634
+ },
635
+ "50331": {
636
+ "content": "[unused46]",
637
+ "lstrip": false,
638
+ "normalized": true,
639
+ "rstrip": false,
640
+ "single_word": false,
641
+ "special": false
642
+ },
643
+ "50332": {
644
+ "content": "[unused47]",
645
+ "lstrip": false,
646
+ "normalized": true,
647
+ "rstrip": false,
648
+ "single_word": false,
649
+ "special": false
650
+ },
651
+ "50333": {
652
+ "content": "[unused48]",
653
+ "lstrip": false,
654
+ "normalized": true,
655
+ "rstrip": false,
656
+ "single_word": false,
657
+ "special": false
658
+ },
659
+ "50334": {
660
+ "content": "[unused49]",
661
+ "lstrip": false,
662
+ "normalized": true,
663
+ "rstrip": false,
664
+ "single_word": false,
665
+ "special": false
666
+ },
667
+ "50335": {
668
+ "content": "[unused50]",
669
+ "lstrip": false,
670
+ "normalized": true,
671
+ "rstrip": false,
672
+ "single_word": false,
673
+ "special": false
674
+ },
675
+ "50336": {
676
+ "content": "[unused51]",
677
+ "lstrip": false,
678
+ "normalized": true,
679
+ "rstrip": false,
680
+ "single_word": false,
681
+ "special": false
682
+ },
683
+ "50337": {
684
+ "content": "[unused52]",
685
+ "lstrip": false,
686
+ "normalized": true,
687
+ "rstrip": false,
688
+ "single_word": false,
689
+ "special": false
690
+ },
691
+ "50338": {
692
+ "content": "[unused53]",
693
+ "lstrip": false,
694
+ "normalized": true,
695
+ "rstrip": false,
696
+ "single_word": false,
697
+ "special": false
698
+ },
699
+ "50339": {
700
+ "content": "[unused54]",
701
+ "lstrip": false,
702
+ "normalized": true,
703
+ "rstrip": false,
704
+ "single_word": false,
705
+ "special": false
706
+ },
707
+ "50340": {
708
+ "content": "[unused55]",
709
+ "lstrip": false,
710
+ "normalized": true,
711
+ "rstrip": false,
712
+ "single_word": false,
713
+ "special": false
714
+ },
715
+ "50341": {
716
+ "content": "[unused56]",
717
+ "lstrip": false,
718
+ "normalized": true,
719
+ "rstrip": false,
720
+ "single_word": false,
721
+ "special": false
722
+ },
723
+ "50342": {
724
+ "content": "[unused57]",
725
+ "lstrip": false,
726
+ "normalized": true,
727
+ "rstrip": false,
728
+ "single_word": false,
729
+ "special": false
730
+ },
731
+ "50343": {
732
+ "content": "[unused58]",
733
+ "lstrip": false,
734
+ "normalized": true,
735
+ "rstrip": false,
736
+ "single_word": false,
737
+ "special": false
738
+ },
739
+ "50344": {
740
+ "content": "[unused59]",
741
+ "lstrip": false,
742
+ "normalized": true,
743
+ "rstrip": false,
744
+ "single_word": false,
745
+ "special": false
746
+ },
747
+ "50345": {
748
+ "content": "[unused60]",
749
+ "lstrip": false,
750
+ "normalized": true,
751
+ "rstrip": false,
752
+ "single_word": false,
753
+ "special": false
754
+ },
755
+ "50346": {
756
+ "content": "[unused61]",
757
+ "lstrip": false,
758
+ "normalized": true,
759
+ "rstrip": false,
760
+ "single_word": false,
761
+ "special": false
762
+ },
763
+ "50347": {
764
+ "content": "[unused62]",
765
+ "lstrip": false,
766
+ "normalized": true,
767
+ "rstrip": false,
768
+ "single_word": false,
769
+ "special": false
770
+ },
771
+ "50348": {
772
+ "content": "[unused63]",
773
+ "lstrip": false,
774
+ "normalized": true,
775
+ "rstrip": false,
776
+ "single_word": false,
777
+ "special": false
778
+ },
779
+ "50349": {
780
+ "content": "[unused64]",
781
+ "lstrip": false,
782
+ "normalized": true,
783
+ "rstrip": false,
784
+ "single_word": false,
785
+ "special": false
786
+ },
787
+ "50350": {
788
+ "content": "[unused65]",
789
+ "lstrip": false,
790
+ "normalized": true,
791
+ "rstrip": false,
792
+ "single_word": false,
793
+ "special": false
794
+ },
795
+ "50351": {
796
+ "content": "[unused66]",
797
+ "lstrip": false,
798
+ "normalized": true,
799
+ "rstrip": false,
800
+ "single_word": false,
801
+ "special": false
802
+ },
803
+ "50352": {
804
+ "content": "[unused67]",
805
+ "lstrip": false,
806
+ "normalized": true,
807
+ "rstrip": false,
808
+ "single_word": false,
809
+ "special": false
810
+ },
811
+ "50353": {
812
+ "content": "[unused68]",
813
+ "lstrip": false,
814
+ "normalized": true,
815
+ "rstrip": false,
816
+ "single_word": false,
817
+ "special": false
818
+ },
819
+ "50354": {
820
+ "content": "[unused69]",
821
+ "lstrip": false,
822
+ "normalized": true,
823
+ "rstrip": false,
824
+ "single_word": false,
825
+ "special": false
826
+ },
827
+ "50355": {
828
+ "content": "[unused70]",
829
+ "lstrip": false,
830
+ "normalized": true,
831
+ "rstrip": false,
832
+ "single_word": false,
833
+ "special": false
834
+ },
835
+ "50356": {
836
+ "content": "[unused71]",
837
+ "lstrip": false,
838
+ "normalized": true,
839
+ "rstrip": false,
840
+ "single_word": false,
841
+ "special": false
842
+ },
843
+ "50357": {
844
+ "content": "[unused72]",
845
+ "lstrip": false,
846
+ "normalized": true,
847
+ "rstrip": false,
848
+ "single_word": false,
849
+ "special": false
850
+ },
851
+ "50358": {
852
+ "content": "[unused73]",
853
+ "lstrip": false,
854
+ "normalized": true,
855
+ "rstrip": false,
856
+ "single_word": false,
857
+ "special": false
858
+ },
859
+ "50359": {
860
+ "content": "[unused74]",
861
+ "lstrip": false,
862
+ "normalized": true,
863
+ "rstrip": false,
864
+ "single_word": false,
865
+ "special": false
866
+ },
867
+ "50360": {
868
+ "content": "[unused75]",
869
+ "lstrip": false,
870
+ "normalized": true,
871
+ "rstrip": false,
872
+ "single_word": false,
873
+ "special": false
874
+ },
875
+ "50361": {
876
+ "content": "[unused76]",
877
+ "lstrip": false,
878
+ "normalized": true,
879
+ "rstrip": false,
880
+ "single_word": false,
881
+ "special": false
882
+ },
883
+ "50362": {
884
+ "content": "[unused77]",
885
+ "lstrip": false,
886
+ "normalized": true,
887
+ "rstrip": false,
888
+ "single_word": false,
889
+ "special": false
890
+ },
891
+ "50363": {
892
+ "content": "[unused78]",
893
+ "lstrip": false,
894
+ "normalized": true,
895
+ "rstrip": false,
896
+ "single_word": false,
897
+ "special": false
898
+ },
899
+ "50364": {
900
+ "content": "[unused79]",
901
+ "lstrip": false,
902
+ "normalized": true,
903
+ "rstrip": false,
904
+ "single_word": false,
905
+ "special": false
906
+ },
907
+ "50365": {
908
+ "content": "[unused80]",
909
+ "lstrip": false,
910
+ "normalized": true,
911
+ "rstrip": false,
912
+ "single_word": false,
913
+ "special": false
914
+ },
915
+ "50366": {
916
+ "content": "[unused81]",
917
+ "lstrip": false,
918
+ "normalized": true,
919
+ "rstrip": false,
920
+ "single_word": false,
921
+ "special": false
922
+ },
923
+ "50367": {
924
+ "content": "[unused82]",
925
+ "lstrip": false,
926
+ "normalized": true,
927
+ "rstrip": false,
928
+ "single_word": false,
929
+ "special": false
930
+ }
931
+ },
932
+ "clean_up_tokenization_spaces": true,
933
+ "cls_token": "[CLS]",
934
+ "extra_special_tokens": {},
935
+ "mask_token": "[MASK]",
936
+ "model_input_names": [
937
+ "input_ids",
938
+ "attention_mask"
939
+ ],
940
+ "model_max_length": 1000000000000000019884624838656,
941
+ "pad_token": "[PAD]",
942
+ "padding_side": "left",
943
+ "sep_token": "[SEP]",
944
+ "tokenizer_class": "PreTrainedTokenizerFast",
945
+ "unk_token": "[UNK]"
946
+ }