JohanHeinsen commited on
Commit
59638dd
·
verified ·
1 Parent(s): db0d238

Add new SentenceTransformer model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,425 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - generated_from_trainer
7
+ - dataset_size:83246
8
+ - loss:CosineSimilarityLoss
9
+ base_model: CALDISS-AAU/DA-BERT_Old_News_V3
10
+ widget:
11
+ - source_sentence: og jeg derefter er underrettet, at han i Mandags Morges Kl. 5,
12
+ skal have paa en tyvagtig Maade krøbet over mit Plankeværk ind i min Gaard gaaet
13
+ ind i sit Kammer, taget endeel Sager, og derefter er han igien gaaet bort, uden
14
+ at nogen veed hvor han er bleven af, saa nødes jeg herved at efterlyse bemeldte
15
+ Niels Stephensen, med Anmodning at han vil indfinde sig, for at giøre Rigtighed
16
+ for anseelige Penge og videre som ham er betroet
17
+ sentences:
18
+ - saa bliver han end videre herved efterlyst, med Erindring til alle og enhver,
19
+ ikke enten at huse, hæle eller i Tieneste antage denne Tyv, men de hannem maatte
20
+ forekomme, anmode hannem at anholde og mig derom notificere
21
+ - ombedes ingen at huse eller hæle med ham, ei heller at antage ham i nogen Slags
22
+ Tieneste
23
+ - saa ombedes alle og enhver, om bemeldte Pige skulde forlange noget i mit Navn,
24
+ hun da ei maa betroes
25
+ - source_sentence: da han ved alle Leyligheder, og Steder her i Landet er eftersøgt,
26
+ men ey bleven funden eller nogen Oplysning om ham faaet
27
+ sentences:
28
+ - og hans Broder Hans Witt, der skal opholde sig i Sielland,
29
+ - er Natten imellem den 6te og 2de et Mai sidstleden undvigt Reserven Jørnen Andersen
30
+ Dannemarre.
31
+ - En skikkelig Karl søger Gaardskarls Plads;
32
+ - source_sentence: uden at hans Opholdsted har været at opdage
33
+ sentences:
34
+ - er efter at have begaaet stor Utroskab bortløben den 12 hujus af min Tieneste
35
+ - som haver paa en utilbørlig Maade forladt Staden og Mesterens Arbeide, og er paa
36
+ Herberget og ved Laden skyldig
37
+ - igen at indfinde sig paa Districter, og søge den Tieneste, hvortil han har fæstet
38
+ sig, hos Bonden Oluf Nielsen i Pederstrup Sogn i bemeldte Ballerup
39
+ - source_sentence: En tyk undersætsig Pige ved Navn Kirsten, som har tient mig fra
40
+ Michelsdag af, er bortgaaet af sin Tieneste
41
+ sentences:
42
+ - 'saaledes undvigt begge uden allerringeste beføyet Aarsag nemlig: den Ene navnlig
43
+ Ole Espensen af Anders Andersens Læg i Høng, 23 a 24 Aar gammel [...] og den Anden
44
+ navnlig Niels Hansen [...] aldeles uden mindste Føye af sin Tieneste undvigt og
45
+ vides ikke hvor er'
46
+ - den Mulct den sig paadrager, i Følge Forordningen af 4de Februar. 1733. den 19de
47
+ Post som slig Bortrømt huser eller hæle
48
+ - indkaldes herved sub pæna præclusi & perpetui silentii, med Aar og Dags Varsel
49
+ alle Dem som formeene sig at have noget at fordre hos S. T. Hr. Assessor Mathias
50
+ Leth Sommerhielm, eller i hens Boe efter hans ved Døden afgangne Hustrue, Frue
51
+ Friderikke Dorothea Sehestedt paa Tomb Gaard i Smaalenenes Amt, af hvad Beskaffenhed
52
+ og under hvad Titul deres Fordringer end maatte være der ved at fremkomme og saadant
53
+ deres havende Krav inden forbemeldte Tids Forløb for ham at anmelde og beviisliggiøre.
54
+ - source_sentence: saa finde vi os beføyet, saavel for at besørge de publiqve Midlers
55
+ Stkkerhed, som i Anledning af den Nød og Armod hans Undvigelse har foraarsaget
56
+ hans efterladte Kone og smaa Børn her, at lade ham herved eftelyse og indkalde
57
+ til at indfinde sig her Byen
58
+ sentences:
59
+ - men ingen veed om ham at sige.
60
+ - Friseur W. har logeret hos mig i lille Kongensgade og absenteret sig i April Maaned,
61
+ samt efterladt sig en luk Kuffert, saa bedes han at komme til mig inden 8 Dage
62
+ for at gjøre Rigtighed, hvis ikke, sættes hans efterladte Tøi paa Auction og forbeholdes
63
+ min Ret.
64
+ - thi anmodes enhver som træffer denne Karl at paagribe ham og imod de sædvanlige
65
+ Indbringerpenge at levere ham til Compagniet.
66
+ pipeline_tag: sentence-similarity
67
+ library_name: sentence-transformers
68
+ metrics:
69
+ - pearson_cosine
70
+ - spearman_cosine
71
+ model-index:
72
+ - name: SentenceTransformer based on CALDISS-AAU/DA-BERT_Old_News_V3
73
+ results:
74
+ - task:
75
+ type: semantic-similarity
76
+ name: Semantic Similarity
77
+ dataset:
78
+ name: validation
79
+ type: validation
80
+ metrics:
81
+ - type: pearson_cosine
82
+ value: 0.8931725385239138
83
+ name: Pearson Cosine
84
+ - type: spearman_cosine
85
+ value: 0.8380469537534867
86
+ name: Spearman Cosine
87
+ ---
88
+
89
+ # SentenceTransformer based on CALDISS-AAU/DA-BERT_Old_News_V3
90
+
91
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [CALDISS-AAU/DA-BERT_Old_News_V3](https://huggingface.co/CALDISS-AAU/DA-BERT_Old_News_V3). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
92
+
93
+ ## Model Details
94
+
95
+ ### Model Description
96
+ - **Model Type:** Sentence Transformer
97
+ - **Base model:** [CALDISS-AAU/DA-BERT_Old_News_V3](https://huggingface.co/CALDISS-AAU/DA-BERT_Old_News_V3) <!-- at revision 4ee1d0adc4713b6ca089da767a2e5bbf07b3a9aa -->
98
+ - **Maximum Sequence Length:** 514 tokens
99
+ - **Output Dimensionality:** 768 dimensions
100
+ - **Similarity Function:** Cosine Similarity
101
+ <!-- - **Training Dataset:** Unknown -->
102
+ <!-- - **Language:** Unknown -->
103
+ <!-- - **License:** Unknown -->
104
+
105
+ ### Model Sources
106
+
107
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
108
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
109
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
110
+
111
+ ### Full Model Architecture
112
+
113
+ ```
114
+ SentenceTransformer(
115
+ (0): Transformer({'max_seq_length': 514, 'do_lower_case': False}) with Transformer model: XLMRobertaModel
116
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
117
+ )
118
+ ```
119
+
120
+ ## Usage
121
+
122
+ ### Direct Usage (Sentence Transformers)
123
+
124
+ First install the Sentence Transformers library:
125
+
126
+ ```bash
127
+ pip install -U sentence-transformers
128
+ ```
129
+
130
+ Then you can load this model and run inference.
131
+ ```python
132
+ from sentence_transformers import SentenceTransformer
133
+
134
+ # Download from the 🤗 Hub
135
+ model = SentenceTransformer("JohanHeinsen/Run_Away_Action_Embedding_model_v0.1")
136
+ # Run inference
137
+ sentences = [
138
+ 'saa finde vi os beføyet, saavel for at besørge de publiqve Midlers Stkkerhed, som i Anledning af den Nød og Armod hans Undvigelse har foraarsaget hans efterladte Kone og smaa Børn her, at lade ham herved eftelyse og indkalde til at indfinde sig her Byen',
139
+ 'thi anmodes enhver som træffer denne Karl at paagribe ham og imod de sædvanlige Indbringerpenge at levere ham til Compagniet.',
140
+ 'men ingen veed om ham at sige.',
141
+ ]
142
+ embeddings = model.encode(sentences)
143
+ print(embeddings.shape)
144
+ # [3, 768]
145
+
146
+ # Get the similarity scores for the embeddings
147
+ similarities = model.similarity(embeddings, embeddings)
148
+ print(similarities.shape)
149
+ # [3, 3]
150
+ ```
151
+
152
+ <!--
153
+ ### Direct Usage (Transformers)
154
+
155
+ <details><summary>Click to see the direct usage in Transformers</summary>
156
+
157
+ </details>
158
+ -->
159
+
160
+ <!--
161
+ ### Downstream Usage (Sentence Transformers)
162
+
163
+ You can finetune this model on your own dataset.
164
+
165
+ <details><summary>Click to expand</summary>
166
+
167
+ </details>
168
+ -->
169
+
170
+ <!--
171
+ ### Out-of-Scope Use
172
+
173
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
174
+ -->
175
+
176
+ ## Evaluation
177
+
178
+ ### Metrics
179
+
180
+ #### Semantic Similarity
181
+
182
+ * Dataset: `validation`
183
+ * Evaluated with [<code>EmbeddingSimilarityEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.EmbeddingSimilarityEvaluator)
184
+
185
+ | Metric | Value |
186
+ |:--------------------|:----------|
187
+ | pearson_cosine | 0.8932 |
188
+ | **spearman_cosine** | **0.838** |
189
+
190
+ <!--
191
+ ## Bias, Risks and Limitations
192
+
193
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
194
+ -->
195
+
196
+ <!--
197
+ ### Recommendations
198
+
199
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
200
+ -->
201
+
202
+ ## Training Details
203
+
204
+ ### Training Dataset
205
+
206
+ #### Unnamed Dataset
207
+
208
+ * Size: 83,246 training samples
209
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
210
+ * Approximate statistics based on the first 1000 samples:
211
+ | | sentence_0 | sentence_1 | label |
212
+ |:--------|:-----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:------------------------------------------------|
213
+ | type | string | string | int |
214
+ | details | <ul><li>min: 6 tokens</li><li>mean: 28.53 tokens</li><li>max: 176 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 30.7 tokens</li><li>max: 514 tokens</li></ul> | <ul><li>0: ~47.20%</li><li>1: ~52.80%</li></ul> |
215
+ * Samples:
216
+ | sentence_0 | sentence_1 | label |
217
+ |:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------|
218
+ | <code>Han er født i Kjøbenhavn, er 33 Aar gammel, har staaet ved Kongens Regiment som Tambour i 9 Aar, og gaaer derfor skjæbt paa det venstre Bern har tjent Hr. Fabriker Schrøder, og paa Maderup gaard som Gaardskarl, en kort Tid.</code> | <code>som tiener paa Hiørnet af gl. Mynt og Svertegaden</code> | <code>1</code> |
219
+ | <code>efterlyses han herved med Anmodning, at han, af hvem han skulde forekomme, maatte anholdes og mig derom meddeles Underretning</code> | <code>Da min Lærredreng Friderich Senggrav er den 12 Aug. undvigt af sin Lære</code> | <code>0</code> |
220
+ | <code>give det tilkiende hos A. J. Schottlænder, boende i Pilestræde Nr. 11 Litr. Dførste Sal.</code> | <code>Niels Pedersen Eistrup, som i 4 Aar har tjent hos mig for Under-Knegt, er fra mig undvigt den 24 December uden at giøre Rigtighed for det ham var betroet</code> | <code>0</code> |
221
+ * Loss: [<code>CosineSimilarityLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cosinesimilarityloss) with these parameters:
222
+ ```json
223
+ {
224
+ "loss_fct": "torch.nn.modules.loss.MSELoss"
225
+ }
226
+ ```
227
+
228
+ ### Training Hyperparameters
229
+ #### Non-Default Hyperparameters
230
+
231
+ - `per_device_train_batch_size`: 16
232
+ - `per_device_eval_batch_size`: 16
233
+ - `num_train_epochs`: 2
234
+ - `multi_dataset_batch_sampler`: round_robin
235
+
236
+ #### All Hyperparameters
237
+ <details><summary>Click to expand</summary>
238
+
239
+ - `overwrite_output_dir`: False
240
+ - `do_predict`: False
241
+ - `eval_strategy`: no
242
+ - `prediction_loss_only`: True
243
+ - `per_device_train_batch_size`: 16
244
+ - `per_device_eval_batch_size`: 16
245
+ - `per_gpu_train_batch_size`: None
246
+ - `per_gpu_eval_batch_size`: None
247
+ - `gradient_accumulation_steps`: 1
248
+ - `eval_accumulation_steps`: None
249
+ - `torch_empty_cache_steps`: None
250
+ - `learning_rate`: 5e-05
251
+ - `weight_decay`: 0.0
252
+ - `adam_beta1`: 0.9
253
+ - `adam_beta2`: 0.999
254
+ - `adam_epsilon`: 1e-08
255
+ - `max_grad_norm`: 1
256
+ - `num_train_epochs`: 2
257
+ - `max_steps`: -1
258
+ - `lr_scheduler_type`: linear
259
+ - `lr_scheduler_kwargs`: {}
260
+ - `warmup_ratio`: 0.0
261
+ - `warmup_steps`: 0
262
+ - `log_level`: passive
263
+ - `log_level_replica`: warning
264
+ - `log_on_each_node`: True
265
+ - `logging_nan_inf_filter`: True
266
+ - `save_safetensors`: True
267
+ - `save_on_each_node`: False
268
+ - `save_only_model`: False
269
+ - `restore_callback_states_from_checkpoint`: False
270
+ - `no_cuda`: False
271
+ - `use_cpu`: False
272
+ - `use_mps_device`: False
273
+ - `seed`: 42
274
+ - `data_seed`: None
275
+ - `jit_mode_eval`: False
276
+ - `use_ipex`: False
277
+ - `bf16`: False
278
+ - `fp16`: False
279
+ - `fp16_opt_level`: O1
280
+ - `half_precision_backend`: auto
281
+ - `bf16_full_eval`: False
282
+ - `fp16_full_eval`: False
283
+ - `tf32`: None
284
+ - `local_rank`: 0
285
+ - `ddp_backend`: None
286
+ - `tpu_num_cores`: None
287
+ - `tpu_metrics_debug`: False
288
+ - `debug`: []
289
+ - `dataloader_drop_last`: False
290
+ - `dataloader_num_workers`: 0
291
+ - `dataloader_prefetch_factor`: None
292
+ - `past_index`: -1
293
+ - `disable_tqdm`: False
294
+ - `remove_unused_columns`: True
295
+ - `label_names`: None
296
+ - `load_best_model_at_end`: False
297
+ - `ignore_data_skip`: False
298
+ - `fsdp`: []
299
+ - `fsdp_min_num_params`: 0
300
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
301
+ - `tp_size`: 0
302
+ - `fsdp_transformer_layer_cls_to_wrap`: None
303
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
304
+ - `deepspeed`: None
305
+ - `label_smoothing_factor`: 0.0
306
+ - `optim`: adamw_torch
307
+ - `optim_args`: None
308
+ - `adafactor`: False
309
+ - `group_by_length`: False
310
+ - `length_column_name`: length
311
+ - `ddp_find_unused_parameters`: None
312
+ - `ddp_bucket_cap_mb`: None
313
+ - `ddp_broadcast_buffers`: False
314
+ - `dataloader_pin_memory`: True
315
+ - `dataloader_persistent_workers`: False
316
+ - `skip_memory_metrics`: True
317
+ - `use_legacy_prediction_loop`: False
318
+ - `push_to_hub`: False
319
+ - `resume_from_checkpoint`: None
320
+ - `hub_model_id`: None
321
+ - `hub_strategy`: every_save
322
+ - `hub_private_repo`: None
323
+ - `hub_always_push`: False
324
+ - `gradient_checkpointing`: False
325
+ - `gradient_checkpointing_kwargs`: None
326
+ - `include_inputs_for_metrics`: False
327
+ - `include_for_metrics`: []
328
+ - `eval_do_concat_batches`: True
329
+ - `fp16_backend`: auto
330
+ - `push_to_hub_model_id`: None
331
+ - `push_to_hub_organization`: None
332
+ - `mp_parameters`:
333
+ - `auto_find_batch_size`: False
334
+ - `full_determinism`: False
335
+ - `torchdynamo`: None
336
+ - `ray_scope`: last
337
+ - `ddp_timeout`: 1800
338
+ - `torch_compile`: False
339
+ - `torch_compile_backend`: None
340
+ - `torch_compile_mode`: None
341
+ - `include_tokens_per_second`: False
342
+ - `include_num_input_tokens_seen`: False
343
+ - `neftune_noise_alpha`: None
344
+ - `optim_target_modules`: None
345
+ - `batch_eval_metrics`: False
346
+ - `eval_on_start`: False
347
+ - `use_liger_kernel`: False
348
+ - `eval_use_gather_object`: False
349
+ - `average_tokens_across_devices`: False
350
+ - `prompts`: None
351
+ - `batch_sampler`: batch_sampler
352
+ - `multi_dataset_batch_sampler`: round_robin
353
+
354
+ </details>
355
+
356
+ ### Training Logs
357
+ | Epoch | Step | Training Loss | validation_spearman_cosine |
358
+ |:------:|:-----:|:-------------:|:--------------------------:|
359
+ | 0.0961 | 500 | 0.1933 | - |
360
+ | 0.1922 | 1000 | 0.1191 | - |
361
+ | 0.2883 | 1500 | 0.1016 | - |
362
+ | 0.3844 | 2000 | 0.0914 | - |
363
+ | 0.4805 | 2500 | 0.0841 | - |
364
+ | 0.5766 | 3000 | 0.079 | - |
365
+ | 0.6727 | 3500 | 0.0757 | - |
366
+ | 0.7688 | 4000 | 0.0732 | - |
367
+ | 0.8649 | 4500 | 0.0686 | - |
368
+ | 0.9610 | 5000 | 0.0665 | - |
369
+ | 1.0 | 5203 | - | 0.8284 |
370
+ | 1.0571 | 5500 | 0.0627 | - |
371
+ | 1.1532 | 6000 | 0.0548 | - |
372
+ | 1.2493 | 6500 | 0.0557 | - |
373
+ | 1.3454 | 7000 | 0.0562 | - |
374
+ | 1.4415 | 7500 | 0.0557 | - |
375
+ | 1.5376 | 8000 | 0.0527 | - |
376
+ | 1.6337 | 8500 | 0.0504 | - |
377
+ | 1.7298 | 9000 | 0.0539 | - |
378
+ | 1.8259 | 9500 | 0.054 | - |
379
+ | 1.9220 | 10000 | 0.0501 | - |
380
+ | 2.0 | 10406 | - | 0.8380 |
381
+
382
+
383
+ ### Framework Versions
384
+ - Python: 3.11.12
385
+ - Sentence Transformers: 4.1.0
386
+ - Transformers: 4.51.3
387
+ - PyTorch: 2.7.0
388
+ - Accelerate: 1.6.0
389
+ - Datasets: 2.19.2
390
+ - Tokenizers: 0.21.1
391
+
392
+ ## Citation
393
+
394
+ ### BibTeX
395
+
396
+ #### Sentence Transformers
397
+ ```bibtex
398
+ @inproceedings{reimers-2019-sentence-bert,
399
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
400
+ author = "Reimers, Nils and Gurevych, Iryna",
401
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
402
+ month = "11",
403
+ year = "2019",
404
+ publisher = "Association for Computational Linguistics",
405
+ url = "https://arxiv.org/abs/1908.10084",
406
+ }
407
+ ```
408
+
409
+ <!--
410
+ ## Glossary
411
+
412
+ *Clearly define terms in order to be accessible across audiences.*
413
+ -->
414
+
415
+ <!--
416
+ ## Model Card Authors
417
+
418
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
419
+ -->
420
+
421
+ <!--
422
+ ## Model Card Contact
423
+
424
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
425
+ -->
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "XLMRobertaModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "classifier_dropout": null,
8
+ "dtype": "float32",
9
+ "eos_token_id": 2,
10
+ "gradient_checkpointing": false,
11
+ "hidden_act": "gelu",
12
+ "hidden_dropout_prob": 0.1,
13
+ "hidden_size": 768,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 3072,
16
+ "layer_norm_eps": 1e-05,
17
+ "max_position_embeddings": 514,
18
+ "model_type": "xlm-roberta",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 12,
21
+ "pad_token_id": 1,
22
+ "position_embedding_type": "absolute",
23
+ "torch_dtype": "float32",
24
+ "transformers_version": "4.51.3",
25
+ "type_vocab_size": 1,
26
+ "use_cache": true,
27
+ "vocab_size": 50005
28
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "4.1.0",
4
+ "transformers": "4.51.3",
5
+ "pytorch": "2.7.0"
6
+ },
7
+ "prompts": {},
8
+ "default_prompt_name": null,
9
+ "similarity_fn_name": "cosine"
10
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:520c3ec1179dc00e10e5639e640ac1e4ffb1e97ad3f0a4f0d9c395f1c022a91e
3
+ size 497806184
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 514,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "extra_special_tokens": {},
49
+ "mask_token": "[MASK]",
50
+ "max_length": 512,
51
+ "model_max_length": 1000000000000000019884624838656,
52
+ "never_split": null,
53
+ "pad_token": "[PAD]",
54
+ "sep_token": "[SEP]",
55
+ "stride": 256,
56
+ "strip_accents": null,
57
+ "tokenize_chinese_chars": true,
58
+ "tokenizer_class": "BertTokenizer",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
+ "unk_token": "[UNK]"
62
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff