koffiwind commited on
Commit
619ca08
·
verified ·
1 Parent(s): c1b20cb

Upload 11 files

Browse files
jobs_model/1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 1024,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
jobs_model/README.md ADDED
@@ -0,0 +1,464 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - sentence-transformers
4
+ - sentence-similarity
5
+ - feature-extraction
6
+ - generated_from_trainer
7
+ - dataset_size:6
8
+ - loss:MatryoshkaLoss
9
+ - loss:MultipleNegativesRankingLoss
10
+ base_model: Snowflake/snowflake-arctic-embed-l
11
+ pipeline_tag: sentence-similarity
12
+ library_name: sentence-transformers
13
+ metrics:
14
+ - cosine_accuracy@1
15
+ - cosine_accuracy@3
16
+ - cosine_accuracy@5
17
+ - cosine_accuracy@10
18
+ - cosine_precision@1
19
+ - cosine_precision@3
20
+ - cosine_precision@5
21
+ - cosine_precision@10
22
+ - cosine_recall@1
23
+ - cosine_recall@3
24
+ - cosine_recall@5
25
+ - cosine_recall@10
26
+ - cosine_ndcg@10
27
+ - cosine_mrr@10
28
+ - cosine_map@100
29
+ model-index:
30
+ - name: SentenceTransformer based on Snowflake/snowflake-arctic-embed-l
31
+ results:
32
+ - task:
33
+ type: information-retrieval
34
+ name: Information Retrieval
35
+ dataset:
36
+ name: Unknown
37
+ type: unknown
38
+ metrics:
39
+ - type: cosine_accuracy@1
40
+ value: 1.0
41
+ name: Cosine Accuracy@1
42
+ - type: cosine_accuracy@3
43
+ value: 1.0
44
+ name: Cosine Accuracy@3
45
+ - type: cosine_accuracy@5
46
+ value: 1.0
47
+ name: Cosine Accuracy@5
48
+ - type: cosine_accuracy@10
49
+ value: 1.0
50
+ name: Cosine Accuracy@10
51
+ - type: cosine_precision@1
52
+ value: 1.0
53
+ name: Cosine Precision@1
54
+ - type: cosine_precision@3
55
+ value: 0.3333333333333333
56
+ name: Cosine Precision@3
57
+ - type: cosine_precision@5
58
+ value: 0.2
59
+ name: Cosine Precision@5
60
+ - type: cosine_precision@10
61
+ value: 0.1
62
+ name: Cosine Precision@10
63
+ - type: cosine_recall@1
64
+ value: 1.0
65
+ name: Cosine Recall@1
66
+ - type: cosine_recall@3
67
+ value: 1.0
68
+ name: Cosine Recall@3
69
+ - type: cosine_recall@5
70
+ value: 1.0
71
+ name: Cosine Recall@5
72
+ - type: cosine_recall@10
73
+ value: 1.0
74
+ name: Cosine Recall@10
75
+ - type: cosine_ndcg@10
76
+ value: 1.0
77
+ name: Cosine Ndcg@10
78
+ - type: cosine_mrr@10
79
+ value: 1.0
80
+ name: Cosine Mrr@10
81
+ - type: cosine_map@100
82
+ value: 1.0
83
+ name: Cosine Map@100
84
+ ---
85
+
86
+ # SentenceTransformer based on Snowflake/snowflake-arctic-embed-l
87
+
88
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Snowflake/snowflake-arctic-embed-l](https://huggingface.co/Snowflake/snowflake-arctic-embed-l). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
89
+
90
+ ## Model Details
91
+
92
+ ### Model Description
93
+ - **Model Type:** Sentence Transformer
94
+ - **Base model:** [Snowflake/snowflake-arctic-embed-l](https://huggingface.co/Snowflake/snowflake-arctic-embed-l) <!-- at revision d8fb21ca8d905d2832ee8b96c894d3298964346b -->
95
+ - **Maximum Sequence Length:** 512 tokens
96
+ - **Output Dimensionality:** 1024 dimensions
97
+ - **Similarity Function:** Cosine Similarity
98
+ <!-- - **Training Dataset:** Unknown -->
99
+ <!-- - **Language:** Unknown -->
100
+ <!-- - **License:** Unknown -->
101
+
102
+ ### Model Sources
103
+
104
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
105
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers)
106
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
107
+
108
+ ### Full Model Architecture
109
+
110
+ ```
111
+ SentenceTransformer(
112
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel
113
+ (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
114
+ (2): Normalize()
115
+ )
116
+ ```
117
+
118
+ ## Usage
119
+
120
+ ### Direct Usage (Sentence Transformers)
121
+
122
+ First install the Sentence Transformers library:
123
+
124
+ ```bash
125
+ pip install -U sentence-transformers
126
+ ```
127
+
128
+ Then you can load this model and run inference.
129
+ ```python
130
+ from sentence_transformers import SentenceTransformer
131
+
132
+ # Download from the 🤗 Hub
133
+ model = SentenceTransformer("sentence_transformers_model_id")
134
+ # Run inference
135
+ sentences = [
136
+ 'The weather is lovely today.',
137
+ "It's so sunny outside!",
138
+ 'He drove to the stadium.',
139
+ ]
140
+ embeddings = model.encode(sentences)
141
+ print(embeddings.shape)
142
+ # [3, 1024]
143
+
144
+ # Get the similarity scores for the embeddings
145
+ similarities = model.similarity(embeddings, embeddings)
146
+ print(similarities.shape)
147
+ # [3, 3]
148
+ ```
149
+
150
+ <!--
151
+ ### Direct Usage (Transformers)
152
+
153
+ <details><summary>Click to see the direct usage in Transformers</summary>
154
+
155
+ </details>
156
+ -->
157
+
158
+ <!--
159
+ ### Downstream Usage (Sentence Transformers)
160
+
161
+ You can finetune this model on your own dataset.
162
+
163
+ <details><summary>Click to expand</summary>
164
+
165
+ </details>
166
+ -->
167
+
168
+ <!--
169
+ ### Out-of-Scope Use
170
+
171
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
172
+ -->
173
+
174
+ ## Evaluation
175
+
176
+ ### Metrics
177
+
178
+ #### Information Retrieval
179
+
180
+ * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
181
+
182
+ | Metric | Value |
183
+ |:--------------------|:--------|
184
+ | cosine_accuracy@1 | 1.0 |
185
+ | cosine_accuracy@3 | 1.0 |
186
+ | cosine_accuracy@5 | 1.0 |
187
+ | cosine_accuracy@10 | 1.0 |
188
+ | cosine_precision@1 | 1.0 |
189
+ | cosine_precision@3 | 0.3333 |
190
+ | cosine_precision@5 | 0.2 |
191
+ | cosine_precision@10 | 0.1 |
192
+ | cosine_recall@1 | 1.0 |
193
+ | cosine_recall@3 | 1.0 |
194
+ | cosine_recall@5 | 1.0 |
195
+ | cosine_recall@10 | 1.0 |
196
+ | **cosine_ndcg@10** | **1.0** |
197
+ | cosine_mrr@10 | 1.0 |
198
+ | cosine_map@100 | 1.0 |
199
+
200
+ <!--
201
+ ## Bias, Risks and Limitations
202
+
203
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
204
+ -->
205
+
206
+ <!--
207
+ ### Recommendations
208
+
209
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
210
+ -->
211
+
212
+ ## Training Details
213
+
214
+ ### Training Dataset
215
+
216
+ #### Unnamed Dataset
217
+
218
+ * Size: 6 training samples
219
+ * Columns: <code>sentence_0</code> and <code>sentence_1</code>
220
+ * Approximate statistics based on the first 6 samples:
221
+ | | sentence_0 | sentence_1 |
222
+ |:--------|:-----------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
223
+ | type | string | string |
224
+ | details | <ul><li>min: 23 tokens</li><li>mean: 24.33 tokens</li><li>max: 26 tokens</li></ul> | <ul><li>min: 212 tokens</li><li>mean: 224.0 tokens</li><li>max: 244 tokens</li></ul> |
225
+ * Samples:
226
+ | sentence_0 | sentence_1 |
227
+ |:-------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
228
+ | <code>Quelles technologies un développeur Full Stack doit-il maîtriser selon la description fournie ?</code> | <code>{"job_name": "D\u00e9veloppeur Full Stack", "description": "Le d\u00e9veloppeur Full Stack est responsable de la conception et du d\u00e9veloppement complet d\u2019une application web, du front-end au back-end. Il doit ma\u00eetriser plusieurs technologies comme JavaScript, Node.js, React, Vue.js, Python, SQL et NoSQL. Il joue un r\u00f4le cl\u00e9 dans les \u00e9quipes tech car il peut intervenir sur tous les aspects du d\u00e9veloppement, de l\u2019interface utilisateur \u00e0 la gestion des bases de donn\u00e9es.", "spec": "Freelance / Salari\u00e9, Remote-friendly, Travail seul ou en \u00e9quipe"}</code> |
229
+ | <code>Quel est le rôle principal d'un développeur Full Stack dans une équipe technique ?</code> | <code>{"job_name": "D\u00e9veloppeur Full Stack", "description": "Le d\u00e9veloppeur Full Stack est responsable de la conception et du d\u00e9veloppement complet d\u2019une application web, du front-end au back-end. Il doit ma\u00eetriser plusieurs technologies comme JavaScript, Node.js, React, Vue.js, Python, SQL et NoSQL. Il joue un r\u00f4le cl\u00e9 dans les \u00e9quipes tech car il peut intervenir sur tous les aspects du d\u00e9veloppement, de l\u2019interface utilisateur \u00e0 la gestion des bases de donn\u00e9es.", "spec": "Freelance / Salari\u00e9, Remote-friendly, Travail seul ou en \u00e9quipe"}</code> |
230
+ | <code>Quels langages de programmation utilise un développeur Backend dans son travail ?</code> | <code>{"job_name": "D\u00e9veloppeur Backend", "description": "Le d\u00e9veloppeur Backend s\u2019occupe de la logique m\u00e9tier d\u2019une application, la gestion des serveurs et des bases de donn\u00e9es. Il travaille avec des langages comme Python, Java, PHP, Ruby, SQL et GraphQL et est souvent en charge des APIs, de la s\u00e9curit\u00e9 et des performances des applications. Il collabore avec les \u00e9quipes Frontend et DevOps pour assurer un fonctionnement optimal du produit.", "spec": "Freelance / Salari\u00e9, Remote-friendly, Plut\u00f4t solitaire"}</code> |
231
+ * Loss: [<code>MatryoshkaLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#matryoshkaloss) with these parameters:
232
+ ```json
233
+ {
234
+ "loss": "MultipleNegativesRankingLoss",
235
+ "matryoshka_dims": [
236
+ 768,
237
+ 512,
238
+ 256,
239
+ 128,
240
+ 64
241
+ ],
242
+ "matryoshka_weights": [
243
+ 1,
244
+ 1,
245
+ 1,
246
+ 1,
247
+ 1
248
+ ],
249
+ "n_dims_per_step": -1
250
+ }
251
+ ```
252
+
253
+ ### Training Hyperparameters
254
+ #### Non-Default Hyperparameters
255
+
256
+ - `eval_strategy`: steps
257
+ - `per_device_train_batch_size`: 10
258
+ - `per_device_eval_batch_size`: 10
259
+ - `num_train_epochs`: 10
260
+ - `multi_dataset_batch_sampler`: round_robin
261
+
262
+ #### All Hyperparameters
263
+ <details><summary>Click to expand</summary>
264
+
265
+ - `overwrite_output_dir`: False
266
+ - `do_predict`: False
267
+ - `eval_strategy`: steps
268
+ - `prediction_loss_only`: True
269
+ - `per_device_train_batch_size`: 10
270
+ - `per_device_eval_batch_size`: 10
271
+ - `per_gpu_train_batch_size`: None
272
+ - `per_gpu_eval_batch_size`: None
273
+ - `gradient_accumulation_steps`: 1
274
+ - `eval_accumulation_steps`: None
275
+ - `torch_empty_cache_steps`: None
276
+ - `learning_rate`: 5e-05
277
+ - `weight_decay`: 0.0
278
+ - `adam_beta1`: 0.9
279
+ - `adam_beta2`: 0.999
280
+ - `adam_epsilon`: 1e-08
281
+ - `max_grad_norm`: 1
282
+ - `num_train_epochs`: 10
283
+ - `max_steps`: -1
284
+ - `lr_scheduler_type`: linear
285
+ - `lr_scheduler_kwargs`: {}
286
+ - `warmup_ratio`: 0.0
287
+ - `warmup_steps`: 0
288
+ - `log_level`: passive
289
+ - `log_level_replica`: warning
290
+ - `log_on_each_node`: True
291
+ - `logging_nan_inf_filter`: True
292
+ - `save_safetensors`: True
293
+ - `save_on_each_node`: False
294
+ - `save_only_model`: False
295
+ - `restore_callback_states_from_checkpoint`: False
296
+ - `no_cuda`: False
297
+ - `use_cpu`: False
298
+ - `use_mps_device`: False
299
+ - `seed`: 42
300
+ - `data_seed`: None
301
+ - `jit_mode_eval`: False
302
+ - `use_ipex`: False
303
+ - `bf16`: False
304
+ - `fp16`: False
305
+ - `fp16_opt_level`: O1
306
+ - `half_precision_backend`: auto
307
+ - `bf16_full_eval`: False
308
+ - `fp16_full_eval`: False
309
+ - `tf32`: None
310
+ - `local_rank`: 0
311
+ - `ddp_backend`: None
312
+ - `tpu_num_cores`: None
313
+ - `tpu_metrics_debug`: False
314
+ - `debug`: []
315
+ - `dataloader_drop_last`: False
316
+ - `dataloader_num_workers`: 0
317
+ - `dataloader_prefetch_factor`: None
318
+ - `past_index`: -1
319
+ - `disable_tqdm`: False
320
+ - `remove_unused_columns`: True
321
+ - `label_names`: None
322
+ - `load_best_model_at_end`: False
323
+ - `ignore_data_skip`: False
324
+ - `fsdp`: []
325
+ - `fsdp_min_num_params`: 0
326
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
327
+ - `fsdp_transformer_layer_cls_to_wrap`: None
328
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
329
+ - `deepspeed`: None
330
+ - `label_smoothing_factor`: 0.0
331
+ - `optim`: adamw_torch
332
+ - `optim_args`: None
333
+ - `adafactor`: False
334
+ - `group_by_length`: False
335
+ - `length_column_name`: length
336
+ - `ddp_find_unused_parameters`: None
337
+ - `ddp_bucket_cap_mb`: None
338
+ - `ddp_broadcast_buffers`: False
339
+ - `dataloader_pin_memory`: True
340
+ - `dataloader_persistent_workers`: False
341
+ - `skip_memory_metrics`: True
342
+ - `use_legacy_prediction_loop`: False
343
+ - `push_to_hub`: False
344
+ - `resume_from_checkpoint`: None
345
+ - `hub_model_id`: None
346
+ - `hub_strategy`: every_save
347
+ - `hub_private_repo`: None
348
+ - `hub_always_push`: False
349
+ - `gradient_checkpointing`: False
350
+ - `gradient_checkpointing_kwargs`: None
351
+ - `include_inputs_for_metrics`: False
352
+ - `include_for_metrics`: []
353
+ - `eval_do_concat_batches`: True
354
+ - `fp16_backend`: auto
355
+ - `push_to_hub_model_id`: None
356
+ - `push_to_hub_organization`: None
357
+ - `mp_parameters`:
358
+ - `auto_find_batch_size`: False
359
+ - `full_determinism`: False
360
+ - `torchdynamo`: None
361
+ - `ray_scope`: last
362
+ - `ddp_timeout`: 1800
363
+ - `torch_compile`: False
364
+ - `torch_compile_backend`: None
365
+ - `torch_compile_mode`: None
366
+ - `dispatch_batches`: None
367
+ - `split_batches`: None
368
+ - `include_tokens_per_second`: False
369
+ - `include_num_input_tokens_seen`: False
370
+ - `neftune_noise_alpha`: None
371
+ - `optim_target_modules`: None
372
+ - `batch_eval_metrics`: False
373
+ - `eval_on_start`: False
374
+ - `use_liger_kernel`: False
375
+ - `eval_use_gather_object`: False
376
+ - `average_tokens_across_devices`: False
377
+ - `prompts`: None
378
+ - `batch_sampler`: batch_sampler
379
+ - `multi_dataset_batch_sampler`: round_robin
380
+
381
+ </details>
382
+
383
+ ### Training Logs
384
+ | Epoch | Step | cosine_ndcg@10 |
385
+ |:-----:|:----:|:--------------:|
386
+ | 1.0 | 1 | 1.0 |
387
+ | 2.0 | 2 | 1.0 |
388
+ | 3.0 | 3 | 1.0 |
389
+ | 4.0 | 4 | 1.0 |
390
+ | 5.0 | 5 | 1.0 |
391
+ | 6.0 | 6 | 1.0 |
392
+ | 7.0 | 7 | 1.0 |
393
+ | 8.0 | 8 | 1.0 |
394
+ | 9.0 | 9 | 1.0 |
395
+ | 10.0 | 10 | 1.0 |
396
+
397
+
398
+ ### Framework Versions
399
+ - Python: 3.13.1
400
+ - Sentence Transformers: 3.4.1
401
+ - Transformers: 4.49.0
402
+ - PyTorch: 2.6.0+cu124
403
+ - Accelerate: 1.4.0
404
+ - Datasets: 3.3.2
405
+ - Tokenizers: 0.21.0
406
+
407
+ ## Citation
408
+
409
+ ### BibTeX
410
+
411
+ #### Sentence Transformers
412
+ ```bibtex
413
+ @inproceedings{reimers-2019-sentence-bert,
414
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
415
+ author = "Reimers, Nils and Gurevych, Iryna",
416
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
417
+ month = "11",
418
+ year = "2019",
419
+ publisher = "Association for Computational Linguistics",
420
+ url = "https://arxiv.org/abs/1908.10084",
421
+ }
422
+ ```
423
+
424
+ #### MatryoshkaLoss
425
+ ```bibtex
426
+ @misc{kusupati2024matryoshka,
427
+ title={Matryoshka Representation Learning},
428
+ author={Aditya Kusupati and Gantavya Bhatt and Aniket Rege and Matthew Wallingford and Aditya Sinha and Vivek Ramanujan and William Howard-Snyder and Kaifeng Chen and Sham Kakade and Prateek Jain and Ali Farhadi},
429
+ year={2024},
430
+ eprint={2205.13147},
431
+ archivePrefix={arXiv},
432
+ primaryClass={cs.LG}
433
+ }
434
+ ```
435
+
436
+ #### MultipleNegativesRankingLoss
437
+ ```bibtex
438
+ @misc{henderson2017efficient,
439
+ title={Efficient Natural Language Response Suggestion for Smart Reply},
440
+ author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil},
441
+ year={2017},
442
+ eprint={1705.00652},
443
+ archivePrefix={arXiv},
444
+ primaryClass={cs.CL}
445
+ }
446
+ ```
447
+
448
+ <!--
449
+ ## Glossary
450
+
451
+ *Clearly define terms in order to be accessible across audiences.*
452
+ -->
453
+
454
+ <!--
455
+ ## Model Card Authors
456
+
457
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
458
+ -->
459
+
460
+ <!--
461
+ ## Model Card Contact
462
+
463
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
464
+ -->
jobs_model/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Snowflake/snowflake-arctic-embed-l",
3
+ "architectures": [
4
+ "BertModel"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 1024,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 4096,
13
+ "layer_norm_eps": 1e-12,
14
+ "max_position_embeddings": 512,
15
+ "model_type": "bert",
16
+ "num_attention_heads": 16,
17
+ "num_hidden_layers": 24,
18
+ "pad_token_id": 0,
19
+ "position_embedding_type": "absolute",
20
+ "torch_dtype": "float32",
21
+ "transformers_version": "4.49.0",
22
+ "type_vocab_size": 2,
23
+ "use_cache": true,
24
+ "vocab_size": 30522
25
+ }
jobs_model/config_sentence_transformers.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "3.4.1",
4
+ "transformers": "4.49.0",
5
+ "pytorch": "2.6.0+cu124"
6
+ },
7
+ "prompts": {
8
+ "query": "Represent this sentence for searching relevant passages: "
9
+ },
10
+ "default_prompt_name": null,
11
+ "similarity_fn_name": "cosine"
12
+ }
jobs_model/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46e36bf399c152c51818e1d0423bfe628d79fc9536dbf9d4fcc93078827d6c99
3
+ size 1336413848
jobs_model/modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
jobs_model/sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
jobs_model/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
jobs_model/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
jobs_model/tokenizer_config.json ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": true,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "max_length": 512,
50
+ "model_max_length": 512,
51
+ "pad_to_multiple_of": null,
52
+ "pad_token": "[PAD]",
53
+ "pad_token_type_id": 0,
54
+ "padding_side": "right",
55
+ "sep_token": "[SEP]",
56
+ "stride": 0,
57
+ "strip_accents": null,
58
+ "tokenize_chinese_chars": true,
59
+ "tokenizer_class": "BertTokenizer",
60
+ "truncation_side": "right",
61
+ "truncation_strategy": "longest_first",
62
+ "unk_token": "[UNK]"
63
+ }
jobs_model/vocab.txt ADDED
The diff for this file is too large to render. See raw diff