tomaarsen HF Staff commited on
Commit
bceae88
·
verified ·
1 Parent(s): 6722f33

Add new SentenceTransformer model

Browse files
1_Pooling/config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false,
7
+ "pooling_mode_weightedmean_tokens": false,
8
+ "pooling_mode_lasttoken": false,
9
+ "include_prompt": true
10
+ }
README.md ADDED
@@ -0,0 +1,601 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ tags:
6
+ - sentence-transformers
7
+ - sentence-similarity
8
+ - feature-extraction
9
+ - dense
10
+ - generated_from_trainer
11
+ - dataset_size:90000
12
+ - loss:CachedMultipleNegativesRankingLoss
13
+ base_model: microsoft/mpnet-base
14
+ widget:
15
+ - source_sentence: what is the difference between trojan virus and worm?
16
+ sentences:
17
+ - Worms spread from computer to computer, but unlike a virus, it has the capability
18
+ to travel without any help from a person. ... A Trojan horse is not a virus. It
19
+ is a destructive program that looks as a genuine application. Unlike viruses,
20
+ Trojan horses do not replicate themselves but they can be just as destructive.
21
+ - You're usually no longer infectious 24 hours after starting a course of antibiotics,
22
+ but this time period can sometimes vary. For example, the antibiotics may take
23
+ longer to work if your body takes longer to absorb them, or if you're taking other
24
+ medicine that interacts with the antibiotics.
25
+ - Eating salt raises the amount of sodium in your bloodstream and wrecks the delicate
26
+ balance, reducing the ability of your kidneys to remove the water. The result
27
+ is a higher blood pressure due to the extra fluid and extra strain on the delicate
28
+ blood vessels leading to the kidneys.
29
+ - source_sentence: which are the neighbouring countries of pakistan?
30
+ sentences:
31
+ - Pakistan is bordered by India on the east, the Arabian Sea on the south, Iran
32
+ on the southwest, and Afghanistan on the west and north; in the northeast is the
33
+ disputed territory (with India) of Kashmir, of which the part occupied by Pakistan
34
+ borders on China.
35
+ - Age is a big factor when it comes to how much sleep a dog needs. Just as human
36
+ babies need a lot of sleep, the AKC notes your puppy needs 15-20 hours of sleep
37
+ a day to help his central nervous system, immune system and muscles develop properly.
38
+ - 'Step 1: Connect your iPhone to your computer using n USB cable through any of
39
+ the USB ports available on your computer. Step 2: Open iTunes, click the "Files"
40
+ tab and check the boxes to sync or transfer your files. Step 3: Select your desired
41
+ destination folder for the files and click "Sync" to complete the transfer.'
42
+ - source_sentence: what can you do with 1gb of data?
43
+ sentences:
44
+ - You could even contact your email provider, complain that somebody else is using
45
+ your email address, and say that you are worried about your account being compromised.
46
+ They're very unlikely to do anything, but if something goes wrong, at least you
47
+ can prove you forewarned them.
48
+ - 1) Under Section 80CCD(1), investment in Atal Pension Yojana or NPS up to ₹ 1.5
49
+ lakh qualifies for income tax deduction. But remember that the total amount of
50
+ deduction under sections 80C, 80CCC and 80CCD cannot exceed ₹ 1.5 lakh.
51
+ - 1GB (or 1024MB) of data lets you send or receive about 1,000 emails and browse
52
+ the Internet for about 20 hours every month. (This limit relates only to your
53
+ 1GB mobile data allocation; if you are an 'inclusive mobile broadband customer'
54
+ you also get 2000 BT Wi-fi wi-fi minutes every month.)
55
+ - source_sentence: how many carbon atoms are in carbon dioxide?
56
+ sentences:
57
+ - For CO2 there is one atom of carbon and two atoms of oxygen. For H2O, there is
58
+ one atom of oxygen and two atoms of hydrogen. A molecule can be made of only one
59
+ type of atom.
60
+ - Avian influenza refers to the disease caused by infection with avian (bird) influenza
61
+ (flu) Type A viruses. These viruses occur naturally among wild aquatic birds worldwide
62
+ and can infect domestic poultry and other bird and animal species. Avian flu viruses
63
+ do not normally infect humans.
64
+ - At the end of "Inception," Dom Cobb (Leonardo DiCaprio) finally returns home to
65
+ his kids after spending a long time in the dream world. Cobb carries a little
66
+ top with him. If the top keeps spinning, that means he is in a dream. ... The
67
+ final shot shows the top spinning, but it never reveals whether it falls over.
68
+ - source_sentence: is duchenne muscular dystrophy a dominant or recessive trait?
69
+ sentences:
70
+ - Duchenne muscular dystrophy is inherited in an X-linked recessive pattern. Males
71
+ have only one copy of the X chromosome from their mother and one copy of the Y
72
+ chromosome from their father. If their X chromosome has a DMD gene mutation, they
73
+ will have Duchenne muscular dystrophy.
74
+ - An automatic transmission will downshift for you when you drive uphill. However,
75
+ for moderately steep slopes, it's wise to shift to the gear range marked D2, 2,
76
+ or L to ascend and descend the hill. For steep slopes that you can't ascend at
77
+ a speed faster than 10 mph (about 15 kph), shift to D1 or 1.
78
+ - The dream suggests captivity and it refers to your fear of punishment. Another
79
+ interpretation of this dream refers to a need to do what you feel is right in
80
+ waking life. Being in jail suggests that your feelings may be trapped by a limited
81
+ mind and body. ... Jail also suggests repressed feelings.
82
+ datasets:
83
+ - sentence-transformers/gooaq
84
+ pipeline_tag: sentence-similarity
85
+ library_name: sentence-transformers
86
+ metrics:
87
+ - cosine_accuracy@1
88
+ - cosine_accuracy@3
89
+ - cosine_accuracy@5
90
+ - cosine_accuracy@10
91
+ - cosine_precision@1
92
+ - cosine_precision@3
93
+ - cosine_precision@5
94
+ - cosine_precision@10
95
+ - cosine_recall@1
96
+ - cosine_recall@3
97
+ - cosine_recall@5
98
+ - cosine_recall@10
99
+ - cosine_ndcg@10
100
+ - cosine_mrr@10
101
+ - cosine_map@100
102
+ co2_eq_emissions:
103
+ emissions: 25.654063970832134
104
+ energy_consumed: 0.09585932386288172
105
+ source: codecarbon
106
+ training_type: fine-tuning
107
+ on_cloud: false
108
+ cpu_model: 13th Gen Intel(R) Core(TM) i7-13700K
109
+ ram_total_size: 31.777088165283203
110
+ hours_used: 0.27
111
+ hardware_used: 1 x NVIDIA GeForce RTX 3090
112
+ model-index:
113
+ - name: MPNet base trained on GooAQ triplets using CachedMultipleNegativesRankingLoss
114
+ with GradCache
115
+ results:
116
+ - task:
117
+ type: information-retrieval
118
+ name: Information Retrieval
119
+ dataset:
120
+ name: gooaq dev
121
+ type: gooaq-dev
122
+ metrics:
123
+ - type: cosine_accuracy@1
124
+ value: 0.7568
125
+ name: Cosine Accuracy@1
126
+ - type: cosine_accuracy@3
127
+ value: 0.8954
128
+ name: Cosine Accuracy@3
129
+ - type: cosine_accuracy@5
130
+ value: 0.9344
131
+ name: Cosine Accuracy@5
132
+ - type: cosine_accuracy@10
133
+ value: 0.9645
134
+ name: Cosine Accuracy@10
135
+ - type: cosine_precision@1
136
+ value: 0.7568
137
+ name: Cosine Precision@1
138
+ - type: cosine_precision@3
139
+ value: 0.2984666666666666
140
+ name: Cosine Precision@3
141
+ - type: cosine_precision@5
142
+ value: 0.18688000000000002
143
+ name: Cosine Precision@5
144
+ - type: cosine_precision@10
145
+ value: 0.09645
146
+ name: Cosine Precision@10
147
+ - type: cosine_recall@1
148
+ value: 0.7568
149
+ name: Cosine Recall@1
150
+ - type: cosine_recall@3
151
+ value: 0.8954
152
+ name: Cosine Recall@3
153
+ - type: cosine_recall@5
154
+ value: 0.9344
155
+ name: Cosine Recall@5
156
+ - type: cosine_recall@10
157
+ value: 0.9645
158
+ name: Cosine Recall@10
159
+ - type: cosine_ndcg@10
160
+ value: 0.8651990399153616
161
+ name: Cosine Ndcg@10
162
+ - type: cosine_mrr@10
163
+ value: 0.8328272619047579
164
+ name: Cosine Mrr@10
165
+ - type: cosine_map@100
166
+ value: 0.8345033162131785
167
+ name: Cosine Map@100
168
+ ---
169
+
170
+ # MPNet base trained on GooAQ triplets using CachedMultipleNegativesRankingLoss with GradCache
171
+
172
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [microsoft/mpnet-base](https://huggingface.co/microsoft/mpnet-base) on the [gooaq](https://huggingface.co/datasets/sentence-transformers/gooaq) dataset. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
173
+
174
+ ## Model Details
175
+
176
+ ### Model Description
177
+ - **Model Type:** Sentence Transformer
178
+ - **Base model:** [microsoft/mpnet-base](https://huggingface.co/microsoft/mpnet-base) <!-- at revision 6996ce1e91bd2a9c7d7f61daec37463394f73f09 -->
179
+ - **Maximum Sequence Length:** 512 tokens
180
+ - **Output Dimensionality:** 768 dimensions
181
+ - **Similarity Function:** Cosine Similarity
182
+ - **Training Dataset:**
183
+ - [gooaq](https://huggingface.co/datasets/sentence-transformers/gooaq)
184
+ - **Language:** en
185
+ - **License:** apache-2.0
186
+
187
+ ### Model Sources
188
+
189
+ - **Documentation:** [Sentence Transformers Documentation](https://sbert.net)
190
+ - **Repository:** [Sentence Transformers on GitHub](https://github.com/huggingface/sentence-transformers)
191
+ - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers)
192
+
193
+ ### Full Model Architecture
194
+
195
+ ```
196
+ SentenceTransformer(
197
+ (0): Transformer({'max_seq_length': 512, 'do_lower_case': False, 'architecture': 'MPNetModel'})
198
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
199
+ )
200
+ ```
201
+
202
+ ## Usage
203
+
204
+ ### Direct Usage (Sentence Transformers)
205
+
206
+ First install the Sentence Transformers library:
207
+
208
+ ```bash
209
+ pip install -U sentence-transformers
210
+ ```
211
+
212
+ Then you can load this model and run inference.
213
+ ```python
214
+ from sentence_transformers import SentenceTransformer
215
+
216
+ # Download from the 🤗 Hub
217
+ model = SentenceTransformer("tomaarsen/mpnet-base-gooaq-cmnrl-1024bs-GradCache")
218
+ # Run inference
219
+ queries = [
220
+ "is duchenne muscular dystrophy a dominant or recessive trait?",
221
+ ]
222
+ documents = [
223
+ 'Duchenne muscular dystrophy is inherited in an X-linked recessive pattern. Males have only one copy of the X chromosome from their mother and one copy of the Y chromosome from their father. If their X chromosome has a DMD gene mutation, they will have Duchenne muscular dystrophy.',
224
+ 'The dream suggests captivity and it refers to your fear of punishment. Another interpretation of this dream refers to a need to do what you feel is right in waking life. Being in jail suggests that your feelings may be trapped by a limited mind and body. ... Jail also suggests repressed feelings.',
225
+ "An automatic transmission will downshift for you when you drive uphill. However, for moderately steep slopes, it's wise to shift to the gear range marked D2, 2, or L to ascend and descend the hill. For steep slopes that you can't ascend at a speed faster than 10 mph (about 15 kph), shift to D1 or 1.",
226
+ ]
227
+ query_embeddings = model.encode_query(queries)
228
+ document_embeddings = model.encode_document(documents)
229
+ print(query_embeddings.shape, document_embeddings.shape)
230
+ # [1, 768] [3, 768]
231
+
232
+ # Get the similarity scores for the embeddings
233
+ similarities = model.similarity(query_embeddings, document_embeddings)
234
+ print(similarities)
235
+ # tensor([[0.8208, 0.1993, 0.1582]])
236
+ ```
237
+
238
+ <!--
239
+ ### Direct Usage (Transformers)
240
+
241
+ <details><summary>Click to see the direct usage in Transformers</summary>
242
+
243
+ </details>
244
+ -->
245
+
246
+ <!--
247
+ ### Downstream Usage (Sentence Transformers)
248
+
249
+ You can finetune this model on your own dataset.
250
+
251
+ <details><summary>Click to expand</summary>
252
+
253
+ </details>
254
+ -->
255
+
256
+ <!--
257
+ ### Out-of-Scope Use
258
+
259
+ *List how the model may foreseeably be misused and address what users ought not to do with the model.*
260
+ -->
261
+
262
+ ## Evaluation
263
+
264
+ ### Metrics
265
+
266
+ #### Information Retrieval
267
+
268
+ * Dataset: `gooaq-dev`
269
+ * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
270
+
271
+ | Metric | Value |
272
+ |:--------------------|:-----------|
273
+ | cosine_accuracy@1 | 0.7568 |
274
+ | cosine_accuracy@3 | 0.8954 |
275
+ | cosine_accuracy@5 | 0.9344 |
276
+ | cosine_accuracy@10 | 0.9645 |
277
+ | cosine_precision@1 | 0.7568 |
278
+ | cosine_precision@3 | 0.2985 |
279
+ | cosine_precision@5 | 0.1869 |
280
+ | cosine_precision@10 | 0.0964 |
281
+ | cosine_recall@1 | 0.7568 |
282
+ | cosine_recall@3 | 0.8954 |
283
+ | cosine_recall@5 | 0.9344 |
284
+ | cosine_recall@10 | 0.9645 |
285
+ | **cosine_ndcg@10** | **0.8652** |
286
+ | cosine_mrr@10 | 0.8328 |
287
+ | cosine_map@100 | 0.8345 |
288
+
289
+ <!--
290
+ ## Bias, Risks and Limitations
291
+
292
+ *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
293
+ -->
294
+
295
+ <!--
296
+ ### Recommendations
297
+
298
+ *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
299
+ -->
300
+
301
+ ## Training Details
302
+
303
+ ### Training Dataset
304
+
305
+ #### gooaq
306
+
307
+ * Dataset: [gooaq](https://huggingface.co/datasets/sentence-transformers/gooaq) at [b089f72](https://huggingface.co/datasets/sentence-transformers/gooaq/tree/b089f728748a068b7bc5234e5bcf5b25e3c8279c)
308
+ * Size: 90,000 training samples
309
+ * Columns: <code>question</code> and <code>answer</code>
310
+ * Approximate statistics based on the first 1000 samples:
311
+ | | question | answer |
312
+ |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
313
+ | type | string | string |
314
+ | details | <ul><li>min: 8 tokens</li><li>mean: 11.83 tokens</li><li>max: 20 tokens</li></ul> | <ul><li>min: 15 tokens</li><li>mean: 60.45 tokens</li><li>max: 180 tokens</li></ul> |
315
+ * Samples:
316
+ | question | answer |
317
+ |:--------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
318
+ | <code>how long does halifax take to transfer mortgage funds?</code> | <code>Bear in mind that the speed of application will vary depending on your own personal circumstances and the lender's present day-to-day performance. In some cases, applications can be approved by the lender within 24 hours, while some can take weeks or even months.</code> |
319
+ | <code>can you get a false pregnancy test?</code> | <code>In very rare cases, you can have a false-positive result. This means you're not pregnant but the test says you are. You could have a false-positive result if you have blood or protein in your pee. Certain drugs, such as tranquilizers, anticonvulsants, hypnotics, and fertility drugs, could cause false-positive results.</code> |
320
+ | <code>are ahead of its time?</code> | <code>Definition of ahead of one's/its time : too advanced or modern to be understood or appreciated during the time when one lives or works As a director, he was ahead of his time.</code> |
321
+ * Loss: [<code>CachedMultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss) with these parameters:
322
+ ```json
323
+ {
324
+ "scale": 20.0,
325
+ "similarity_fct": "cos_sim",
326
+ "mini_batch_size": 64,
327
+ "gather_across_devices": false,
328
+ "use_cont_accum": false,
329
+ "cache_size": 0,
330
+ "prev_cache": false
331
+ }
332
+ ```
333
+
334
+ ### Evaluation Dataset
335
+
336
+ #### gooaq
337
+
338
+ * Dataset: [gooaq](https://huggingface.co/datasets/sentence-transformers/gooaq) at [b089f72](https://huggingface.co/datasets/sentence-transformers/gooaq/tree/b089f728748a068b7bc5234e5bcf5b25e3c8279c)
339
+ * Size: 10,000 evaluation samples
340
+ * Columns: <code>question</code> and <code>answer</code>
341
+ * Approximate statistics based on the first 1000 samples:
342
+ | | question | answer |
343
+ |:--------|:----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
344
+ | type | string | string |
345
+ | details | <ul><li>min: 8 tokens</li><li>mean: 11.93 tokens</li><li>max: 25 tokens</li></ul> | <ul><li>min: 14 tokens</li><li>mean: 60.84 tokens</li><li>max: 127 tokens</li></ul> |
346
+ * Samples:
347
+ | question | answer |
348
+ |:-----------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
349
+ | <code>should you take ibuprofen with high blood pressure?</code> | <code>In general, people with high blood pressure should use acetaminophen or possibly aspirin for over-the-counter pain relief. Unless your health care provider has said it's OK, you should not use ibuprofen, ketoprofen, or naproxen sodium. If aspirin or acetaminophen doesn't help with your pain, call your doctor.</code> |
350
+ | <code>how old do you have to be to work in sc?</code> | <code>The general minimum age of employment for South Carolina youth is 14, although the state allows younger children who are performers to work in show business. If their families are agricultural workers, children younger than age 14 may also participate in farm labor.</code> |
351
+ | <code>how to write a topic proposal for a research paper?</code> | <code>['Write down the main topic of your paper. ... ', 'Write two or three short sentences under the main topic that explain why you chose that topic. ... ', 'Write a thesis sentence that states the angle and purpose of your research paper. ... ', 'List the items you will cover in the body of the paper that support your thesis statement.']</code> |
352
+ * Loss: [<code>CachedMultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedmultiplenegativesrankingloss) with these parameters:
353
+ ```json
354
+ {
355
+ "scale": 20.0,
356
+ "similarity_fct": "cos_sim",
357
+ "mini_batch_size": 64,
358
+ "gather_across_devices": false,
359
+ "use_cont_accum": false,
360
+ "cache_size": 0,
361
+ "prev_cache": false
362
+ }
363
+ ```
364
+
365
+ ### Training Hyperparameters
366
+ #### Non-Default Hyperparameters
367
+
368
+ - `eval_strategy`: steps
369
+ - `per_device_train_batch_size`: 1024
370
+ - `per_device_eval_batch_size`: 1024
371
+ - `learning_rate`: 8e-05
372
+ - `num_train_epochs`: 1
373
+ - `warmup_ratio`: 0.1
374
+ - `bf16`: True
375
+ - `batch_sampler`: no_duplicates
376
+
377
+ #### All Hyperparameters
378
+ <details><summary>Click to expand</summary>
379
+
380
+ - `overwrite_output_dir`: False
381
+ - `do_predict`: False
382
+ - `eval_strategy`: steps
383
+ - `prediction_loss_only`: True
384
+ - `per_device_train_batch_size`: 1024
385
+ - `per_device_eval_batch_size`: 1024
386
+ - `per_gpu_train_batch_size`: None
387
+ - `per_gpu_eval_batch_size`: None
388
+ - `gradient_accumulation_steps`: 1
389
+ - `eval_accumulation_steps`: None
390
+ - `torch_empty_cache_steps`: None
391
+ - `learning_rate`: 8e-05
392
+ - `weight_decay`: 0.0
393
+ - `adam_beta1`: 0.9
394
+ - `adam_beta2`: 0.999
395
+ - `adam_epsilon`: 1e-08
396
+ - `max_grad_norm`: 1.0
397
+ - `num_train_epochs`: 1
398
+ - `max_steps`: -1
399
+ - `lr_scheduler_type`: linear
400
+ - `lr_scheduler_kwargs`: None
401
+ - `warmup_ratio`: 0.1
402
+ - `warmup_steps`: 0
403
+ - `log_level`: passive
404
+ - `log_level_replica`: warning
405
+ - `log_on_each_node`: True
406
+ - `logging_nan_inf_filter`: True
407
+ - `save_safetensors`: True
408
+ - `save_on_each_node`: False
409
+ - `save_only_model`: False
410
+ - `restore_callback_states_from_checkpoint`: False
411
+ - `no_cuda`: False
412
+ - `use_cpu`: False
413
+ - `use_mps_device`: False
414
+ - `seed`: 42
415
+ - `data_seed`: None
416
+ - `jit_mode_eval`: False
417
+ - `bf16`: True
418
+ - `fp16`: False
419
+ - `fp16_opt_level`: O1
420
+ - `half_precision_backend`: auto
421
+ - `bf16_full_eval`: False
422
+ - `fp16_full_eval`: False
423
+ - `tf32`: None
424
+ - `local_rank`: 0
425
+ - `ddp_backend`: None
426
+ - `tpu_num_cores`: None
427
+ - `tpu_metrics_debug`: False
428
+ - `debug`: []
429
+ - `dataloader_drop_last`: False
430
+ - `dataloader_num_workers`: 0
431
+ - `dataloader_prefetch_factor`: None
432
+ - `past_index`: -1
433
+ - `disable_tqdm`: False
434
+ - `remove_unused_columns`: True
435
+ - `label_names`: None
436
+ - `load_best_model_at_end`: False
437
+ - `ignore_data_skip`: False
438
+ - `fsdp`: []
439
+ - `fsdp_min_num_params`: 0
440
+ - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}
441
+ - `fsdp_transformer_layer_cls_to_wrap`: None
442
+ - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None}
443
+ - `parallelism_config`: None
444
+ - `deepspeed`: None
445
+ - `label_smoothing_factor`: 0.0
446
+ - `optim`: adamw_torch_fused
447
+ - `optim_args`: None
448
+ - `adafactor`: False
449
+ - `group_by_length`: False
450
+ - `length_column_name`: length
451
+ - `project`: huggingface
452
+ - `trackio_space_id`: trackio
453
+ - `ddp_find_unused_parameters`: None
454
+ - `ddp_bucket_cap_mb`: None
455
+ - `ddp_broadcast_buffers`: False
456
+ - `dataloader_pin_memory`: True
457
+ - `dataloader_persistent_workers`: False
458
+ - `skip_memory_metrics`: True
459
+ - `use_legacy_prediction_loop`: False
460
+ - `push_to_hub`: False
461
+ - `resume_from_checkpoint`: None
462
+ - `hub_model_id`: None
463
+ - `hub_strategy`: every_save
464
+ - `hub_private_repo`: None
465
+ - `hub_always_push`: False
466
+ - `hub_revision`: None
467
+ - `gradient_checkpointing`: False
468
+ - `gradient_checkpointing_kwargs`: None
469
+ - `include_inputs_for_metrics`: False
470
+ - `include_for_metrics`: []
471
+ - `eval_do_concat_batches`: True
472
+ - `fp16_backend`: auto
473
+ - `push_to_hub_model_id`: None
474
+ - `push_to_hub_organization`: None
475
+ - `mp_parameters`:
476
+ - `auto_find_batch_size`: False
477
+ - `full_determinism`: False
478
+ - `torchdynamo`: None
479
+ - `ray_scope`: last
480
+ - `ddp_timeout`: 1800
481
+ - `torch_compile`: False
482
+ - `torch_compile_backend`: None
483
+ - `torch_compile_mode`: None
484
+ - `include_tokens_per_second`: False
485
+ - `include_num_input_tokens_seen`: no
486
+ - `neftune_noise_alpha`: None
487
+ - `optim_target_modules`: None
488
+ - `batch_eval_metrics`: False
489
+ - `eval_on_start`: False
490
+ - `use_liger_kernel`: False
491
+ - `liger_kernel_config`: None
492
+ - `eval_use_gather_object`: False
493
+ - `average_tokens_across_devices`: True
494
+ - `prompts`: None
495
+ - `batch_sampler`: no_duplicates
496
+ - `multi_dataset_batch_sampler`: proportional
497
+ - `router_mapping`: {}
498
+ - `learning_rate_mapping`: {}
499
+
500
+ </details>
501
+
502
+ ### Training Logs
503
+ | Epoch | Step | Training Loss | Validation Loss | gooaq-dev_cosine_ndcg@10 |
504
+ |:------:|:----:|:-------------:|:---------------:|:------------------------:|
505
+ | -1 | -1 | - | - | 0.2154 |
506
+ | 0.0114 | 1 | 6.4064 | - | - |
507
+ | 0.0568 | 5 | 5.8799 | - | - |
508
+ | 0.1023 | 9 | - | 1.1303 | 0.6342 |
509
+ | 0.1136 | 10 | 2.4927 | - | - |
510
+ | 0.1705 | 15 | 1.0794 | - | - |
511
+ | 0.2045 | 18 | - | 0.5062 | 0.7918 |
512
+ | 0.2273 | 20 | 0.7681 | - | - |
513
+ | 0.2841 | 25 | 0.622 | - | - |
514
+ | 0.3068 | 27 | - | 0.3818 | 0.8250 |
515
+ | 0.3409 | 30 | 0.5329 | - | - |
516
+ | 0.3977 | 35 | 0.4671 | - | - |
517
+ | 0.4091 | 36 | - | 0.3244 | 0.8404 |
518
+ | 0.4545 | 40 | 0.4472 | - | - |
519
+ | 0.5114 | 45 | 0.4308 | 0.2978 | 0.8496 |
520
+ | 0.5682 | 50 | 0.4292 | - | - |
521
+ | 0.6136 | 54 | - | 0.2803 | 0.8554 |
522
+ | 0.625 | 55 | 0.3973 | - | - |
523
+ | 0.6818 | 60 | 0.3789 | - | - |
524
+ | 0.7159 | 63 | - | 0.2673 | 0.8601 |
525
+ | 0.7386 | 65 | 0.3477 | - | - |
526
+ | 0.7955 | 70 | 0.3725 | - | - |
527
+ | 0.8182 | 72 | - | 0.2548 | 0.8632 |
528
+ | 0.8523 | 75 | 0.3541 | - | - |
529
+ | 0.9091 | 80 | 0.352 | - | - |
530
+ | 0.9205 | 81 | - | 0.2498 | 0.8644 |
531
+ | 0.9659 | 85 | 0.339 | - | - |
532
+ | -1 | -1 | - | - | 0.8652 |
533
+
534
+
535
+ ### Environmental Impact
536
+ Carbon emissions were measured using [CodeCarbon](https://github.com/mlco2/codecarbon).
537
+ - **Energy Consumed**: 0.096 kWh
538
+ - **Carbon Emitted**: 0.026 kg of CO2
539
+ - **Hours Used**: 0.27 hours
540
+
541
+ ### Training Hardware
542
+ - **On Cloud**: No
543
+ - **GPU Model**: 1 x NVIDIA GeForce RTX 3090
544
+ - **CPU Model**: 13th Gen Intel(R) Core(TM) i7-13700K
545
+ - **RAM Size**: 31.78 GB
546
+
547
+ ### Framework Versions
548
+ - Python: 3.11.6
549
+ - Sentence Transformers: 5.3.0.dev0
550
+ - Transformers: 4.57.6
551
+ - PyTorch: 2.10.0+cu128
552
+ - Accelerate: 1.6.0
553
+ - Datasets: 4.5.0
554
+ - Tokenizers: 0.22.2
555
+
556
+ ## Citation
557
+
558
+ ### BibTeX
559
+
560
+ #### Sentence Transformers
561
+ ```bibtex
562
+ @inproceedings{reimers-2019-sentence-bert,
563
+ title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks",
564
+ author = "Reimers, Nils and Gurevych, Iryna",
565
+ booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing",
566
+ month = "11",
567
+ year = "2019",
568
+ publisher = "Association for Computational Linguistics",
569
+ url = "https://arxiv.org/abs/1908.10084",
570
+ }
571
+ ```
572
+
573
+ #### CachedMultipleNegativesRankingLoss
574
+ ```bibtex
575
+ @misc{gao2021scaling,
576
+ title={Scaling Deep Contrastive Learning Batch Size under Memory Limited Setup},
577
+ author={Luyu Gao and Yunyi Zhang and Jiawei Han and Jamie Callan},
578
+ year={2021},
579
+ eprint={2101.06983},
580
+ archivePrefix={arXiv},
581
+ primaryClass={cs.LG}
582
+ }
583
+ ```
584
+
585
+ <!--
586
+ ## Glossary
587
+
588
+ *Clearly define terms in order to be accessible across audiences.*
589
+ -->
590
+
591
+ <!--
592
+ ## Model Card Authors
593
+
594
+ *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
595
+ -->
596
+
597
+ <!--
598
+ ## Model Card Contact
599
+
600
+ *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
601
+ -->
config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MPNetModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
+ "dtype": "float32",
8
+ "eos_token_id": 2,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "mpnet",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
20
+ "relative_attention_num_buckets": 32,
21
+ "transformers_version": "4.57.6",
22
+ "vocab_size": 30527
23
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "SentenceTransformer",
3
+ "__version__": {
4
+ "sentence_transformers": "5.3.0.dev0",
5
+ "transformers": "4.57.6",
6
+ "pytorch": "2.10.0+cu128"
7
+ },
8
+ "prompts": {
9
+ "query": "",
10
+ "document": ""
11
+ },
12
+ "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
14
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b07336e3fb4977ab89a0a0a50b6617e815f5206c5a473efdc4d0e0ef7aae7330
3
+ size 437967672
modules.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ }
14
+ ]
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "104": {
36
+ "content": "[UNK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "30526": {
44
+ "content": "<mask>",
45
+ "lstrip": true,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ }
51
+ },
52
+ "bos_token": "<s>",
53
+ "clean_up_tokenization_spaces": false,
54
+ "cls_token": "<s>",
55
+ "do_lower_case": true,
56
+ "eos_token": "</s>",
57
+ "extra_special_tokens": {},
58
+ "mask_token": "<mask>",
59
+ "model_max_length": 512,
60
+ "pad_token": "<pad>",
61
+ "sep_token": "</s>",
62
+ "strip_accents": null,
63
+ "tokenize_chinese_chars": true,
64
+ "tokenizer_class": "MPNetTokenizer",
65
+ "unk_token": "[UNK]"
66
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff