radoslavralev commited on
Commit
ecbdb74
·
verified ·
1 Parent(s): e99a637

Add new SentenceTransformer model

Browse files
Files changed (2) hide show
  1. README.md +98 -105
  2. config_sentence_transformers.json +1 -1
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:111470
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: thenlper/gte-small
11
  widget:
12
  - source_sentence: when was the first elephant brought to america
13
  sentences:
@@ -132,7 +132,7 @@ metrics:
132
  - cosine_mrr@10
133
  - cosine_map@100
134
  model-index:
135
- - name: SentenceTransformer based on thenlper/gte-small
136
  results:
137
  - task:
138
  type: information-retrieval
@@ -142,49 +142,49 @@ model-index:
142
  type: NanoMSMARCO
143
  metrics:
144
  - type: cosine_accuracy@1
145
- value: 0.16
146
  name: Cosine Accuracy@1
147
  - type: cosine_accuracy@3
148
- value: 0.32
149
  name: Cosine Accuracy@3
150
  - type: cosine_accuracy@5
151
- value: 0.4
152
  name: Cosine Accuracy@5
153
  - type: cosine_accuracy@10
154
- value: 0.54
155
  name: Cosine Accuracy@10
156
  - type: cosine_precision@1
157
- value: 0.16
158
  name: Cosine Precision@1
159
  - type: cosine_precision@3
160
- value: 0.10666666666666666
161
  name: Cosine Precision@3
162
  - type: cosine_precision@5
163
- value: 0.08
164
  name: Cosine Precision@5
165
  - type: cosine_precision@10
166
- value: 0.05400000000000001
167
  name: Cosine Precision@10
168
  - type: cosine_recall@1
169
- value: 0.16
170
  name: Cosine Recall@1
171
  - type: cosine_recall@3
172
- value: 0.32
173
  name: Cosine Recall@3
174
  - type: cosine_recall@5
175
- value: 0.4
176
  name: Cosine Recall@5
177
  - type: cosine_recall@10
178
- value: 0.54
179
  name: Cosine Recall@10
180
  - type: cosine_ndcg@10
181
- value: 0.32698862634234876
182
  name: Cosine Ndcg@10
183
  - type: cosine_mrr@10
184
- value: 0.2620793650793651
185
  name: Cosine Mrr@10
186
  - type: cosine_map@100
187
- value: 0.2747949118190278
188
  name: Cosine Map@100
189
  - task:
190
  type: information-retrieval
@@ -194,49 +194,49 @@ model-index:
194
  type: NanoNQ
195
  metrics:
196
  - type: cosine_accuracy@1
197
- value: 0.22
198
  name: Cosine Accuracy@1
199
  - type: cosine_accuracy@3
200
- value: 0.44
201
  name: Cosine Accuracy@3
202
  - type: cosine_accuracy@5
203
- value: 0.52
204
  name: Cosine Accuracy@5
205
  - type: cosine_accuracy@10
206
- value: 0.64
207
  name: Cosine Accuracy@10
208
  - type: cosine_precision@1
209
- value: 0.22
210
  name: Cosine Precision@1
211
  - type: cosine_precision@3
212
- value: 0.14666666666666664
213
  name: Cosine Precision@3
214
  - type: cosine_precision@5
215
- value: 0.10400000000000001
216
  name: Cosine Precision@5
217
  - type: cosine_precision@10
218
- value: 0.06400000000000002
219
  name: Cosine Precision@10
220
  - type: cosine_recall@1
221
- value: 0.22
222
  name: Cosine Recall@1
223
  - type: cosine_recall@3
224
- value: 0.42
225
  name: Cosine Recall@3
226
  - type: cosine_recall@5
227
- value: 0.49
228
  name: Cosine Recall@5
229
  - type: cosine_recall@10
230
- value: 0.6
231
  name: Cosine Recall@10
232
  - type: cosine_ndcg@10
233
- value: 0.39877036805974797
234
  name: Cosine Ndcg@10
235
  - type: cosine_mrr@10
236
- value: 0.3438015873015873
237
  name: Cosine Mrr@10
238
  - type: cosine_map@100
239
- value: 0.3445409270682024
240
  name: Cosine Map@100
241
  - task:
242
  type: nano-beir
@@ -246,61 +246,61 @@ model-index:
246
  type: NanoBEIR_mean
247
  metrics:
248
  - type: cosine_accuracy@1
249
- value: 0.19
250
  name: Cosine Accuracy@1
251
  - type: cosine_accuracy@3
252
- value: 0.38
253
  name: Cosine Accuracy@3
254
  - type: cosine_accuracy@5
255
- value: 0.46
256
  name: Cosine Accuracy@5
257
  - type: cosine_accuracy@10
258
- value: 0.5900000000000001
259
  name: Cosine Accuracy@10
260
  - type: cosine_precision@1
261
- value: 0.19
262
  name: Cosine Precision@1
263
  - type: cosine_precision@3
264
- value: 0.12666666666666665
265
  name: Cosine Precision@3
266
  - type: cosine_precision@5
267
- value: 0.092
268
  name: Cosine Precision@5
269
  - type: cosine_precision@10
270
- value: 0.05900000000000001
271
  name: Cosine Precision@10
272
  - type: cosine_recall@1
273
- value: 0.19
274
  name: Cosine Recall@1
275
  - type: cosine_recall@3
276
- value: 0.37
277
  name: Cosine Recall@3
278
  - type: cosine_recall@5
279
- value: 0.445
280
  name: Cosine Recall@5
281
  - type: cosine_recall@10
282
- value: 0.5700000000000001
283
  name: Cosine Recall@10
284
  - type: cosine_ndcg@10
285
- value: 0.3628794972010484
286
  name: Cosine Ndcg@10
287
  - type: cosine_mrr@10
288
- value: 0.30294047619047615
289
  name: Cosine Mrr@10
290
  - type: cosine_map@100
291
- value: 0.3096679194436151
292
  name: Cosine Map@100
293
  ---
294
 
295
- # SentenceTransformer based on thenlper/gte-small
296
 
297
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [thenlper/gte-small](https://huggingface.co/thenlper/gte-small). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
298
 
299
  ## Model Details
300
 
301
  ### Model Description
302
  - **Model Type:** Sentence Transformer
303
- - **Base model:** [thenlper/gte-small](https://huggingface.co/thenlper/gte-small) <!-- at revision 17e1f347d17fe144873b1201da91788898c639cd -->
304
  - **Maximum Sequence Length:** 128 tokens
305
  - **Output Dimensionality:** 384 dimensions
306
  - **Similarity Function:** Cosine Similarity
@@ -353,9 +353,9 @@ print(embeddings.shape)
353
  # Get the similarity scores for the embeddings
354
  similarities = model.similarity(embeddings, embeddings)
355
  print(similarities)
356
- # tensor([[1.0001, 1.0001, 0.4386],
357
- # [1.0001, 1.0001, 0.4386],
358
- # [0.4386, 0.4386, 1.0000]])
359
  ```
360
 
361
  <!--
@@ -393,21 +393,21 @@ You can finetune this model on your own dataset.
393
 
394
  | Metric | NanoMSMARCO | NanoNQ |
395
  |:--------------------|:------------|:-----------|
396
- | cosine_accuracy@1 | 0.16 | 0.22 |
397
- | cosine_accuracy@3 | 0.32 | 0.44 |
398
- | cosine_accuracy@5 | 0.4 | 0.52 |
399
- | cosine_accuracy@10 | 0.54 | 0.64 |
400
- | cosine_precision@1 | 0.16 | 0.22 |
401
- | cosine_precision@3 | 0.1067 | 0.1467 |
402
- | cosine_precision@5 | 0.08 | 0.104 |
403
- | cosine_precision@10 | 0.054 | 0.064 |
404
- | cosine_recall@1 | 0.16 | 0.22 |
405
- | cosine_recall@3 | 0.32 | 0.42 |
406
- | cosine_recall@5 | 0.4 | 0.49 |
407
- | cosine_recall@10 | 0.54 | 0.6 |
408
- | **cosine_ndcg@10** | **0.327** | **0.3988** |
409
- | cosine_mrr@10 | 0.2621 | 0.3438 |
410
- | cosine_map@100 | 0.2748 | 0.3445 |
411
 
412
  #### Nano BEIR
413
 
@@ -425,21 +425,21 @@ You can finetune this model on your own dataset.
425
 
426
  | Metric | Value |
427
  |:--------------------|:-----------|
428
- | cosine_accuracy@1 | 0.19 |
429
- | cosine_accuracy@3 | 0.38 |
430
- | cosine_accuracy@5 | 0.46 |
431
- | cosine_accuracy@10 | 0.59 |
432
- | cosine_precision@1 | 0.19 |
433
- | cosine_precision@3 | 0.1267 |
434
- | cosine_precision@5 | 0.092 |
435
- | cosine_precision@10 | 0.059 |
436
- | cosine_recall@1 | 0.19 |
437
- | cosine_recall@3 | 0.37 |
438
- | cosine_recall@5 | 0.445 |
439
- | cosine_recall@10 | 0.57 |
440
- | **cosine_ndcg@10** | **0.3629** |
441
- | cosine_mrr@10 | 0.3029 |
442
- | cosine_map@100 | 0.3097 |
443
 
444
  <!--
445
  ## Bias, Risks and Limitations
@@ -475,7 +475,7 @@ You can finetune this model on your own dataset.
475
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
476
  ```json
477
  {
478
- "scale": 3.0,
479
  "similarity_fct": "cos_sim",
480
  "gather_across_devices": false
481
  }
@@ -501,7 +501,7 @@ You can finetune this model on your own dataset.
501
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
502
  ```json
503
  {
504
- "scale": 3.0,
505
  "similarity_fct": "cos_sim",
506
  "gather_across_devices": false
507
  }
@@ -513,9 +513,9 @@ You can finetune this model on your own dataset.
513
  - `eval_strategy`: steps
514
  - `per_device_train_batch_size`: 128
515
  - `per_device_eval_batch_size`: 128
516
- - `learning_rate`: 8e-05
517
- - `weight_decay`: 0.005
518
- - `max_steps`: 3375
519
  - `warmup_ratio`: 0.1
520
  - `fp16`: True
521
  - `dataloader_drop_last`: True
@@ -542,14 +542,14 @@ You can finetune this model on your own dataset.
542
  - `gradient_accumulation_steps`: 1
543
  - `eval_accumulation_steps`: None
544
  - `torch_empty_cache_steps`: None
545
- - `learning_rate`: 8e-05
546
- - `weight_decay`: 0.005
547
  - `adam_beta1`: 0.9
548
  - `adam_beta2`: 0.999
549
  - `adam_epsilon`: 1e-08
550
  - `max_grad_norm`: 1.0
551
  - `num_train_epochs`: 3.0
552
- - `max_steps`: 3375
553
  - `lr_scheduler_type`: linear
554
  - `lr_scheduler_kwargs`: {}
555
  - `warmup_ratio`: 0.1
@@ -656,20 +656,13 @@ You can finetune this model on your own dataset.
656
  ### Training Logs
657
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
658
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
659
- | 0 | 0 | - | 4.9445 | 0.6259 | 0.6583 | 0.6421 |
660
- | 0.2874 | 250 | 3.6887 | 3.0013 | 0.4676 | 0.4424 | 0.4550 |
661
- | 0.5747 | 500 | 3.0661 | 2.9415 | 0.4647 | 0.4688 | 0.4667 |
662
- | 0.8621 | 750 | 3.0125 | 2.9161 | 0.3994 | 0.4479 | 0.4237 |
663
- | 1.1494 | 1000 | 2.9691 | 2.9044 | 0.3845 | 0.4090 | 0.3968 |
664
- | 1.4368 | 1250 | 2.9407 | 2.8981 | 0.3614 | 0.3858 | 0.3736 |
665
- | 1.7241 | 1500 | 2.9321 | 2.8893 | 0.3182 | 0.3811 | 0.3496 |
666
- | 2.0115 | 1750 | 2.9227 | 2.8817 | 0.3444 | 0.3973 | 0.3708 |
667
- | 2.2989 | 2000 | 2.8854 | 2.8807 | 0.3088 | 0.3730 | 0.3409 |
668
- | 2.5862 | 2250 | 2.8832 | 2.8744 | 0.3251 | 0.3968 | 0.3610 |
669
- | 2.8736 | 2500 | 2.8857 | 2.8730 | 0.3504 | 0.4101 | 0.3802 |
670
- | 3.1609 | 2750 | 2.8677 | 2.8714 | 0.3233 | 0.4021 | 0.3627 |
671
- | 3.4483 | 3000 | 2.86 | 2.8697 | 0.3239 | 0.4106 | 0.3673 |
672
- | 3.7356 | 3250 | 2.8584 | 2.8686 | 0.3270 | 0.3988 | 0.3629 |
673
 
674
 
675
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:111470
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
  - source_sentence: when was the first elephant brought to america
13
  sentences:
 
132
  - cosine_mrr@10
133
  - cosine_map@100
134
  model-index:
135
+ - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
136
  results:
137
  - task:
138
  type: information-retrieval
 
142
  type: NanoMSMARCO
143
  metrics:
144
  - type: cosine_accuracy@1
145
+ value: 0.32
146
  name: Cosine Accuracy@1
147
  - type: cosine_accuracy@3
148
+ value: 0.5
149
  name: Cosine Accuracy@3
150
  - type: cosine_accuracy@5
151
+ value: 0.6
152
  name: Cosine Accuracy@5
153
  - type: cosine_accuracy@10
154
+ value: 0.76
155
  name: Cosine Accuracy@10
156
  - type: cosine_precision@1
157
+ value: 0.32
158
  name: Cosine Precision@1
159
  - type: cosine_precision@3
160
+ value: 0.16666666666666663
161
  name: Cosine Precision@3
162
  - type: cosine_precision@5
163
+ value: 0.12000000000000002
164
  name: Cosine Precision@5
165
  - type: cosine_precision@10
166
+ value: 0.07600000000000001
167
  name: Cosine Precision@10
168
  - type: cosine_recall@1
169
+ value: 0.32
170
  name: Cosine Recall@1
171
  - type: cosine_recall@3
172
+ value: 0.5
173
  name: Cosine Recall@3
174
  - type: cosine_recall@5
175
+ value: 0.6
176
  name: Cosine Recall@5
177
  - type: cosine_recall@10
178
+ value: 0.76
179
  name: Cosine Recall@10
180
  - type: cosine_ndcg@10
181
+ value: 0.5174146339399069
182
  name: Cosine Ndcg@10
183
  - type: cosine_mrr@10
184
+ value: 0.4427063492063491
185
  name: Cosine Mrr@10
186
  - type: cosine_map@100
187
+ value: 0.452501292753926
188
  name: Cosine Map@100
189
  - task:
190
  type: information-retrieval
 
194
  type: NanoNQ
195
  metrics:
196
  - type: cosine_accuracy@1
197
+ value: 0.54
198
  name: Cosine Accuracy@1
199
  - type: cosine_accuracy@3
200
+ value: 0.66
201
  name: Cosine Accuracy@3
202
  - type: cosine_accuracy@5
203
+ value: 0.68
204
  name: Cosine Accuracy@5
205
  - type: cosine_accuracy@10
206
+ value: 0.74
207
  name: Cosine Accuracy@10
208
  - type: cosine_precision@1
209
+ value: 0.54
210
  name: Cosine Precision@1
211
  - type: cosine_precision@3
212
+ value: 0.22
213
  name: Cosine Precision@3
214
  - type: cosine_precision@5
215
+ value: 0.136
216
  name: Cosine Precision@5
217
  - type: cosine_precision@10
218
+ value: 0.08
219
  name: Cosine Precision@10
220
  - type: cosine_recall@1
221
+ value: 0.51
222
  name: Cosine Recall@1
223
  - type: cosine_recall@3
224
+ value: 0.62
225
  name: Cosine Recall@3
226
  - type: cosine_recall@5
227
+ value: 0.64
228
  name: Cosine Recall@5
229
  - type: cosine_recall@10
230
+ value: 0.72
231
  name: Cosine Recall@10
232
  - type: cosine_ndcg@10
233
+ value: 0.6171839770040762
234
  name: Cosine Ndcg@10
235
  - type: cosine_mrr@10
236
+ value: 0.6030555555555556
237
  name: Cosine Mrr@10
238
  - type: cosine_map@100
239
+ value: 0.5845310002947148
240
  name: Cosine Map@100
241
  - task:
242
  type: nano-beir
 
246
  type: NanoBEIR_mean
247
  metrics:
248
  - type: cosine_accuracy@1
249
+ value: 0.43000000000000005
250
  name: Cosine Accuracy@1
251
  - type: cosine_accuracy@3
252
+ value: 0.5800000000000001
253
  name: Cosine Accuracy@3
254
  - type: cosine_accuracy@5
255
+ value: 0.64
256
  name: Cosine Accuracy@5
257
  - type: cosine_accuracy@10
258
+ value: 0.75
259
  name: Cosine Accuracy@10
260
  - type: cosine_precision@1
261
+ value: 0.43000000000000005
262
  name: Cosine Precision@1
263
  - type: cosine_precision@3
264
+ value: 0.1933333333333333
265
  name: Cosine Precision@3
266
  - type: cosine_precision@5
267
+ value: 0.128
268
  name: Cosine Precision@5
269
  - type: cosine_precision@10
270
+ value: 0.07800000000000001
271
  name: Cosine Precision@10
272
  - type: cosine_recall@1
273
+ value: 0.41500000000000004
274
  name: Cosine Recall@1
275
  - type: cosine_recall@3
276
+ value: 0.56
277
  name: Cosine Recall@3
278
  - type: cosine_recall@5
279
+ value: 0.62
280
  name: Cosine Recall@5
281
  - type: cosine_recall@10
282
+ value: 0.74
283
  name: Cosine Recall@10
284
  - type: cosine_ndcg@10
285
+ value: 0.5672993054719916
286
  name: Cosine Ndcg@10
287
  - type: cosine_mrr@10
288
+ value: 0.5228809523809523
289
  name: Cosine Mrr@10
290
  - type: cosine_map@100
291
+ value: 0.5185161465243204
292
  name: Cosine Map@100
293
  ---
294
 
295
+ # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
296
 
297
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
298
 
299
  ## Model Details
300
 
301
  ### Model Description
302
  - **Model Type:** Sentence Transformer
303
+ - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
304
  - **Maximum Sequence Length:** 128 tokens
305
  - **Output Dimensionality:** 384 dimensions
306
  - **Similarity Function:** Cosine Similarity
 
353
  # Get the similarity scores for the embeddings
354
  similarities = model.similarity(embeddings, embeddings)
355
  print(similarities)
356
+ # tensor([[1.0000, 1.0000, 0.8824],
357
+ # [1.0000, 1.0000, 0.8824],
358
+ # [0.8824, 0.8824, 1.0000]])
359
  ```
360
 
361
  <!--
 
393
 
394
  | Metric | NanoMSMARCO | NanoNQ |
395
  |:--------------------|:------------|:-----------|
396
+ | cosine_accuracy@1 | 0.32 | 0.54 |
397
+ | cosine_accuracy@3 | 0.5 | 0.66 |
398
+ | cosine_accuracy@5 | 0.6 | 0.68 |
399
+ | cosine_accuracy@10 | 0.76 | 0.74 |
400
+ | cosine_precision@1 | 0.32 | 0.54 |
401
+ | cosine_precision@3 | 0.1667 | 0.22 |
402
+ | cosine_precision@5 | 0.12 | 0.136 |
403
+ | cosine_precision@10 | 0.076 | 0.08 |
404
+ | cosine_recall@1 | 0.32 | 0.51 |
405
+ | cosine_recall@3 | 0.5 | 0.62 |
406
+ | cosine_recall@5 | 0.6 | 0.64 |
407
+ | cosine_recall@10 | 0.76 | 0.72 |
408
+ | **cosine_ndcg@10** | **0.5174** | **0.6172** |
409
+ | cosine_mrr@10 | 0.4427 | 0.6031 |
410
+ | cosine_map@100 | 0.4525 | 0.5845 |
411
 
412
  #### Nano BEIR
413
 
 
425
 
426
  | Metric | Value |
427
  |:--------------------|:-----------|
428
+ | cosine_accuracy@1 | 0.43 |
429
+ | cosine_accuracy@3 | 0.58 |
430
+ | cosine_accuracy@5 | 0.64 |
431
+ | cosine_accuracy@10 | 0.75 |
432
+ | cosine_precision@1 | 0.43 |
433
+ | cosine_precision@3 | 0.1933 |
434
+ | cosine_precision@5 | 0.128 |
435
+ | cosine_precision@10 | 0.078 |
436
+ | cosine_recall@1 | 0.415 |
437
+ | cosine_recall@3 | 0.56 |
438
+ | cosine_recall@5 | 0.62 |
439
+ | cosine_recall@10 | 0.74 |
440
+ | **cosine_ndcg@10** | **0.5673** |
441
+ | cosine_mrr@10 | 0.5229 |
442
+ | cosine_map@100 | 0.5185 |
443
 
444
  <!--
445
  ## Bias, Risks and Limitations
 
475
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
476
  ```json
477
  {
478
+ "scale": 20.0,
479
  "similarity_fct": "cos_sim",
480
  "gather_across_devices": false
481
  }
 
501
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
502
  ```json
503
  {
504
+ "scale": 20.0,
505
  "similarity_fct": "cos_sim",
506
  "gather_across_devices": false
507
  }
 
513
  - `eval_strategy`: steps
514
  - `per_device_train_batch_size`: 128
515
  - `per_device_eval_batch_size`: 128
516
+ - `learning_rate`: 0.0001
517
+ - `weight_decay`: 0.001
518
+ - `max_steps`: 1687
519
  - `warmup_ratio`: 0.1
520
  - `fp16`: True
521
  - `dataloader_drop_last`: True
 
542
  - `gradient_accumulation_steps`: 1
543
  - `eval_accumulation_steps`: None
544
  - `torch_empty_cache_steps`: None
545
+ - `learning_rate`: 0.0001
546
+ - `weight_decay`: 0.001
547
  - `adam_beta1`: 0.9
548
  - `adam_beta2`: 0.999
549
  - `adam_epsilon`: 1e-08
550
  - `max_grad_norm`: 1.0
551
  - `num_train_epochs`: 3.0
552
+ - `max_steps`: 1687
553
  - `lr_scheduler_type`: linear
554
  - `lr_scheduler_kwargs`: {}
555
  - `warmup_ratio`: 0.1
 
656
  ### Training Logs
657
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
658
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
659
+ | 0 | 0 | - | 0.1310 | 0.5540 | 0.5931 | 0.5735 |
660
+ | 0.2874 | 250 | 0.1078 | 0.0793 | 0.5375 | 0.5386 | 0.5380 |
661
+ | 0.5747 | 500 | 0.0893 | 0.0673 | 0.5031 | 0.6009 | 0.5520 |
662
+ | 0.8621 | 750 | 0.081 | 0.0605 | 0.5414 | 0.5786 | 0.5600 |
663
+ | 1.1494 | 1000 | 0.0593 | 0.0565 | 0.5158 | 0.5786 | 0.5472 |
664
+ | 1.4368 | 1250 | 0.0422 | 0.0537 | 0.5300 | 0.6107 | 0.5704 |
665
+ | 1.7241 | 1500 | 0.0402 | 0.0514 | 0.5174 | 0.6172 | 0.5673 |
 
 
 
 
 
 
 
666
 
667
 
668
  ### Framework Versions
config_sentence_transformers.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "model_type": "SentenceTransformer",
3
  "__version__": {
4
  "sentence_transformers": "5.2.0",
5
  "transformers": "4.57.3",
6
  "pytorch": "2.9.1+cu128"
7
  },
 
8
  "prompts": {
9
  "query": "",
10
  "document": ""
 
1
  {
 
2
  "__version__": {
3
  "sentence_transformers": "5.2.0",
4
  "transformers": "4.57.3",
5
  "pytorch": "2.9.1+cu128"
6
  },
7
+ "model_type": "SentenceTransformer",
8
  "prompts": {
9
  "query": "",
10
  "document": ""