radoslavralev commited on
Commit
b58b8f6
·
verified ·
1 Parent(s): ab531ff

Add new SentenceTransformer model

Browse files
1_Pooling/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "word_embedding_dimension": 768,
3
- "pooling_mode_cls_token": true,
4
- "pooling_mode_mean_tokens": false,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
 
1
  {
2
+ "word_embedding_dimension": 384,
3
+ "pooling_mode_cls_token": false,
4
+ "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:90000
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: Alibaba-NLP/gte-modernbert-base
11
  widget:
12
  - source_sentence: what is the maximum i can contribute to a traditional ira
13
  sentences:
@@ -121,7 +121,7 @@ metrics:
121
  - cosine_mrr@10
122
  - cosine_map@100
123
  model-index:
124
- - name: SentenceTransformer based on Alibaba-NLP/gte-modernbert-base
125
  results:
126
  - task:
127
  type: information-retrieval
@@ -131,49 +131,49 @@ model-index:
131
  type: NanoMSMARCO
132
  metrics:
133
  - type: cosine_accuracy@1
134
- value: 0.44
135
  name: Cosine Accuracy@1
136
  - type: cosine_accuracy@3
137
- value: 0.66
138
  name: Cosine Accuracy@3
139
  - type: cosine_accuracy@5
140
- value: 0.72
141
  name: Cosine Accuracy@5
142
  - type: cosine_accuracy@10
143
- value: 0.84
144
  name: Cosine Accuracy@10
145
  - type: cosine_precision@1
146
- value: 0.44
147
  name: Cosine Precision@1
148
  - type: cosine_precision@3
149
- value: 0.22
150
  name: Cosine Precision@3
151
  - type: cosine_precision@5
152
- value: 0.14400000000000002
153
  name: Cosine Precision@5
154
  - type: cosine_precision@10
155
- value: 0.08399999999999999
156
  name: Cosine Precision@10
157
  - type: cosine_recall@1
158
- value: 0.44
159
  name: Cosine Recall@1
160
  - type: cosine_recall@3
161
- value: 0.66
162
  name: Cosine Recall@3
163
  - type: cosine_recall@5
164
- value: 0.72
165
  name: Cosine Recall@5
166
  - type: cosine_recall@10
167
- value: 0.84
168
  name: Cosine Recall@10
169
  - type: cosine_ndcg@10
170
- value: 0.6329590083197414
171
  name: Cosine Ndcg@10
172
  - type: cosine_mrr@10
173
- value: 0.5681349206349207
174
  name: Cosine Mrr@10
175
  - type: cosine_map@100
176
- value: 0.5748098465861624
177
  name: Cosine Map@100
178
  - task:
179
  type: information-retrieval
@@ -183,49 +183,49 @@ model-index:
183
  type: NanoNQ
184
  metrics:
185
  - type: cosine_accuracy@1
186
- value: 0.32
187
  name: Cosine Accuracy@1
188
  - type: cosine_accuracy@3
189
- value: 0.54
190
  name: Cosine Accuracy@3
191
  - type: cosine_accuracy@5
192
  value: 0.64
193
  name: Cosine Accuracy@5
194
  - type: cosine_accuracy@10
195
- value: 0.74
196
  name: Cosine Accuracy@10
197
  - type: cosine_precision@1
198
- value: 0.32
199
  name: Cosine Precision@1
200
  - type: cosine_precision@3
201
- value: 0.18666666666666665
202
  name: Cosine Precision@3
203
  - type: cosine_precision@5
204
- value: 0.132
205
  name: Cosine Precision@5
206
  - type: cosine_precision@10
207
- value: 0.08
208
  name: Cosine Precision@10
209
  - type: cosine_recall@1
210
- value: 0.3
211
  name: Cosine Recall@1
212
  - type: cosine_recall@3
213
- value: 0.52
214
  name: Cosine Recall@3
215
  - type: cosine_recall@5
216
- value: 0.6
217
  name: Cosine Recall@5
218
  - type: cosine_recall@10
219
- value: 0.71
220
  name: Cosine Recall@10
221
  - type: cosine_ndcg@10
222
- value: 0.49769715371181095
223
  name: Cosine Ndcg@10
224
  - type: cosine_mrr@10
225
- value: 0.43952380952380954
226
  name: Cosine Mrr@10
227
  - type: cosine_map@100
228
- value: 0.43617752668168885
229
  name: Cosine Map@100
230
  - task:
231
  type: nano-beir
@@ -235,63 +235,63 @@ model-index:
235
  type: NanoBEIR_mean
236
  metrics:
237
  - type: cosine_accuracy@1
238
- value: 0.38
239
  name: Cosine Accuracy@1
240
  - type: cosine_accuracy@3
241
- value: 0.6000000000000001
242
  name: Cosine Accuracy@3
243
  - type: cosine_accuracy@5
244
- value: 0.6799999999999999
245
  name: Cosine Accuracy@5
246
  - type: cosine_accuracy@10
247
- value: 0.79
248
  name: Cosine Accuracy@10
249
  - type: cosine_precision@1
250
- value: 0.38
251
  name: Cosine Precision@1
252
  - type: cosine_precision@3
253
- value: 0.2033333333333333
254
  name: Cosine Precision@3
255
  - type: cosine_precision@5
256
  value: 0.138
257
  name: Cosine Precision@5
258
  - type: cosine_precision@10
259
- value: 0.08199999999999999
260
  name: Cosine Precision@10
261
  - type: cosine_recall@1
262
- value: 0.37
263
  name: Cosine Recall@1
264
  - type: cosine_recall@3
265
- value: 0.5900000000000001
266
  name: Cosine Recall@3
267
  - type: cosine_recall@5
268
- value: 0.6599999999999999
269
  name: Cosine Recall@5
270
  - type: cosine_recall@10
271
- value: 0.7749999999999999
272
  name: Cosine Recall@10
273
  - type: cosine_ndcg@10
274
- value: 0.5653280810157761
275
  name: Cosine Ndcg@10
276
  - type: cosine_mrr@10
277
- value: 0.503829365079365
278
  name: Cosine Mrr@10
279
  - type: cosine_map@100
280
- value: 0.5054936866339257
281
  name: Cosine Map@100
282
  ---
283
 
284
- # SentenceTransformer based on Alibaba-NLP/gte-modernbert-base
285
 
286
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
287
 
288
  ## Model Details
289
 
290
  ### Model Description
291
  - **Model Type:** Sentence Transformer
292
- - **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision e7f32e3c00f91d699e8c43b53106206bcc72bb22 -->
293
  - **Maximum Sequence Length:** 128 tokens
294
- - **Output Dimensionality:** 768 dimensions
295
  - **Similarity Function:** Cosine Similarity
296
  <!-- - **Training Dataset:** Unknown -->
297
  <!-- - **Language:** Unknown -->
@@ -307,8 +307,9 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [A
307
 
308
  ```
309
  SentenceTransformer(
310
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'ModernBertModel'})
311
- (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
 
312
  )
313
  ```
314
 
@@ -336,14 +337,14 @@ sentences = [
336
  ]
337
  embeddings = model.encode(sentences)
338
  print(embeddings.shape)
339
- # [3, 768]
340
 
341
  # Get the similarity scores for the embeddings
342
  similarities = model.similarity(embeddings, embeddings)
343
  print(similarities)
344
- # tensor([[0.9999, 0.3909, 0.5755],
345
- # [0.3909, 0.9999, 0.1633],
346
- # [0.5755, 0.1633, 0.9999]])
347
  ```
348
 
349
  <!--
@@ -379,23 +380,23 @@ You can finetune this model on your own dataset.
379
  * Datasets: `NanoMSMARCO` and `NanoNQ`
380
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
381
 
382
- | Metric | NanoMSMARCO | NanoNQ |
383
- |:--------------------|:------------|:-----------|
384
- | cosine_accuracy@1 | 0.44 | 0.32 |
385
- | cosine_accuracy@3 | 0.66 | 0.54 |
386
- | cosine_accuracy@5 | 0.72 | 0.64 |
387
- | cosine_accuracy@10 | 0.84 | 0.74 |
388
- | cosine_precision@1 | 0.44 | 0.32 |
389
- | cosine_precision@3 | 0.22 | 0.1867 |
390
- | cosine_precision@5 | 0.144 | 0.132 |
391
- | cosine_precision@10 | 0.084 | 0.08 |
392
- | cosine_recall@1 | 0.44 | 0.3 |
393
- | cosine_recall@3 | 0.66 | 0.52 |
394
- | cosine_recall@5 | 0.72 | 0.6 |
395
- | cosine_recall@10 | 0.84 | 0.71 |
396
- | **cosine_ndcg@10** | **0.633** | **0.4977** |
397
- | cosine_mrr@10 | 0.5681 | 0.4395 |
398
- | cosine_map@100 | 0.5748 | 0.4362 |
399
 
400
  #### Nano BEIR
401
 
@@ -413,21 +414,21 @@ You can finetune this model on your own dataset.
413
 
414
  | Metric | Value |
415
  |:--------------------|:-----------|
416
- | cosine_accuracy@1 | 0.38 |
417
- | cosine_accuracy@3 | 0.6 |
418
- | cosine_accuracy@5 | 0.68 |
419
- | cosine_accuracy@10 | 0.79 |
420
- | cosine_precision@1 | 0.38 |
421
- | cosine_precision@3 | 0.2033 |
422
  | cosine_precision@5 | 0.138 |
423
- | cosine_precision@10 | 0.082 |
424
- | cosine_recall@1 | 0.37 |
425
- | cosine_recall@3 | 0.59 |
426
- | cosine_recall@5 | 0.66 |
427
- | cosine_recall@10 | 0.775 |
428
- | **cosine_ndcg@10** | **0.5653** |
429
- | cosine_mrr@10 | 0.5038 |
430
- | cosine_map@100 | 0.5055 |
431
 
432
  <!--
433
  ## Bias, Risks and Limitations
@@ -450,10 +451,10 @@ You can finetune this model on your own dataset.
450
  * Size: 90,000 training samples
451
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
452
  * Approximate statistics based on the first 1000 samples:
453
- | | anchor | positive | negative |
454
- |:--------|:--------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
455
- | type | string | string | string |
456
- | details | <ul><li>min: 4 tokens</li><li>mean: 9.3 tokens</li><li>max: 27 tokens</li></ul> | <ul><li>min: 20 tokens</li><li>mean: 79.39 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 17 tokens</li><li>mean: 78.77 tokens</li><li>max: 128 tokens</li></ul> |
457
  * Samples:
458
  | anchor | positive | negative |
459
  |:--------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
@@ -476,10 +477,10 @@ You can finetune this model on your own dataset.
476
  * Size: 10,000 evaluation samples
477
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
478
  * Approximate statistics based on the first 1000 samples:
479
- | | anchor | positive | negative |
480
- |:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
481
- | type | string | string | string |
482
- | details | <ul><li>min: 4 tokens</li><li>mean: 9.33 tokens</li><li>max: 21 tokens</li></ul> | <ul><li>min: 17 tokens</li><li>mean: 79.7 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 18 tokens</li><li>mean: 77.72 tokens</li><li>max: 128 tokens</li></ul> |
483
  * Samples:
484
  | anchor | positive | negative |
485
  |:----------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
@@ -644,19 +645,19 @@ You can finetune this model on your own dataset.
644
  ### Training Logs
645
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
646
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
647
- | 0 | 0 | - | 2.7123 | 0.6530 | 0.6552 | 0.6541 |
648
- | 0.3556 | 250 | 1.5463 | 0.9474 | 0.6490 | 0.5678 | 0.6084 |
649
- | 0.7112 | 500 | 0.9209 | 0.8966 | 0.6648 | 0.5657 | 0.6153 |
650
- | 1.0669 | 750 | 0.8879 | 0.8806 | 0.6618 | 0.5579 | 0.6099 |
651
- | 1.4225 | 1000 | 0.8374 | 0.8719 | 0.6586 | 0.5354 | 0.5970 |
652
- | 1.7781 | 1250 | 0.831 | 0.8651 | 0.6657 | 0.5295 | 0.5976 |
653
- | 2.1337 | 1500 | 0.8096 | 0.8626 | 0.6443 | 0.5006 | 0.5725 |
654
- | 2.4893 | 1750 | 0.7797 | 0.8602 | 0.6449 | 0.5081 | 0.5765 |
655
- | 2.8450 | 2000 | 0.7782 | 0.8579 | 0.6376 | 0.5033 | 0.5704 |
656
- | 3.2006 | 2250 | 0.7561 | 0.8611 | 0.6428 | 0.5007 | 0.5718 |
657
- | 3.5562 | 2500 | 0.7494 | 0.8598 | 0.6482 | 0.4921 | 0.5701 |
658
- | 3.9118 | 2750 | 0.7436 | 0.8607 | 0.6427 | 0.4982 | 0.5704 |
659
- | 4.2674 | 3000 | 0.733 | 0.8612 | 0.6330 | 0.4977 | 0.5653 |
660
 
661
 
662
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:90000
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: thenlper/gte-small
11
  widget:
12
  - source_sentence: what is the maximum i can contribute to a traditional ira
13
  sentences:
 
121
  - cosine_mrr@10
122
  - cosine_map@100
123
  model-index:
124
+ - name: SentenceTransformer based on thenlper/gte-small
125
  results:
126
  - task:
127
  type: information-retrieval
 
131
  type: NanoMSMARCO
132
  metrics:
133
  - type: cosine_accuracy@1
134
+ value: 0.38
135
  name: Cosine Accuracy@1
136
  - type: cosine_accuracy@3
137
+ value: 0.56
138
  name: Cosine Accuracy@3
139
  - type: cosine_accuracy@5
140
+ value: 0.68
141
  name: Cosine Accuracy@5
142
  - type: cosine_accuracy@10
143
+ value: 0.72
144
  name: Cosine Accuracy@10
145
  - type: cosine_precision@1
146
+ value: 0.38
147
  name: Cosine Precision@1
148
  - type: cosine_precision@3
149
+ value: 0.18666666666666668
150
  name: Cosine Precision@3
151
  - type: cosine_precision@5
152
+ value: 0.136
153
  name: Cosine Precision@5
154
  - type: cosine_precision@10
155
+ value: 0.07200000000000001
156
  name: Cosine Precision@10
157
  - type: cosine_recall@1
158
+ value: 0.38
159
  name: Cosine Recall@1
160
  - type: cosine_recall@3
161
+ value: 0.56
162
  name: Cosine Recall@3
163
  - type: cosine_recall@5
164
+ value: 0.68
165
  name: Cosine Recall@5
166
  - type: cosine_recall@10
167
+ value: 0.72
168
  name: Cosine Recall@10
169
  - type: cosine_ndcg@10
170
+ value: 0.5560017600077842
171
  name: Cosine Ndcg@10
172
  - type: cosine_mrr@10
173
+ value: 0.5023333333333333
174
  name: Cosine Mrr@10
175
  - type: cosine_map@100
176
+ value: 0.5147876135794199
177
  name: Cosine Map@100
178
  - task:
179
  type: information-retrieval
 
183
  type: NanoNQ
184
  metrics:
185
  - type: cosine_accuracy@1
186
+ value: 0.4
187
  name: Cosine Accuracy@1
188
  - type: cosine_accuracy@3
189
+ value: 0.56
190
  name: Cosine Accuracy@3
191
  - type: cosine_accuracy@5
192
  value: 0.64
193
  name: Cosine Accuracy@5
194
  - type: cosine_accuracy@10
195
+ value: 0.76
196
  name: Cosine Accuracy@10
197
  - type: cosine_precision@1
198
+ value: 0.4
199
  name: Cosine Precision@1
200
  - type: cosine_precision@3
201
+ value: 0.19333333333333333
202
  name: Cosine Precision@3
203
  - type: cosine_precision@5
204
+ value: 0.14
205
  name: Cosine Precision@5
206
  - type: cosine_precision@10
207
+ value: 0.08199999999999999
208
  name: Cosine Precision@10
209
  - type: cosine_recall@1
210
+ value: 0.37
211
  name: Cosine Recall@1
212
  - type: cosine_recall@3
213
+ value: 0.54
214
  name: Cosine Recall@3
215
  - type: cosine_recall@5
216
+ value: 0.63
217
  name: Cosine Recall@5
218
  - type: cosine_recall@10
219
+ value: 0.74
220
  name: Cosine Recall@10
221
  - type: cosine_ndcg@10
222
+ value: 0.5529706231693197
223
  name: Cosine Ndcg@10
224
  - type: cosine_mrr@10
225
+ value: 0.505468253968254
226
  name: Cosine Mrr@10
227
  - type: cosine_map@100
228
+ value: 0.4971957129558554
229
  name: Cosine Map@100
230
  - task:
231
  type: nano-beir
 
235
  type: NanoBEIR_mean
236
  metrics:
237
  - type: cosine_accuracy@1
238
+ value: 0.39
239
  name: Cosine Accuracy@1
240
  - type: cosine_accuracy@3
241
+ value: 0.56
242
  name: Cosine Accuracy@3
243
  - type: cosine_accuracy@5
244
+ value: 0.66
245
  name: Cosine Accuracy@5
246
  - type: cosine_accuracy@10
247
+ value: 0.74
248
  name: Cosine Accuracy@10
249
  - type: cosine_precision@1
250
+ value: 0.39
251
  name: Cosine Precision@1
252
  - type: cosine_precision@3
253
+ value: 0.19
254
  name: Cosine Precision@3
255
  - type: cosine_precision@5
256
  value: 0.138
257
  name: Cosine Precision@5
258
  - type: cosine_precision@10
259
+ value: 0.077
260
  name: Cosine Precision@10
261
  - type: cosine_recall@1
262
+ value: 0.375
263
  name: Cosine Recall@1
264
  - type: cosine_recall@3
265
+ value: 0.55
266
  name: Cosine Recall@3
267
  - type: cosine_recall@5
268
+ value: 0.655
269
  name: Cosine Recall@5
270
  - type: cosine_recall@10
271
+ value: 0.73
272
  name: Cosine Recall@10
273
  - type: cosine_ndcg@10
274
+ value: 0.554486191588552
275
  name: Cosine Ndcg@10
276
  - type: cosine_mrr@10
277
+ value: 0.5039007936507937
278
  name: Cosine Mrr@10
279
  - type: cosine_map@100
280
+ value: 0.5059916632676377
281
  name: Cosine Map@100
282
  ---
283
 
284
+ # SentenceTransformer based on thenlper/gte-small
285
 
286
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [thenlper/gte-small](https://huggingface.co/thenlper/gte-small). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
287
 
288
  ## Model Details
289
 
290
  ### Model Description
291
  - **Model Type:** Sentence Transformer
292
+ - **Base model:** [thenlper/gte-small](https://huggingface.co/thenlper/gte-small) <!-- at revision 17e1f347d17fe144873b1201da91788898c639cd -->
293
  - **Maximum Sequence Length:** 128 tokens
294
+ - **Output Dimensionality:** 384 dimensions
295
  - **Similarity Function:** Cosine Similarity
296
  <!-- - **Training Dataset:** Unknown -->
297
  <!-- - **Language:** Unknown -->
 
307
 
308
  ```
309
  SentenceTransformer(
310
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'BertModel'})
311
+ (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
312
+ (2): Normalize()
313
  )
314
  ```
315
 
 
337
  ]
338
  embeddings = model.encode(sentences)
339
  print(embeddings.shape)
340
+ # [3, 384]
341
 
342
  # Get the similarity scores for the embeddings
343
  similarities = model.similarity(embeddings, embeddings)
344
  print(similarities)
345
+ # tensor([[1.0000, 0.7010, 0.4682],
346
+ # [0.7010, 1.0000, 0.2656],
347
+ # [0.4682, 0.2656, 1.0001]])
348
  ```
349
 
350
  <!--
 
380
  * Datasets: `NanoMSMARCO` and `NanoNQ`
381
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
382
 
383
+ | Metric | NanoMSMARCO | NanoNQ |
384
+ |:--------------------|:------------|:----------|
385
+ | cosine_accuracy@1 | 0.38 | 0.4 |
386
+ | cosine_accuracy@3 | 0.56 | 0.56 |
387
+ | cosine_accuracy@5 | 0.68 | 0.64 |
388
+ | cosine_accuracy@10 | 0.72 | 0.76 |
389
+ | cosine_precision@1 | 0.38 | 0.4 |
390
+ | cosine_precision@3 | 0.1867 | 0.1933 |
391
+ | cosine_precision@5 | 0.136 | 0.14 |
392
+ | cosine_precision@10 | 0.072 | 0.082 |
393
+ | cosine_recall@1 | 0.38 | 0.37 |
394
+ | cosine_recall@3 | 0.56 | 0.54 |
395
+ | cosine_recall@5 | 0.68 | 0.63 |
396
+ | cosine_recall@10 | 0.72 | 0.74 |
397
+ | **cosine_ndcg@10** | **0.556** | **0.553** |
398
+ | cosine_mrr@10 | 0.5023 | 0.5055 |
399
+ | cosine_map@100 | 0.5148 | 0.4972 |
400
 
401
  #### Nano BEIR
402
 
 
414
 
415
  | Metric | Value |
416
  |:--------------------|:-----------|
417
+ | cosine_accuracy@1 | 0.39 |
418
+ | cosine_accuracy@3 | 0.56 |
419
+ | cosine_accuracy@5 | 0.66 |
420
+ | cosine_accuracy@10 | 0.74 |
421
+ | cosine_precision@1 | 0.39 |
422
+ | cosine_precision@3 | 0.19 |
423
  | cosine_precision@5 | 0.138 |
424
+ | cosine_precision@10 | 0.077 |
425
+ | cosine_recall@1 | 0.375 |
426
+ | cosine_recall@3 | 0.55 |
427
+ | cosine_recall@5 | 0.655 |
428
+ | cosine_recall@10 | 0.73 |
429
+ | **cosine_ndcg@10** | **0.5545** |
430
+ | cosine_mrr@10 | 0.5039 |
431
+ | cosine_map@100 | 0.506 |
432
 
433
  <!--
434
  ## Bias, Risks and Limitations
 
451
  * Size: 90,000 training samples
452
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
453
  * Approximate statistics based on the first 1000 samples:
454
+ | | anchor | positive | negative |
455
+ |:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
456
+ | type | string | string | string |
457
+ | details | <ul><li>min: 4 tokens</li><li>mean: 9.18 tokens</li><li>max: 43 tokens</li></ul> | <ul><li>min: 19 tokens</li><li>mean: 78.75 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 17 tokens</li><li>mean: 77.97 tokens</li><li>max: 128 tokens</li></ul> |
458
  * Samples:
459
  | anchor | positive | negative |
460
  |:--------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
 
477
  * Size: 10,000 evaluation samples
478
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
479
  * Approximate statistics based on the first 1000 samples:
480
+ | | anchor | positive | negative |
481
+ |:--------|:---------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------|
482
+ | type | string | string | string |
483
+ | details | <ul><li>min: 4 tokens</li><li>mean: 9.18 tokens</li><li>max: 25 tokens</li></ul> | <ul><li>min: 16 tokens</li><li>mean: 78.83 tokens</li><li>max: 128 tokens</li></ul> | <ul><li>min: 18 tokens</li><li>mean: 76.86 tokens</li><li>max: 128 tokens</li></ul> |
484
  * Samples:
485
  | anchor | positive | negative |
486
  |:----------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
 
645
  ### Training Logs
646
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
647
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
648
+ | 0 | 0 | - | 4.1735 | 0.6259 | 0.6583 | 0.6421 |
649
+ | 0.3556 | 250 | 3.3399 | 1.1023 | 0.5061 | 0.5036 | 0.5048 |
650
+ | 0.7112 | 500 | 1.2152 | 0.9648 | 0.5609 | 0.5241 | 0.5425 |
651
+ | 1.0669 | 750 | 1.1157 | 0.9385 | 0.5568 | 0.5444 | 0.5506 |
652
+ | 1.4225 | 1000 | 1.0721 | 0.9241 | 0.5523 | 0.5520 | 0.5521 |
653
+ | 1.7781 | 1250 | 1.0561 | 0.9159 | 0.5677 | 0.5454 | 0.5565 |
654
+ | 2.1337 | 1500 | 1.0411 | 0.9116 | 0.5505 | 0.5527 | 0.5516 |
655
+ | 2.4893 | 1750 | 1.0248 | 0.9073 | 0.5433 | 0.5631 | 0.5532 |
656
+ | 2.8450 | 2000 | 1.022 | 0.9042 | 0.5424 | 0.5648 | 0.5536 |
657
+ | 3.2006 | 2250 | 1.0079 | 0.9015 | 0.5557 | 0.5626 | 0.5592 |
658
+ | 3.5562 | 2500 | 1.0092 | 0.8998 | 0.5560 | 0.5590 | 0.5575 |
659
+ | 3.9118 | 2750 | 0.9982 | 0.8986 | 0.5560 | 0.5535 | 0.5548 |
660
+ | 4.2674 | 3000 | 0.9947 | 0.8986 | 0.5560 | 0.5530 | 0.5545 |
661
 
662
 
663
  ### Framework Versions
config_sentence_transformers.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "__version__": {
3
  "sentence_transformers": "5.2.0",
4
  "transformers": "4.57.3",
@@ -9,6 +10,5 @@
9
  "document": ""
10
  },
11
  "default_prompt_name": null,
12
- "similarity_fn_name": "cosine",
13
- "model_type": "SentenceTransformer"
14
  }
 
1
  {
2
+ "model_type": "SentenceTransformer",
3
  "__version__": {
4
  "sentence_transformers": "5.2.0",
5
  "transformers": "4.57.3",
 
10
  "document": ""
11
  },
12
  "default_prompt_name": null,
13
+ "similarity_fn_name": "cosine"
 
14
  }
modules.json CHANGED
@@ -10,5 +10,11 @@
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
 
 
 
 
 
 
13
  }
14
  ]
 
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
  }
20
  ]