radoslavralev commited on
Commit
569b003
·
verified ·
1 Parent(s): ef4e0fd

Add new SentenceTransformer model

Browse files
1_Pooling/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "word_embedding_dimension": 384,
3
- "pooling_mode_cls_token": false,
4
- "pooling_mode_mean_tokens": true,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
 
1
  {
2
+ "word_embedding_dimension": 768,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
  "pooling_mode_max_tokens": false,
6
  "pooling_mode_mean_sqrt_len_tokens": false,
7
  "pooling_mode_weightedmean_tokens": false,
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:111468
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L12-v2
11
  widget:
12
  - source_sentence: What is something you do (or don’t do), even though you feel conflicted
13
  about it?
@@ -60,7 +60,7 @@ metrics:
60
  - cosine_mrr@10
61
  - cosine_map@100
62
  model-index:
63
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
64
  results:
65
  - task:
66
  type: information-retrieval
@@ -70,49 +70,49 @@ model-index:
70
  type: NanoMSMARCO
71
  metrics:
72
  - type: cosine_accuracy@1
73
- value: 0.22
74
  name: Cosine Accuracy@1
75
  - type: cosine_accuracy@3
76
- value: 0.42
77
  name: Cosine Accuracy@3
78
  - type: cosine_accuracy@5
79
- value: 0.56
80
  name: Cosine Accuracy@5
81
  - type: cosine_accuracy@10
82
- value: 0.64
83
  name: Cosine Accuracy@10
84
  - type: cosine_precision@1
85
- value: 0.22
86
  name: Cosine Precision@1
87
  - type: cosine_precision@3
88
- value: 0.13999999999999999
89
  name: Cosine Precision@3
90
  - type: cosine_precision@5
91
- value: 0.11200000000000002
92
  name: Cosine Precision@5
93
  - type: cosine_precision@10
94
- value: 0.06400000000000002
95
  name: Cosine Precision@10
96
  - type: cosine_recall@1
97
- value: 0.22
98
  name: Cosine Recall@1
99
  - type: cosine_recall@3
100
- value: 0.42
101
  name: Cosine Recall@3
102
  - type: cosine_recall@5
103
- value: 0.56
104
  name: Cosine Recall@5
105
  - type: cosine_recall@10
106
- value: 0.64
107
  name: Cosine Recall@10
108
  - type: cosine_ndcg@10
109
- value: 0.4253611031601517
110
  name: Cosine Ndcg@10
111
  - type: cosine_mrr@10
112
- value: 0.3573571428571428
113
  name: Cosine Mrr@10
114
  - type: cosine_map@100
115
- value: 0.37029891648053065
116
  name: Cosine Map@100
117
  - task:
118
  type: information-retrieval
@@ -122,49 +122,49 @@ model-index:
122
  type: NanoNQ
123
  metrics:
124
  - type: cosine_accuracy@1
125
- value: 0.18
126
  name: Cosine Accuracy@1
127
  - type: cosine_accuracy@3
128
- value: 0.34
129
  name: Cosine Accuracy@3
130
  - type: cosine_accuracy@5
131
- value: 0.4
132
  name: Cosine Accuracy@5
133
  - type: cosine_accuracy@10
134
- value: 0.48
135
  name: Cosine Accuracy@10
136
  - type: cosine_precision@1
137
- value: 0.18
138
  name: Cosine Precision@1
139
  - type: cosine_precision@3
140
- value: 0.11333333333333333
141
  name: Cosine Precision@3
142
  - type: cosine_precision@5
143
- value: 0.08
144
  name: Cosine Precision@5
145
  - type: cosine_precision@10
146
- value: 0.04800000000000001
147
  name: Cosine Precision@10
148
  - type: cosine_recall@1
149
- value: 0.16
150
  name: Cosine Recall@1
151
  - type: cosine_recall@3
152
- value: 0.32
153
  name: Cosine Recall@3
154
  - type: cosine_recall@5
155
- value: 0.37
156
  name: Cosine Recall@5
157
  - type: cosine_recall@10
158
- value: 0.45
159
  name: Cosine Recall@10
160
  - type: cosine_ndcg@10
161
- value: 0.30483462455030724
162
  name: Cosine Ndcg@10
163
  - type: cosine_mrr@10
164
- value: 0.27455555555555555
165
  name: Cosine Mrr@10
166
  - type: cosine_map@100
167
- value: 0.26647869540792385
168
  name: Cosine Map@100
169
  - task:
170
  type: nano-beir
@@ -174,63 +174,63 @@ model-index:
174
  type: NanoBEIR_mean
175
  metrics:
176
  - type: cosine_accuracy@1
177
- value: 0.2
178
  name: Cosine Accuracy@1
179
  - type: cosine_accuracy@3
180
- value: 0.38
181
  name: Cosine Accuracy@3
182
  - type: cosine_accuracy@5
183
- value: 0.48000000000000004
184
  name: Cosine Accuracy@5
185
  - type: cosine_accuracy@10
186
- value: 0.56
187
  name: Cosine Accuracy@10
188
  - type: cosine_precision@1
189
- value: 0.2
190
  name: Cosine Precision@1
191
  - type: cosine_precision@3
192
- value: 0.12666666666666665
193
  name: Cosine Precision@3
194
  - type: cosine_precision@5
195
- value: 0.096
196
  name: Cosine Precision@5
197
  - type: cosine_precision@10
198
- value: 0.05600000000000001
199
  name: Cosine Precision@10
200
  - type: cosine_recall@1
201
- value: 0.19
202
  name: Cosine Recall@1
203
  - type: cosine_recall@3
204
- value: 0.37
205
  name: Cosine Recall@3
206
  - type: cosine_recall@5
207
- value: 0.465
208
  name: Cosine Recall@5
209
  - type: cosine_recall@10
210
- value: 0.545
211
  name: Cosine Recall@10
212
  - type: cosine_ndcg@10
213
- value: 0.36509786385522947
214
  name: Cosine Ndcg@10
215
  - type: cosine_mrr@10
216
- value: 0.3159563492063492
217
  name: Cosine Mrr@10
218
  - type: cosine_map@100
219
- value: 0.31838880594422725
220
  name: Cosine Map@100
221
  ---
222
 
223
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
224
 
225
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
226
 
227
  ## Model Details
228
 
229
  ### Model Description
230
  - **Model Type:** Sentence Transformer
231
- - **Base model:** [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2) <!-- at revision 936af83a2ecce5fe87a09109ff5cbcefe073173a -->
232
  - **Maximum Sequence Length:** 128 tokens
233
- - **Output Dimensionality:** 384 dimensions
234
  - **Similarity Function:** Cosine Similarity
235
  <!-- - **Training Dataset:** Unknown -->
236
  <!-- - **Language:** Unknown -->
@@ -246,9 +246,8 @@ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [s
246
 
247
  ```
248
  SentenceTransformer(
249
- (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'BertModel'})
250
- (1): Pooling({'word_embedding_dimension': 384, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
251
- (2): Normalize()
252
  )
253
  ```
254
 
@@ -276,14 +275,14 @@ sentences = [
276
  ]
277
  embeddings = model.encode(sentences)
278
  print(embeddings.shape)
279
- # [3, 384]
280
 
281
  # Get the similarity scores for the embeddings
282
  similarities = model.similarity(embeddings, embeddings)
283
  print(similarities)
284
- # tensor([[1.0000, 1.0000, 0.4197],
285
- # [1.0000, 1.0000, 0.4197],
286
- # [0.4197, 0.4197, 1.0000]])
287
  ```
288
 
289
  <!--
@@ -321,21 +320,21 @@ You can finetune this model on your own dataset.
321
 
322
  | Metric | NanoMSMARCO | NanoNQ |
323
  |:--------------------|:------------|:-----------|
324
- | cosine_accuracy@1 | 0.22 | 0.18 |
325
- | cosine_accuracy@3 | 0.42 | 0.34 |
326
- | cosine_accuracy@5 | 0.56 | 0.4 |
327
- | cosine_accuracy@10 | 0.64 | 0.48 |
328
- | cosine_precision@1 | 0.22 | 0.18 |
329
- | cosine_precision@3 | 0.14 | 0.1133 |
330
- | cosine_precision@5 | 0.112 | 0.08 |
331
- | cosine_precision@10 | 0.064 | 0.048 |
332
- | cosine_recall@1 | 0.22 | 0.16 |
333
- | cosine_recall@3 | 0.42 | 0.32 |
334
- | cosine_recall@5 | 0.56 | 0.37 |
335
- | cosine_recall@10 | 0.64 | 0.45 |
336
- | **cosine_ndcg@10** | **0.4254** | **0.3048** |
337
- | cosine_mrr@10 | 0.3574 | 0.2746 |
338
- | cosine_map@100 | 0.3703 | 0.2665 |
339
 
340
  #### Nano BEIR
341
 
@@ -351,23 +350,23 @@ You can finetune this model on your own dataset.
351
  }
352
  ```
353
 
354
- | Metric | Value |
355
- |:--------------------|:-----------|
356
- | cosine_accuracy@1 | 0.2 |
357
- | cosine_accuracy@3 | 0.38 |
358
- | cosine_accuracy@5 | 0.48 |
359
- | cosine_accuracy@10 | 0.56 |
360
- | cosine_precision@1 | 0.2 |
361
- | cosine_precision@3 | 0.1267 |
362
- | cosine_precision@5 | 0.096 |
363
- | cosine_precision@10 | 0.056 |
364
- | cosine_recall@1 | 0.19 |
365
- | cosine_recall@3 | 0.37 |
366
- | cosine_recall@5 | 0.465 |
367
- | cosine_recall@10 | 0.545 |
368
- | **cosine_ndcg@10** | **0.3651** |
369
- | cosine_mrr@10 | 0.316 |
370
- | cosine_map@100 | 0.3184 |
371
 
372
  <!--
373
  ## Bias, Risks and Limitations
@@ -390,10 +389,10 @@ You can finetune this model on your own dataset.
390
  * Size: 111,468 training samples
391
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
392
  * Approximate statistics based on the first 1000 samples:
393
- | | anchor | positive | negative |
394
- |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
395
- | type | string | string | string |
396
- | details | <ul><li>min: 6 tokens</li><li>mean: 16.11 tokens</li><li>max: 71 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.16 tokens</li><li>max: 71 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 17.35 tokens</li><li>max: 76 tokens</li></ul> |
397
  * Samples:
398
  | anchor | positive | negative |
399
  |:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------|
@@ -416,10 +415,10 @@ You can finetune this model on your own dataset.
416
  * Size: 12,386 evaluation samples
417
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
418
  * Approximate statistics based on the first 1000 samples:
419
- | | anchor | positive | negative |
420
- |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
421
- | type | string | string | string |
422
- | details | <ul><li>min: 6 tokens</li><li>mean: 16.22 tokens</li><li>max: 62 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.28 tokens</li><li>max: 62 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 17.39 tokens</li><li>max: 66 tokens</li></ul> |
423
  * Samples:
424
  | anchor | positive | negative |
425
  |:------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------|
@@ -584,19 +583,19 @@ You can finetune this model on your own dataset.
584
  ### Training Logs
585
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
586
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
587
- | 0 | 0 | - | 0.6110 | 0.5887 | 0.5786 | 0.5836 |
588
- | 0.2874 | 250 | 0.6051 | 0.4208 | 0.5346 | 0.5338 | 0.5342 |
589
- | 0.5747 | 500 | 0.5163 | 0.3991 | 0.5346 | 0.3965 | 0.4655 |
590
- | 0.8621 | 750 | 0.4918 | 0.3904 | 0.4988 | 0.3821 | 0.4404 |
591
- | 1.1494 | 1000 | 0.4788 | 0.3851 | 0.4855 | 0.3593 | 0.4224 |
592
- | 1.4368 | 1250 | 0.4676 | 0.3814 | 0.4616 | 0.3421 | 0.4019 |
593
- | 1.7241 | 1500 | 0.4636 | 0.3782 | 0.4680 | 0.3354 | 0.4017 |
594
- | 2.0115 | 1750 | 0.4604 | 0.3763 | 0.4306 | 0.3243 | 0.3775 |
595
- | 2.2989 | 2000 | 0.4523 | 0.3749 | 0.4414 | 0.3128 | 0.3771 |
596
- | 2.5862 | 2250 | 0.4497 | 0.3741 | 0.4349 | 0.3106 | 0.3727 |
597
- | 2.8736 | 2500 | 0.4451 | 0.3725 | 0.4324 | 0.3072 | 0.3698 |
598
- | 3.1609 | 2750 | 0.4439 | 0.3721 | 0.4295 | 0.3037 | 0.3666 |
599
- | 3.4483 | 3000 | 0.4411 | 0.3720 | 0.4254 | 0.3048 | 0.3651 |
600
 
601
 
602
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:111468
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: Alibaba-NLP/gte-modernbert-base
11
  widget:
12
  - source_sentence: What is something you do (or don’t do), even though you feel conflicted
13
  about it?
 
60
  - cosine_mrr@10
61
  - cosine_map@100
62
  model-index:
63
+ - name: SentenceTransformer based on Alibaba-NLP/gte-modernbert-base
64
  results:
65
  - task:
66
  type: information-retrieval
 
70
  type: NanoMSMARCO
71
  metrics:
72
  - type: cosine_accuracy@1
73
+ value: 0.16
74
  name: Cosine Accuracy@1
75
  - type: cosine_accuracy@3
76
+ value: 0.3
77
  name: Cosine Accuracy@3
78
  - type: cosine_accuracy@5
79
+ value: 0.32
80
  name: Cosine Accuracy@5
81
  - type: cosine_accuracy@10
82
+ value: 0.4
83
  name: Cosine Accuracy@10
84
  - type: cosine_precision@1
85
+ value: 0.16
86
  name: Cosine Precision@1
87
  - type: cosine_precision@3
88
+ value: 0.09999999999999998
89
  name: Cosine Precision@3
90
  - type: cosine_precision@5
91
+ value: 0.064
92
  name: Cosine Precision@5
93
  - type: cosine_precision@10
94
+ value: 0.04000000000000001
95
  name: Cosine Precision@10
96
  - type: cosine_recall@1
97
+ value: 0.16
98
  name: Cosine Recall@1
99
  - type: cosine_recall@3
100
+ value: 0.3
101
  name: Cosine Recall@3
102
  - type: cosine_recall@5
103
+ value: 0.32
104
  name: Cosine Recall@5
105
  - type: cosine_recall@10
106
+ value: 0.4
107
  name: Cosine Recall@10
108
  - type: cosine_ndcg@10
109
+ value: 0.2797381138841459
110
  name: Cosine Ndcg@10
111
  - type: cosine_mrr@10
112
+ value: 0.24177777777777776
113
  name: Cosine Mrr@10
114
  - type: cosine_map@100
115
+ value: 0.25452652172848145
116
  name: Cosine Map@100
117
  - task:
118
  type: information-retrieval
 
122
  type: NanoNQ
123
  metrics:
124
  - type: cosine_accuracy@1
125
+ value: 0.1
126
  name: Cosine Accuracy@1
127
  - type: cosine_accuracy@3
128
+ value: 0.12
129
  name: Cosine Accuracy@3
130
  - type: cosine_accuracy@5
131
+ value: 0.16
132
  name: Cosine Accuracy@5
133
  - type: cosine_accuracy@10
134
+ value: 0.26
135
  name: Cosine Accuracy@10
136
  - type: cosine_precision@1
137
+ value: 0.1
138
  name: Cosine Precision@1
139
  - type: cosine_precision@3
140
+ value: 0.04
141
  name: Cosine Precision@3
142
  - type: cosine_precision@5
143
+ value: 0.032
144
  name: Cosine Precision@5
145
  - type: cosine_precision@10
146
+ value: 0.026000000000000002
147
  name: Cosine Precision@10
148
  - type: cosine_recall@1
149
+ value: 0.08
150
  name: Cosine Recall@1
151
  - type: cosine_recall@3
152
+ value: 0.09
153
  name: Cosine Recall@3
154
  - type: cosine_recall@5
155
+ value: 0.13
156
  name: Cosine Recall@5
157
  - type: cosine_recall@10
158
+ value: 0.23
159
  name: Cosine Recall@10
160
  - type: cosine_ndcg@10
161
+ value: 0.13826388457522362
162
  name: Cosine Ndcg@10
163
  - type: cosine_mrr@10
164
+ value: 0.12794444444444442
165
  name: Cosine Mrr@10
166
  - type: cosine_map@100
167
+ value: 0.10725400911912068
168
  name: Cosine Map@100
169
  - task:
170
  type: nano-beir
 
174
  type: NanoBEIR_mean
175
  metrics:
176
  - type: cosine_accuracy@1
177
+ value: 0.13
178
  name: Cosine Accuracy@1
179
  - type: cosine_accuracy@3
180
+ value: 0.21
181
  name: Cosine Accuracy@3
182
  - type: cosine_accuracy@5
183
+ value: 0.24
184
  name: Cosine Accuracy@5
185
  - type: cosine_accuracy@10
186
+ value: 0.33
187
  name: Cosine Accuracy@10
188
  - type: cosine_precision@1
189
+ value: 0.13
190
  name: Cosine Precision@1
191
  - type: cosine_precision@3
192
+ value: 0.06999999999999999
193
  name: Cosine Precision@3
194
  - type: cosine_precision@5
195
+ value: 0.048
196
  name: Cosine Precision@5
197
  - type: cosine_precision@10
198
+ value: 0.033
199
  name: Cosine Precision@10
200
  - type: cosine_recall@1
201
+ value: 0.12
202
  name: Cosine Recall@1
203
  - type: cosine_recall@3
204
+ value: 0.195
205
  name: Cosine Recall@3
206
  - type: cosine_recall@5
207
+ value: 0.225
208
  name: Cosine Recall@5
209
  - type: cosine_recall@10
210
+ value: 0.315
211
  name: Cosine Recall@10
212
  - type: cosine_ndcg@10
213
+ value: 0.20900099922968474
214
  name: Cosine Ndcg@10
215
  - type: cosine_mrr@10
216
+ value: 0.1848611111111111
217
  name: Cosine Mrr@10
218
  - type: cosine_map@100
219
+ value: 0.18089026542380107
220
  name: Cosine Map@100
221
  ---
222
 
223
+ # SentenceTransformer based on Alibaba-NLP/gte-modernbert-base
224
 
225
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base). It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
226
 
227
  ## Model Details
228
 
229
  ### Model Description
230
  - **Model Type:** Sentence Transformer
231
+ - **Base model:** [Alibaba-NLP/gte-modernbert-base](https://huggingface.co/Alibaba-NLP/gte-modernbert-base) <!-- at revision e7f32e3c00f91d699e8c43b53106206bcc72bb22 -->
232
  - **Maximum Sequence Length:** 128 tokens
233
+ - **Output Dimensionality:** 768 dimensions
234
  - **Similarity Function:** Cosine Similarity
235
  <!-- - **Training Dataset:** Unknown -->
236
  <!-- - **Language:** Unknown -->
 
246
 
247
  ```
248
  SentenceTransformer(
249
+ (0): Transformer({'max_seq_length': 128, 'do_lower_case': False, 'architecture': 'ModernBertModel'})
250
+ (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True})
 
251
  )
252
  ```
253
 
 
275
  ]
276
  embeddings = model.encode(sentences)
277
  print(embeddings.shape)
278
+ # [3, 768]
279
 
280
  # Get the similarity scores for the embeddings
281
  similarities = model.similarity(embeddings, embeddings)
282
  print(similarities)
283
+ # tensor([[1.0000, 1.0000, 0.1015],
284
+ # [1.0000, 1.0000, 0.1015],
285
+ # [0.1015, 0.1015, 1.0000]])
286
  ```
287
 
288
  <!--
 
320
 
321
  | Metric | NanoMSMARCO | NanoNQ |
322
  |:--------------------|:------------|:-----------|
323
+ | cosine_accuracy@1 | 0.16 | 0.1 |
324
+ | cosine_accuracy@3 | 0.3 | 0.12 |
325
+ | cosine_accuracy@5 | 0.32 | 0.16 |
326
+ | cosine_accuracy@10 | 0.4 | 0.26 |
327
+ | cosine_precision@1 | 0.16 | 0.1 |
328
+ | cosine_precision@3 | 0.1 | 0.04 |
329
+ | cosine_precision@5 | 0.064 | 0.032 |
330
+ | cosine_precision@10 | 0.04 | 0.026 |
331
+ | cosine_recall@1 | 0.16 | 0.08 |
332
+ | cosine_recall@3 | 0.3 | 0.09 |
333
+ | cosine_recall@5 | 0.32 | 0.13 |
334
+ | cosine_recall@10 | 0.4 | 0.23 |
335
+ | **cosine_ndcg@10** | **0.2797** | **0.1383** |
336
+ | cosine_mrr@10 | 0.2418 | 0.1279 |
337
+ | cosine_map@100 | 0.2545 | 0.1073 |
338
 
339
  #### Nano BEIR
340
 
 
350
  }
351
  ```
352
 
353
+ | Metric | Value |
354
+ |:--------------------|:----------|
355
+ | cosine_accuracy@1 | 0.13 |
356
+ | cosine_accuracy@3 | 0.21 |
357
+ | cosine_accuracy@5 | 0.24 |
358
+ | cosine_accuracy@10 | 0.33 |
359
+ | cosine_precision@1 | 0.13 |
360
+ | cosine_precision@3 | 0.07 |
361
+ | cosine_precision@5 | 0.048 |
362
+ | cosine_precision@10 | 0.033 |
363
+ | cosine_recall@1 | 0.12 |
364
+ | cosine_recall@3 | 0.195 |
365
+ | cosine_recall@5 | 0.225 |
366
+ | cosine_recall@10 | 0.315 |
367
+ | **cosine_ndcg@10** | **0.209** |
368
+ | cosine_mrr@10 | 0.1849 |
369
+ | cosine_map@100 | 0.1809 |
370
 
371
  <!--
372
  ## Bias, Risks and Limitations
 
389
  * Size: 111,468 training samples
390
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
391
  * Approximate statistics based on the first 1000 samples:
392
+ | | anchor | positive | negative |
393
+ |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
394
+ | type | string | string | string |
395
+ | details | <ul><li>min: 6 tokens</li><li>mean: 16.1 tokens</li><li>max: 68 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.16 tokens</li><li>max: 68 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 17.35 tokens</li><li>max: 76 tokens</li></ul> |
396
  * Samples:
397
  | anchor | positive | negative |
398
  |:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------|
 
415
  * Size: 12,386 evaluation samples
416
  * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
417
  * Approximate statistics based on the first 1000 samples:
418
+ | | anchor | positive | negative |
419
+ |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
420
+ | type | string | string | string |
421
+ | details | <ul><li>min: 6 tokens</li><li>mean: 16.3 tokens</li><li>max: 62 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.36 tokens</li><li>max: 62 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 17.41 tokens</li><li>max: 65 tokens</li></ul> |
422
  * Samples:
423
  | anchor | positive | negative |
424
  |:------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------|
 
583
  ### Training Logs
584
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
585
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
586
+ | 0 | 0 | - | 2.1536 | 0.6530 | 0.6552 | 0.6541 |
587
+ | 0.2874 | 250 | 0.8149 | 0.3782 | 0.5748 | 0.6239 | 0.5994 |
588
+ | 0.5747 | 500 | 0.3649 | 0.3561 | 0.5529 | 0.5583 | 0.5556 |
589
+ | 0.8621 | 750 | 0.3496 | 0.3456 | 0.5747 | 0.4299 | 0.5023 |
590
+ | 1.1494 | 1000 | 0.3364 | 0.3416 | 0.5479 | 0.4669 | 0.5074 |
591
+ | 1.4368 | 1250 | 0.3231 | 0.3385 | 0.4802 | 0.4214 | 0.4508 |
592
+ | 1.7241 | 1500 | 0.3222 | 0.3349 | 0.4694 | 0.3547 | 0.4120 |
593
+ | 2.0115 | 1750 | 0.3187 | 0.3340 | 0.4398 | 0.2723 | 0.3560 |
594
+ | 2.2989 | 2000 | 0.3051 | 0.3347 | 0.3702 | 0.2144 | 0.2923 |
595
+ | 2.5862 | 2250 | 0.3039 | 0.3339 | 0.4180 | 0.2575 | 0.3377 |
596
+ | 2.8736 | 2500 | 0.2994 | 0.3328 | 0.3687 | 0.2662 | 0.3174 |
597
+ | 3.1609 | 2750 | 0.2969 | 0.3339 | 0.2906 | 0.1283 | 0.2094 |
598
+ | 3.4483 | 3000 | 0.2933 | 0.3343 | 0.2797 | 0.1383 | 0.2090 |
599
 
600
 
601
  ### Framework Versions
config_sentence_transformers.json CHANGED
@@ -4,11 +4,11 @@
4
  "transformers": "4.57.3",
5
  "pytorch": "2.9.1+cu128"
6
  },
7
- "model_type": "SentenceTransformer",
8
  "prompts": {
9
  "query": "",
10
  "document": ""
11
  },
12
  "default_prompt_name": null,
13
- "similarity_fn_name": "cosine"
 
14
  }
 
4
  "transformers": "4.57.3",
5
  "pytorch": "2.9.1+cu128"
6
  },
 
7
  "prompts": {
8
  "query": "",
9
  "document": ""
10
  },
11
  "default_prompt_name": null,
12
+ "similarity_fn_name": "cosine",
13
+ "model_type": "SentenceTransformer"
14
  }
modules.json CHANGED
@@ -10,11 +10,5 @@
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
13
- },
14
- {
15
- "idx": 2,
16
- "name": "2",
17
- "path": "2_Normalize",
18
- "type": "sentence_transformers.models.Normalize"
19
  }
20
  ]
 
10
  "name": "1",
11
  "path": "1_Pooling",
12
  "type": "sentence_transformers.models.Pooling"
 
 
 
 
 
 
13
  }
14
  ]