radoslavralev commited on
Commit
9535836
·
verified ·
1 Parent(s): be1bca8

Add new SentenceTransformer model

Browse files
Files changed (2) hide show
  1. README.md +92 -92
  2. config_sentence_transformers.json +1 -1
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:90000
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: thenlper/gte-small
11
  widget:
12
  - source_sentence: what is the maximum i can contribute to a traditional ira
13
  sentences:
@@ -121,7 +121,7 @@ metrics:
121
  - cosine_mrr@10
122
  - cosine_map@100
123
  model-index:
124
- - name: SentenceTransformer based on thenlper/gte-small
125
  results:
126
  - task:
127
  type: information-retrieval
@@ -134,46 +134,46 @@ model-index:
134
  value: 0.32
135
  name: Cosine Accuracy@1
136
  - type: cosine_accuracy@3
137
- value: 0.56
138
  name: Cosine Accuracy@3
139
  - type: cosine_accuracy@5
140
  value: 0.62
141
  name: Cosine Accuracy@5
142
  - type: cosine_accuracy@10
143
- value: 0.74
144
  name: Cosine Accuracy@10
145
  - type: cosine_precision@1
146
  value: 0.32
147
  name: Cosine Precision@1
148
  - type: cosine_precision@3
149
- value: 0.18666666666666668
150
  name: Cosine Precision@3
151
  - type: cosine_precision@5
152
  value: 0.124
153
  name: Cosine Precision@5
154
  - type: cosine_precision@10
155
- value: 0.07400000000000001
156
  name: Cosine Precision@10
157
  - type: cosine_recall@1
158
  value: 0.32
159
  name: Cosine Recall@1
160
  - type: cosine_recall@3
161
- value: 0.56
162
  name: Cosine Recall@3
163
  - type: cosine_recall@5
164
  value: 0.62
165
  name: Cosine Recall@5
166
  - type: cosine_recall@10
167
- value: 0.74
168
  name: Cosine Recall@10
169
  - type: cosine_ndcg@10
170
- value: 0.522258803283639
171
  name: Cosine Ndcg@10
172
  - type: cosine_mrr@10
173
- value: 0.4534126984126985
174
  name: Cosine Mrr@10
175
  - type: cosine_map@100
176
- value: 0.4638252994171221
177
  name: Cosine Map@100
178
  - task:
179
  type: information-retrieval
@@ -183,49 +183,49 @@ model-index:
183
  type: NanoNQ
184
  metrics:
185
  - type: cosine_accuracy@1
186
- value: 0.32
187
  name: Cosine Accuracy@1
188
  - type: cosine_accuracy@3
189
- value: 0.54
190
  name: Cosine Accuracy@3
191
  - type: cosine_accuracy@5
192
- value: 0.54
193
  name: Cosine Accuracy@5
194
  - type: cosine_accuracy@10
195
- value: 0.74
196
  name: Cosine Accuracy@10
197
  - type: cosine_precision@1
198
- value: 0.32
199
  name: Cosine Precision@1
200
  - type: cosine_precision@3
201
- value: 0.18666666666666665
202
  name: Cosine Precision@3
203
  - type: cosine_precision@5
204
- value: 0.11600000000000002
205
  name: Cosine Precision@5
206
  - type: cosine_precision@10
207
- value: 0.07800000000000001
208
  name: Cosine Precision@10
209
  - type: cosine_recall@1
210
- value: 0.29
211
  name: Cosine Recall@1
212
  - type: cosine_recall@3
213
- value: 0.52
214
  name: Cosine Recall@3
215
  - type: cosine_recall@5
216
- value: 0.53
217
  name: Cosine Recall@5
218
  - type: cosine_recall@10
219
- value: 0.71
220
  name: Cosine Recall@10
221
  - type: cosine_ndcg@10
222
- value: 0.5000975716574514
223
  name: Cosine Ndcg@10
224
  - type: cosine_mrr@10
225
- value: 0.44801587301587303
226
  name: Cosine Mrr@10
227
  - type: cosine_map@100
228
- value: 0.435720694785091
229
  name: Cosine Map@100
230
  - task:
231
  type: nano-beir
@@ -235,61 +235,61 @@ model-index:
235
  type: NanoBEIR_mean
236
  metrics:
237
  - type: cosine_accuracy@1
238
- value: 0.32
239
  name: Cosine Accuracy@1
240
  - type: cosine_accuracy@3
241
- value: 0.55
242
  name: Cosine Accuracy@3
243
  - type: cosine_accuracy@5
244
- value: 0.5800000000000001
245
  name: Cosine Accuracy@5
246
  - type: cosine_accuracy@10
247
- value: 0.74
248
  name: Cosine Accuracy@10
249
  - type: cosine_precision@1
250
- value: 0.32
251
  name: Cosine Precision@1
252
  - type: cosine_precision@3
253
- value: 0.18666666666666665
254
  name: Cosine Precision@3
255
  - type: cosine_precision@5
256
- value: 0.12000000000000001
257
  name: Cosine Precision@5
258
  - type: cosine_precision@10
259
- value: 0.07600000000000001
260
  name: Cosine Precision@10
261
  - type: cosine_recall@1
262
- value: 0.305
263
  name: Cosine Recall@1
264
  - type: cosine_recall@3
265
- value: 0.54
266
  name: Cosine Recall@3
267
  - type: cosine_recall@5
268
- value: 0.575
269
  name: Cosine Recall@5
270
  - type: cosine_recall@10
271
- value: 0.725
272
  name: Cosine Recall@10
273
  - type: cosine_ndcg@10
274
- value: 0.5111781874705452
275
  name: Cosine Ndcg@10
276
  - type: cosine_mrr@10
277
- value: 0.45071428571428573
278
  name: Cosine Mrr@10
279
  - type: cosine_map@100
280
- value: 0.4497729971011065
281
  name: Cosine Map@100
282
  ---
283
 
284
- # SentenceTransformer based on thenlper/gte-small
285
 
286
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [thenlper/gte-small](https://huggingface.co/thenlper/gte-small). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
287
 
288
  ## Model Details
289
 
290
  ### Model Description
291
  - **Model Type:** Sentence Transformer
292
- - **Base model:** [thenlper/gte-small](https://huggingface.co/thenlper/gte-small) <!-- at revision 17e1f347d17fe144873b1201da91788898c639cd -->
293
  - **Maximum Sequence Length:** 128 tokens
294
  - **Output Dimensionality:** 384 dimensions
295
  - **Similarity Function:** Cosine Similarity
@@ -342,9 +342,9 @@ print(embeddings.shape)
342
  # Get the similarity scores for the embeddings
343
  similarities = model.similarity(embeddings, embeddings)
344
  print(similarities)
345
- # tensor([[1.0000, 0.5348, 0.4287],
346
- # [0.5348, 1.0000, 0.3212],
347
- # [0.4287, 0.3212, 1.0000]])
348
  ```
349
 
350
  <!--
@@ -382,21 +382,21 @@ You can finetune this model on your own dataset.
382
 
383
  | Metric | NanoMSMARCO | NanoNQ |
384
  |:--------------------|:------------|:-----------|
385
- | cosine_accuracy@1 | 0.32 | 0.32 |
386
- | cosine_accuracy@3 | 0.56 | 0.54 |
387
- | cosine_accuracy@5 | 0.62 | 0.54 |
388
- | cosine_accuracy@10 | 0.74 | 0.74 |
389
- | cosine_precision@1 | 0.32 | 0.32 |
390
- | cosine_precision@3 | 0.1867 | 0.1867 |
391
- | cosine_precision@5 | 0.124 | 0.116 |
392
- | cosine_precision@10 | 0.074 | 0.078 |
393
- | cosine_recall@1 | 0.32 | 0.29 |
394
- | cosine_recall@3 | 0.56 | 0.52 |
395
- | cosine_recall@5 | 0.62 | 0.53 |
396
- | cosine_recall@10 | 0.74 | 0.71 |
397
- | **cosine_ndcg@10** | **0.5223** | **0.5001** |
398
- | cosine_mrr@10 | 0.4534 | 0.448 |
399
- | cosine_map@100 | 0.4638 | 0.4357 |
400
 
401
  #### Nano BEIR
402
 
@@ -414,21 +414,21 @@ You can finetune this model on your own dataset.
414
 
415
  | Metric | Value |
416
  |:--------------------|:-----------|
417
- | cosine_accuracy@1 | 0.32 |
418
- | cosine_accuracy@3 | 0.55 |
419
- | cosine_accuracy@5 | 0.58 |
420
- | cosine_accuracy@10 | 0.74 |
421
- | cosine_precision@1 | 0.32 |
422
- | cosine_precision@3 | 0.1867 |
423
- | cosine_precision@5 | 0.12 |
424
- | cosine_precision@10 | 0.076 |
425
- | cosine_recall@1 | 0.305 |
426
- | cosine_recall@3 | 0.54 |
427
- | cosine_recall@5 | 0.575 |
428
- | cosine_recall@10 | 0.725 |
429
- | **cosine_ndcg@10** | **0.5112** |
430
- | cosine_mrr@10 | 0.4507 |
431
- | cosine_map@100 | 0.4498 |
432
 
433
  <!--
434
  ## Bias, Risks and Limitations
@@ -502,7 +502,7 @@ You can finetune this model on your own dataset.
502
  - `eval_strategy`: steps
503
  - `per_device_train_batch_size`: 128
504
  - `per_device_eval_batch_size`: 128
505
- - `learning_rate`: 1e-06
506
  - `weight_decay`: 0.001
507
  - `max_steps`: 3000
508
  - `warmup_ratio`: 0.1
@@ -531,7 +531,7 @@ You can finetune this model on your own dataset.
531
  - `gradient_accumulation_steps`: 1
532
  - `eval_accumulation_steps`: None
533
  - `torch_empty_cache_steps`: None
534
- - `learning_rate`: 1e-06
535
  - `weight_decay`: 0.001
536
  - `adam_beta1`: 0.9
537
  - `adam_beta2`: 0.999
@@ -645,19 +645,19 @@ You can finetune this model on your own dataset.
645
  ### Training Logs
646
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
647
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
648
- | 0 | 0 | - | 4.1735 | 0.6259 | 0.6583 | 0.6421 |
649
- | 0.3556 | 250 | 4.3112 | 3.9428 | 0.6062 | 0.6473 | 0.6268 |
650
- | 0.7112 | 500 | 3.8846 | 3.1810 | 0.5977 | 0.6251 | 0.6114 |
651
- | 1.0669 | 750 | 2.9479 | 1.8567 | 0.5628 | 0.5082 | 0.5355 |
652
- | 1.4225 | 1000 | 1.9026 | 1.2819 | 0.5229 | 0.4809 | 0.5019 |
653
- | 1.7781 | 1250 | 1.5448 | 1.1522 | 0.5112 | 0.4845 | 0.4978 |
654
- | 2.1337 | 1500 | 1.4335 | 1.1065 | 0.5183 | 0.4899 | 0.5041 |
655
- | 2.4893 | 1750 | 1.3748 | 1.0820 | 0.5231 | 0.4896 | 0.5064 |
656
- | 2.8450 | 2000 | 1.3445 | 1.0667 | 0.5184 | 0.4962 | 0.5073 |
657
- | 3.2006 | 2250 | 1.3173 | 1.0577 | 0.5231 | 0.4971 | 0.5101 |
658
- | 3.5562 | 2500 | 1.3099 | 1.0517 | 0.5165 | 0.4999 | 0.5082 |
659
- | 3.9118 | 2750 | 1.2945 | 1.0483 | 0.5218 | 0.4999 | 0.5108 |
660
- | 4.2674 | 3000 | 1.288 | 1.0476 | 0.5223 | 0.5001 | 0.5112 |
661
 
662
 
663
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:90000
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
  - source_sentence: what is the maximum i can contribute to a traditional ira
13
  sentences:
 
121
  - cosine_mrr@10
122
  - cosine_map@100
123
  model-index:
124
+ - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
125
  results:
126
  - task:
127
  type: information-retrieval
 
134
  value: 0.32
135
  name: Cosine Accuracy@1
136
  - type: cosine_accuracy@3
137
+ value: 0.54
138
  name: Cosine Accuracy@3
139
  - type: cosine_accuracy@5
140
  value: 0.62
141
  name: Cosine Accuracy@5
142
  - type: cosine_accuracy@10
143
+ value: 0.72
144
  name: Cosine Accuracy@10
145
  - type: cosine_precision@1
146
  value: 0.32
147
  name: Cosine Precision@1
148
  - type: cosine_precision@3
149
+ value: 0.18
150
  name: Cosine Precision@3
151
  - type: cosine_precision@5
152
  value: 0.124
153
  name: Cosine Precision@5
154
  - type: cosine_precision@10
155
+ value: 0.07200000000000001
156
  name: Cosine Precision@10
157
  - type: cosine_recall@1
158
  value: 0.32
159
  name: Cosine Recall@1
160
  - type: cosine_recall@3
161
+ value: 0.54
162
  name: Cosine Recall@3
163
  - type: cosine_recall@5
164
  value: 0.62
165
  name: Cosine Recall@5
166
  - type: cosine_recall@10
167
+ value: 0.72
168
  name: Cosine Recall@10
169
  - type: cosine_ndcg@10
170
+ value: 0.5113600890173079
171
  name: Cosine Ndcg@10
172
  - type: cosine_mrr@10
173
+ value: 0.44588095238095227
174
  name: Cosine Mrr@10
175
  - type: cosine_map@100
176
+ value: 0.46015815095025
177
  name: Cosine Map@100
178
  - task:
179
  type: information-retrieval
 
183
  type: NanoNQ
184
  metrics:
185
  - type: cosine_accuracy@1
186
+ value: 0.34
187
  name: Cosine Accuracy@1
188
  - type: cosine_accuracy@3
189
+ value: 0.5
190
  name: Cosine Accuracy@3
191
  - type: cosine_accuracy@5
192
+ value: 0.58
193
  name: Cosine Accuracy@5
194
  - type: cosine_accuracy@10
195
+ value: 0.62
196
  name: Cosine Accuracy@10
197
  - type: cosine_precision@1
198
+ value: 0.34
199
  name: Cosine Precision@1
200
  - type: cosine_precision@3
201
+ value: 0.18
202
  name: Cosine Precision@3
203
  - type: cosine_precision@5
204
+ value: 0.12400000000000003
205
  name: Cosine Precision@5
206
  - type: cosine_precision@10
207
+ value: 0.068
208
  name: Cosine Precision@10
209
  - type: cosine_recall@1
210
+ value: 0.31
211
  name: Cosine Recall@1
212
  - type: cosine_recall@3
213
+ value: 0.49
214
  name: Cosine Recall@3
215
  - type: cosine_recall@5
216
+ value: 0.56
217
  name: Cosine Recall@5
218
  - type: cosine_recall@10
219
+ value: 0.61
220
  name: Cosine Recall@10
221
  - type: cosine_ndcg@10
222
+ value: 0.47606860320855016
223
  name: Cosine Ndcg@10
224
  - type: cosine_mrr@10
225
+ value: 0.44116666666666665
226
  name: Cosine Mrr@10
227
  - type: cosine_map@100
228
+ value: 0.4435242837774235
229
  name: Cosine Map@100
230
  - task:
231
  type: nano-beir
 
235
  type: NanoBEIR_mean
236
  metrics:
237
  - type: cosine_accuracy@1
238
+ value: 0.33
239
  name: Cosine Accuracy@1
240
  - type: cosine_accuracy@3
241
+ value: 0.52
242
  name: Cosine Accuracy@3
243
  - type: cosine_accuracy@5
244
+ value: 0.6
245
  name: Cosine Accuracy@5
246
  - type: cosine_accuracy@10
247
+ value: 0.6699999999999999
248
  name: Cosine Accuracy@10
249
  - type: cosine_precision@1
250
+ value: 0.33
251
  name: Cosine Precision@1
252
  - type: cosine_precision@3
253
+ value: 0.18
254
  name: Cosine Precision@3
255
  - type: cosine_precision@5
256
+ value: 0.12400000000000001
257
  name: Cosine Precision@5
258
  - type: cosine_precision@10
259
+ value: 0.07
260
  name: Cosine Precision@10
261
  - type: cosine_recall@1
262
+ value: 0.315
263
  name: Cosine Recall@1
264
  - type: cosine_recall@3
265
+ value: 0.515
266
  name: Cosine Recall@3
267
  - type: cosine_recall@5
268
+ value: 0.5900000000000001
269
  name: Cosine Recall@5
270
  - type: cosine_recall@10
271
+ value: 0.665
272
  name: Cosine Recall@10
273
  - type: cosine_ndcg@10
274
+ value: 0.493714346112929
275
  name: Cosine Ndcg@10
276
  - type: cosine_mrr@10
277
+ value: 0.44352380952380943
278
  name: Cosine Mrr@10
279
  - type: cosine_map@100
280
+ value: 0.45184121736383676
281
  name: Cosine Map@100
282
  ---
283
 
284
+ # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
285
 
286
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
287
 
288
  ## Model Details
289
 
290
  ### Model Description
291
  - **Model Type:** Sentence Transformer
292
+ - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
293
  - **Maximum Sequence Length:** 128 tokens
294
  - **Output Dimensionality:** 384 dimensions
295
  - **Similarity Function:** Cosine Similarity
 
342
  # Get the similarity scores for the embeddings
343
  similarities = model.similarity(embeddings, embeddings)
344
  print(similarities)
345
+ # tensor([[1.0000, 0.5834, 0.4906],
346
+ # [0.5834, 1.0000, 0.2633],
347
+ # [0.4906, 0.2633, 1.0000]])
348
  ```
349
 
350
  <!--
 
382
 
383
  | Metric | NanoMSMARCO | NanoNQ |
384
  |:--------------------|:------------|:-----------|
385
+ | cosine_accuracy@1 | 0.32 | 0.34 |
386
+ | cosine_accuracy@3 | 0.54 | 0.5 |
387
+ | cosine_accuracy@5 | 0.62 | 0.58 |
388
+ | cosine_accuracy@10 | 0.72 | 0.62 |
389
+ | cosine_precision@1 | 0.32 | 0.34 |
390
+ | cosine_precision@3 | 0.18 | 0.18 |
391
+ | cosine_precision@5 | 0.124 | 0.124 |
392
+ | cosine_precision@10 | 0.072 | 0.068 |
393
+ | cosine_recall@1 | 0.32 | 0.31 |
394
+ | cosine_recall@3 | 0.54 | 0.49 |
395
+ | cosine_recall@5 | 0.62 | 0.56 |
396
+ | cosine_recall@10 | 0.72 | 0.61 |
397
+ | **cosine_ndcg@10** | **0.5114** | **0.4761** |
398
+ | cosine_mrr@10 | 0.4459 | 0.4412 |
399
+ | cosine_map@100 | 0.4602 | 0.4435 |
400
 
401
  #### Nano BEIR
402
 
 
414
 
415
  | Metric | Value |
416
  |:--------------------|:-----------|
417
+ | cosine_accuracy@1 | 0.33 |
418
+ | cosine_accuracy@3 | 0.52 |
419
+ | cosine_accuracy@5 | 0.6 |
420
+ | cosine_accuracy@10 | 0.67 |
421
+ | cosine_precision@1 | 0.33 |
422
+ | cosine_precision@3 | 0.18 |
423
+ | cosine_precision@5 | 0.124 |
424
+ | cosine_precision@10 | 0.07 |
425
+ | cosine_recall@1 | 0.315 |
426
+ | cosine_recall@3 | 0.515 |
427
+ | cosine_recall@5 | 0.59 |
428
+ | cosine_recall@10 | 0.665 |
429
+ | **cosine_ndcg@10** | **0.4937** |
430
+ | cosine_mrr@10 | 0.4435 |
431
+ | cosine_map@100 | 0.4518 |
432
 
433
  <!--
434
  ## Bias, Risks and Limitations
 
502
  - `eval_strategy`: steps
503
  - `per_device_train_batch_size`: 128
504
  - `per_device_eval_batch_size`: 128
505
+ - `learning_rate`: 1e-05
506
  - `weight_decay`: 0.001
507
  - `max_steps`: 3000
508
  - `warmup_ratio`: 0.1
 
531
  - `gradient_accumulation_steps`: 1
532
  - `eval_accumulation_steps`: None
533
  - `torch_empty_cache_steps`: None
534
+ - `learning_rate`: 1e-05
535
  - `weight_decay`: 0.001
536
  - `adam_beta1`: 0.9
537
  - `adam_beta2`: 0.999
 
645
  ### Training Logs
646
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
647
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
648
+ | 0 | 0 | - | 1.2542 | 0.5540 | 0.5931 | 0.5735 |
649
+ | 0.3556 | 250 | 1.2668 | 0.9745 | 0.5174 | 0.5317 | 0.5246 |
650
+ | 0.7112 | 500 | 1.1283 | 0.9461 | 0.5013 | 0.5121 | 0.5067 |
651
+ | 1.0669 | 750 | 1.1038 | 0.9351 | 0.5139 | 0.5015 | 0.5077 |
652
+ | 1.4225 | 1000 | 1.0775 | 0.9264 | 0.5211 | 0.5268 | 0.5239 |
653
+ | 1.7781 | 1250 | 1.0688 | 0.9222 | 0.5290 | 0.4973 | 0.5132 |
654
+ | 2.1337 | 1500 | 1.0564 | 0.9181 | 0.5224 | 0.4688 | 0.4956 |
655
+ | 2.4893 | 1750 | 1.0411 | 0.9167 | 0.5210 | 0.5014 | 0.5112 |
656
+ | 2.8450 | 2000 | 1.0432 | 0.9141 | 0.5257 | 0.4661 | 0.4959 |
657
+ | 3.2006 | 2250 | 1.0299 | 0.9124 | 0.5122 | 0.4895 | 0.5008 |
658
+ | 3.5562 | 2500 | 1.0308 | 0.9112 | 0.5131 | 0.4848 | 0.4989 |
659
+ | 3.9118 | 2750 | 1.0231 | 0.9106 | 0.5106 | 0.4843 | 0.4975 |
660
+ | 4.2674 | 3000 | 1.0183 | 0.9105 | 0.5114 | 0.4761 | 0.4937 |
661
 
662
 
663
  ### Framework Versions
config_sentence_transformers.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "model_type": "SentenceTransformer",
3
  "__version__": {
4
  "sentence_transformers": "5.2.0",
5
  "transformers": "4.57.3",
6
  "pytorch": "2.9.1+cu128"
7
  },
 
8
  "prompts": {
9
  "query": "",
10
  "document": ""
 
1
  {
 
2
  "__version__": {
3
  "sentence_transformers": "5.2.0",
4
  "transformers": "4.57.3",
5
  "pytorch": "2.9.1+cu128"
6
  },
7
+ "model_type": "SentenceTransformer",
8
  "prompts": {
9
  "query": "",
10
  "document": ""