radoslavralev commited on
Commit
20e41ce
·
verified ·
1 Parent(s): 7bce0e7

Add new SentenceTransformer model

Browse files
Files changed (1) hide show
  1. README.md +95 -95
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:111470
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
  - source_sentence: why are some rocks radioactive
13
  sentences:
@@ -106,7 +106,7 @@ metrics:
106
  - cosine_mrr@10
107
  - cosine_map@100
108
  model-index:
109
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
110
  results:
111
  - task:
112
  type: information-retrieval
@@ -116,49 +116,49 @@ model-index:
116
  type: NanoMSMARCO
117
  metrics:
118
  - type: cosine_accuracy@1
119
- value: 0.32
120
  name: Cosine Accuracy@1
121
  - type: cosine_accuracy@3
122
- value: 0.54
123
  name: Cosine Accuracy@3
124
  - type: cosine_accuracy@5
125
- value: 0.62
126
  name: Cosine Accuracy@5
127
  - type: cosine_accuracy@10
128
- value: 0.74
129
  name: Cosine Accuracy@10
130
  - type: cosine_precision@1
131
- value: 0.32
132
  name: Cosine Precision@1
133
  - type: cosine_precision@3
134
- value: 0.18
135
  name: Cosine Precision@3
136
  - type: cosine_precision@5
137
- value: 0.124
138
  name: Cosine Precision@5
139
  - type: cosine_precision@10
140
- value: 0.07400000000000001
141
  name: Cosine Precision@10
142
  - type: cosine_recall@1
143
- value: 0.32
144
  name: Cosine Recall@1
145
  - type: cosine_recall@3
146
- value: 0.54
147
  name: Cosine Recall@3
148
  - type: cosine_recall@5
149
- value: 0.62
150
  name: Cosine Recall@5
151
  - type: cosine_recall@10
152
- value: 0.74
153
  name: Cosine Recall@10
154
  - type: cosine_ndcg@10
155
- value: 0.5189661789845982
156
  name: Cosine Ndcg@10
157
  - type: cosine_mrr@10
158
- value: 0.4499126984126983
159
  name: Cosine Mrr@10
160
  - type: cosine_map@100
161
- value: 0.4616522119994847
162
  name: Cosine Map@100
163
  - task:
164
  type: information-retrieval
@@ -168,49 +168,49 @@ model-index:
168
  type: NanoNQ
169
  metrics:
170
  - type: cosine_accuracy@1
171
- value: 0.34
172
  name: Cosine Accuracy@1
173
  - type: cosine_accuracy@3
174
- value: 0.54
175
  name: Cosine Accuracy@3
176
  - type: cosine_accuracy@5
177
- value: 0.58
178
  name: Cosine Accuracy@5
179
  - type: cosine_accuracy@10
180
  value: 0.66
181
  name: Cosine Accuracy@10
182
  - type: cosine_precision@1
183
- value: 0.34
184
  name: Cosine Precision@1
185
  - type: cosine_precision@3
186
  value: 0.19333333333333333
187
  name: Cosine Precision@3
188
  - type: cosine_precision@5
189
- value: 0.12400000000000003
190
  name: Cosine Precision@5
191
  - type: cosine_precision@10
192
- value: 0.07200000000000001
193
  name: Cosine Precision@10
194
  - type: cosine_recall@1
195
- value: 0.31
196
  name: Cosine Recall@1
197
  - type: cosine_recall@3
198
- value: 0.52
199
  name: Cosine Recall@3
200
  - type: cosine_recall@5
201
- value: 0.56
202
  name: Cosine Recall@5
203
  - type: cosine_recall@10
204
- value: 0.65
205
  name: Cosine Recall@10
206
  - type: cosine_ndcg@10
207
- value: 0.495936841468424
208
  name: Cosine Ndcg@10
209
  - type: cosine_mrr@10
210
- value: 0.455388888888889
211
  name: Cosine Mrr@10
212
  - type: cosine_map@100
213
- value: 0.45517920150439706
214
  name: Cosine Map@100
215
  - task:
216
  type: nano-beir
@@ -220,61 +220,61 @@ model-index:
220
  type: NanoBEIR_mean
221
  metrics:
222
  - type: cosine_accuracy@1
223
- value: 0.33
224
  name: Cosine Accuracy@1
225
  - type: cosine_accuracy@3
226
- value: 0.54
227
  name: Cosine Accuracy@3
228
  - type: cosine_accuracy@5
229
- value: 0.6
230
  name: Cosine Accuracy@5
231
  - type: cosine_accuracy@10
232
- value: 0.7
233
  name: Cosine Accuracy@10
234
  - type: cosine_precision@1
235
- value: 0.33
236
  name: Cosine Precision@1
237
  - type: cosine_precision@3
238
- value: 0.18666666666666665
239
  name: Cosine Precision@3
240
  - type: cosine_precision@5
241
- value: 0.12400000000000001
242
  name: Cosine Precision@5
243
  - type: cosine_precision@10
244
  value: 0.07300000000000001
245
  name: Cosine Precision@10
246
  - type: cosine_recall@1
247
- value: 0.315
248
  name: Cosine Recall@1
249
  - type: cosine_recall@3
250
- value: 0.53
251
  name: Cosine Recall@3
252
  - type: cosine_recall@5
253
- value: 0.5900000000000001
254
  name: Cosine Recall@5
255
  - type: cosine_recall@10
256
- value: 0.6950000000000001
257
  name: Cosine Recall@10
258
  - type: cosine_ndcg@10
259
- value: 0.5074515102265111
260
  name: Cosine Ndcg@10
261
  - type: cosine_mrr@10
262
- value: 0.45265079365079364
263
  name: Cosine Mrr@10
264
  - type: cosine_map@100
265
- value: 0.4584157067519409
266
  name: Cosine Map@100
267
  ---
268
 
269
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
270
 
271
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
272
 
273
  ## Model Details
274
 
275
  ### Model Description
276
  - **Model Type:** Sentence Transformer
277
- - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
278
  - **Maximum Sequence Length:** 128 tokens
279
  - **Output Dimensionality:** 384 dimensions
280
  - **Similarity Function:** Cosine Similarity
@@ -327,9 +327,9 @@ print(embeddings.shape)
327
  # Get the similarity scores for the embeddings
328
  similarities = model.similarity(embeddings, embeddings)
329
  print(similarities)
330
- # tensor([[1.0000, 1.0000, 0.9705],
331
- # [1.0000, 1.0000, 0.9705],
332
- # [0.9705, 0.9705, 1.0000]])
333
  ```
334
 
335
  <!--
@@ -367,21 +367,21 @@ You can finetune this model on your own dataset.
367
 
368
  | Metric | NanoMSMARCO | NanoNQ |
369
  |:--------------------|:------------|:-----------|
370
- | cosine_accuracy@1 | 0.32 | 0.34 |
371
- | cosine_accuracy@3 | 0.54 | 0.54 |
372
- | cosine_accuracy@5 | 0.62 | 0.58 |
373
- | cosine_accuracy@10 | 0.74 | 0.66 |
374
- | cosine_precision@1 | 0.32 | 0.34 |
375
- | cosine_precision@3 | 0.18 | 0.1933 |
376
- | cosine_precision@5 | 0.124 | 0.124 |
377
- | cosine_precision@10 | 0.074 | 0.072 |
378
- | cosine_recall@1 | 0.32 | 0.31 |
379
- | cosine_recall@3 | 0.54 | 0.52 |
380
- | cosine_recall@5 | 0.62 | 0.56 |
381
- | cosine_recall@10 | 0.74 | 0.65 |
382
- | **cosine_ndcg@10** | **0.519** | **0.4959** |
383
- | cosine_mrr@10 | 0.4499 | 0.4554 |
384
- | cosine_map@100 | 0.4617 | 0.4552 |
385
 
386
  #### Nano BEIR
387
 
@@ -397,23 +397,23 @@ You can finetune this model on your own dataset.
397
  }
398
  ```
399
 
400
- | Metric | Value |
401
- |:--------------------|:-----------|
402
- | cosine_accuracy@1 | 0.33 |
403
- | cosine_accuracy@3 | 0.54 |
404
- | cosine_accuracy@5 | 0.6 |
405
- | cosine_accuracy@10 | 0.7 |
406
- | cosine_precision@1 | 0.33 |
407
- | cosine_precision@3 | 0.1867 |
408
- | cosine_precision@5 | 0.124 |
409
- | cosine_precision@10 | 0.073 |
410
- | cosine_recall@1 | 0.315 |
411
- | cosine_recall@3 | 0.53 |
412
- | cosine_recall@5 | 0.59 |
413
- | cosine_recall@10 | 0.695 |
414
- | **cosine_ndcg@10** | **0.5075** |
415
- | cosine_mrr@10 | 0.4527 |
416
- | cosine_map@100 | 0.4584 |
417
 
418
  <!--
419
  ## Bias, Risks and Limitations
@@ -630,19 +630,19 @@ You can finetune this model on your own dataset.
630
  ### Training Logs
631
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
632
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
633
- | 0 | 0 | - | 1.1445 | 0.5540 | 0.5931 | 0.5735 |
634
- | 0.2874 | 250 | 1.1834 | 0.9135 | 0.4900 | 0.5098 | 0.4999 |
635
- | 0.5747 | 500 | 1.0424 | 0.8651 | 0.4979 | 0.4970 | 0.4974 |
636
- | 0.8621 | 750 | 1.0133 | 0.8524 | 0.5118 | 0.4987 | 0.5052 |
637
- | 1.1494 | 1000 | 0.9885 | 0.8441 | 0.5232 | 0.5085 | 0.5159 |
638
- | 1.4368 | 1250 | 0.9788 | 0.8383 | 0.5204 | 0.5039 | 0.5122 |
639
- | 1.7241 | 1500 | 0.9689 | 0.8339 | 0.5184 | 0.5220 | 0.5202 |
640
- | 2.0115 | 1750 | 0.9641 | 0.8292 | 0.5192 | 0.4993 | 0.5093 |
641
- | 2.2989 | 2000 | 0.9523 | 0.8269 | 0.5173 | 0.4896 | 0.5034 |
642
- | 2.5862 | 2250 | 0.9492 | 0.8248 | 0.5181 | 0.5132 | 0.5157 |
643
- | 2.8736 | 2500 | 0.936 | 0.8241 | 0.5187 | 0.5082 | 0.5134 |
644
- | 3.1609 | 2750 | 0.9427 | 0.8228 | 0.5193 | 0.4957 | 0.5075 |
645
- | 3.4483 | 3000 | 0.9363 | 0.8225 | 0.5190 | 0.4959 | 0.5075 |
646
 
647
 
648
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:111470
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: sentence-transformers/all-MiniLM-L12-v2
11
  widget:
12
  - source_sentence: why are some rocks radioactive
13
  sentences:
 
106
  - cosine_mrr@10
107
  - cosine_map@100
108
  model-index:
109
+ - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
110
  results:
111
  - task:
112
  type: information-retrieval
 
116
  type: NanoMSMARCO
117
  metrics:
118
  - type: cosine_accuracy@1
119
+ value: 0.36
120
  name: Cosine Accuracy@1
121
  - type: cosine_accuracy@3
122
+ value: 0.58
123
  name: Cosine Accuracy@3
124
  - type: cosine_accuracy@5
125
+ value: 0.68
126
  name: Cosine Accuracy@5
127
  - type: cosine_accuracy@10
128
+ value: 0.76
129
  name: Cosine Accuracy@10
130
  - type: cosine_precision@1
131
+ value: 0.36
132
  name: Cosine Precision@1
133
  - type: cosine_precision@3
134
+ value: 0.19333333333333333
135
  name: Cosine Precision@3
136
  - type: cosine_precision@5
137
+ value: 0.136
138
  name: Cosine Precision@5
139
  - type: cosine_precision@10
140
+ value: 0.07600000000000001
141
  name: Cosine Precision@10
142
  - type: cosine_recall@1
143
+ value: 0.36
144
  name: Cosine Recall@1
145
  - type: cosine_recall@3
146
+ value: 0.58
147
  name: Cosine Recall@3
148
  - type: cosine_recall@5
149
+ value: 0.68
150
  name: Cosine Recall@5
151
  - type: cosine_recall@10
152
+ value: 0.76
153
  name: Cosine Recall@10
154
  - type: cosine_ndcg@10
155
+ value: 0.5522122196843595
156
  name: Cosine Ndcg@10
157
  - type: cosine_mrr@10
158
+ value: 0.48633333333333334
159
  name: Cosine Mrr@10
160
  - type: cosine_map@100
161
+ value: 0.4958040190560276
162
  name: Cosine Map@100
163
  - task:
164
  type: information-retrieval
 
168
  type: NanoNQ
169
  metrics:
170
  - type: cosine_accuracy@1
171
+ value: 0.44
172
  name: Cosine Accuracy@1
173
  - type: cosine_accuracy@3
174
+ value: 0.56
175
  name: Cosine Accuracy@3
176
  - type: cosine_accuracy@5
177
+ value: 0.6
178
  name: Cosine Accuracy@5
179
  - type: cosine_accuracy@10
180
  value: 0.66
181
  name: Cosine Accuracy@10
182
  - type: cosine_precision@1
183
+ value: 0.44
184
  name: Cosine Precision@1
185
  - type: cosine_precision@3
186
  value: 0.19333333333333333
187
  name: Cosine Precision@3
188
  - type: cosine_precision@5
189
+ value: 0.128
190
  name: Cosine Precision@5
191
  - type: cosine_precision@10
192
+ value: 0.07
193
  name: Cosine Precision@10
194
  - type: cosine_recall@1
195
+ value: 0.41
196
  name: Cosine Recall@1
197
  - type: cosine_recall@3
198
+ value: 0.53
199
  name: Cosine Recall@3
200
  - type: cosine_recall@5
201
+ value: 0.58
202
  name: Cosine Recall@5
203
  - type: cosine_recall@10
204
+ value: 0.64
205
  name: Cosine Recall@10
206
  - type: cosine_ndcg@10
207
+ value: 0.5277371826996979
208
  name: Cosine Ndcg@10
209
  - type: cosine_mrr@10
210
+ value: 0.505079365079365
211
  name: Cosine Mrr@10
212
  - type: cosine_map@100
213
+ value: 0.5002333969355556
214
  name: Cosine Map@100
215
  - task:
216
  type: nano-beir
 
220
  type: NanoBEIR_mean
221
  metrics:
222
  - type: cosine_accuracy@1
223
+ value: 0.4
224
  name: Cosine Accuracy@1
225
  - type: cosine_accuracy@3
226
+ value: 0.5700000000000001
227
  name: Cosine Accuracy@3
228
  - type: cosine_accuracy@5
229
+ value: 0.64
230
  name: Cosine Accuracy@5
231
  - type: cosine_accuracy@10
232
+ value: 0.71
233
  name: Cosine Accuracy@10
234
  - type: cosine_precision@1
235
+ value: 0.4
236
  name: Cosine Precision@1
237
  - type: cosine_precision@3
238
+ value: 0.19333333333333333
239
  name: Cosine Precision@3
240
  - type: cosine_precision@5
241
+ value: 0.132
242
  name: Cosine Precision@5
243
  - type: cosine_precision@10
244
  value: 0.07300000000000001
245
  name: Cosine Precision@10
246
  - type: cosine_recall@1
247
+ value: 0.385
248
  name: Cosine Recall@1
249
  - type: cosine_recall@3
250
+ value: 0.5549999999999999
251
  name: Cosine Recall@3
252
  - type: cosine_recall@5
253
+ value: 0.63
254
  name: Cosine Recall@5
255
  - type: cosine_recall@10
256
+ value: 0.7
257
  name: Cosine Recall@10
258
  - type: cosine_ndcg@10
259
+ value: 0.5399747011920287
260
  name: Cosine Ndcg@10
261
  - type: cosine_mrr@10
262
+ value: 0.49570634920634915
263
  name: Cosine Mrr@10
264
  - type: cosine_map@100
265
+ value: 0.4980187079957916
266
  name: Cosine Map@100
267
  ---
268
 
269
+ # SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
270
 
271
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
272
 
273
  ## Model Details
274
 
275
  ### Model Description
276
  - **Model Type:** Sentence Transformer
277
+ - **Base model:** [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2) <!-- at revision 936af83a2ecce5fe87a09109ff5cbcefe073173a -->
278
  - **Maximum Sequence Length:** 128 tokens
279
  - **Output Dimensionality:** 384 dimensions
280
  - **Similarity Function:** Cosine Similarity
 
327
  # Get the similarity scores for the embeddings
328
  similarities = model.similarity(embeddings, embeddings)
329
  print(similarities)
330
+ # tensor([[0.9999, 0.9999, 0.9775],
331
+ # [0.9999, 0.9999, 0.9775],
332
+ # [0.9775, 0.9775, 1.0000]])
333
  ```
334
 
335
  <!--
 
367
 
368
  | Metric | NanoMSMARCO | NanoNQ |
369
  |:--------------------|:------------|:-----------|
370
+ | cosine_accuracy@1 | 0.36 | 0.44 |
371
+ | cosine_accuracy@3 | 0.58 | 0.56 |
372
+ | cosine_accuracy@5 | 0.68 | 0.6 |
373
+ | cosine_accuracy@10 | 0.76 | 0.66 |
374
+ | cosine_precision@1 | 0.36 | 0.44 |
375
+ | cosine_precision@3 | 0.1933 | 0.1933 |
376
+ | cosine_precision@5 | 0.136 | 0.128 |
377
+ | cosine_precision@10 | 0.076 | 0.07 |
378
+ | cosine_recall@1 | 0.36 | 0.41 |
379
+ | cosine_recall@3 | 0.58 | 0.53 |
380
+ | cosine_recall@5 | 0.68 | 0.58 |
381
+ | cosine_recall@10 | 0.76 | 0.64 |
382
+ | **cosine_ndcg@10** | **0.5522** | **0.5277** |
383
+ | cosine_mrr@10 | 0.4863 | 0.5051 |
384
+ | cosine_map@100 | 0.4958 | 0.5002 |
385
 
386
  #### Nano BEIR
387
 
 
397
  }
398
  ```
399
 
400
+ | Metric | Value |
401
+ |:--------------------|:---------|
402
+ | cosine_accuracy@1 | 0.4 |
403
+ | cosine_accuracy@3 | 0.57 |
404
+ | cosine_accuracy@5 | 0.64 |
405
+ | cosine_accuracy@10 | 0.71 |
406
+ | cosine_precision@1 | 0.4 |
407
+ | cosine_precision@3 | 0.1933 |
408
+ | cosine_precision@5 | 0.132 |
409
+ | cosine_precision@10 | 0.073 |
410
+ | cosine_recall@1 | 0.385 |
411
+ | cosine_recall@3 | 0.555 |
412
+ | cosine_recall@5 | 0.63 |
413
+ | cosine_recall@10 | 0.7 |
414
+ | **cosine_ndcg@10** | **0.54** |
415
+ | cosine_mrr@10 | 0.4957 |
416
+ | cosine_map@100 | 0.498 |
417
 
418
  <!--
419
  ## Bias, Risks and Limitations
 
630
  ### Training Logs
631
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
632
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
633
+ | 0 | 0 | - | 1.1142 | 0.5887 | 0.5786 | 0.5836 |
634
+ | 0.2874 | 250 | 1.1156 | 0.8606 | 0.5535 | 0.5483 | 0.5509 |
635
+ | 0.5747 | 500 | 0.972 | 0.8202 | 0.5383 | 0.5521 | 0.5452 |
636
+ | 0.8621 | 750 | 0.9475 | 0.8071 | 0.5445 | 0.5372 | 0.5409 |
637
+ | 1.1494 | 1000 | 0.924 | 0.8011 | 0.5444 | 0.5591 | 0.5517 |
638
+ | 1.4368 | 1250 | 0.9108 | 0.7953 | 0.5390 | 0.5372 | 0.5381 |
639
+ | 1.7241 | 1500 | 0.9024 | 0.7896 | 0.5473 | 0.5381 | 0.5427 |
640
+ | 2.0115 | 1750 | 0.8971 | 0.7864 | 0.5539 | 0.5201 | 0.5370 |
641
+ | 2.2989 | 2000 | 0.8815 | 0.7842 | 0.5580 | 0.5329 | 0.5455 |
642
+ | 2.5862 | 2250 | 0.8791 | 0.7824 | 0.5569 | 0.5286 | 0.5427 |
643
+ | 2.8736 | 2500 | 0.8667 | 0.7821 | 0.5565 | 0.5188 | 0.5376 |
644
+ | 3.1609 | 2750 | 0.8709 | 0.7807 | 0.5536 | 0.5213 | 0.5374 |
645
+ | 3.4483 | 3000 | 0.8642 | 0.7802 | 0.5522 | 0.5277 | 0.5400 |
646
 
647
 
648
  ### Framework Versions