radoslavralev commited on
Commit
db2f67d
·
verified ·
1 Parent(s): b0a58a0

Add new SentenceTransformer model

Browse files
Files changed (1) hide show
  1. README.md +99 -99
README.md CHANGED
@@ -7,7 +7,7 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:89998
9
  - loss:MultipleNegativesRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
  - source_sentence: Indian university which follow" international education "type system?
13
  sentences:
@@ -57,7 +57,7 @@ metrics:
57
  - cosine_mrr@10
58
  - cosine_map@100
59
  model-index:
60
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
61
  results:
62
  - task:
63
  type: information-retrieval
@@ -67,49 +67,49 @@ model-index:
67
  type: NanoMSMARCO
68
  metrics:
69
  - type: cosine_accuracy@1
70
- value: 0.26
71
  name: Cosine Accuracy@1
72
  - type: cosine_accuracy@3
73
- value: 0.5
74
  name: Cosine Accuracy@3
75
  - type: cosine_accuracy@5
76
- value: 0.6
77
  name: Cosine Accuracy@5
78
  - type: cosine_accuracy@10
79
- value: 0.74
80
  name: Cosine Accuracy@10
81
  - type: cosine_precision@1
82
- value: 0.26
83
  name: Cosine Precision@1
84
  - type: cosine_precision@3
85
- value: 0.16666666666666669
86
  name: Cosine Precision@3
87
  - type: cosine_precision@5
88
- value: 0.12
89
  name: Cosine Precision@5
90
  - type: cosine_precision@10
91
- value: 0.07400000000000001
92
  name: Cosine Precision@10
93
  - type: cosine_recall@1
94
- value: 0.26
95
  name: Cosine Recall@1
96
  - type: cosine_recall@3
97
- value: 0.5
98
  name: Cosine Recall@3
99
  - type: cosine_recall@5
100
- value: 0.6
101
  name: Cosine Recall@5
102
  - type: cosine_recall@10
103
- value: 0.74
104
  name: Cosine Recall@10
105
  - type: cosine_ndcg@10
106
- value: 0.48774998633566824
107
  name: Cosine Ndcg@10
108
  - type: cosine_mrr@10
109
- value: 0.4093333333333333
110
  name: Cosine Mrr@10
111
  - type: cosine_map@100
112
- value: 0.4245357678657921
113
  name: Cosine Map@100
114
  - task:
115
  type: information-retrieval
@@ -119,49 +119,49 @@ model-index:
119
  type: NanoNQ
120
  metrics:
121
  - type: cosine_accuracy@1
122
- value: 0.34
123
  name: Cosine Accuracy@1
124
  - type: cosine_accuracy@3
125
- value: 0.48
126
  name: Cosine Accuracy@3
127
  - type: cosine_accuracy@5
128
- value: 0.6
129
  name: Cosine Accuracy@5
130
  - type: cosine_accuracy@10
131
  value: 0.68
132
  name: Cosine Accuracy@10
133
  - type: cosine_precision@1
134
- value: 0.34
135
  name: Cosine Precision@1
136
  - type: cosine_precision@3
137
- value: 0.16666666666666663
138
  name: Cosine Precision@3
139
  - type: cosine_precision@5
140
- value: 0.124
141
  name: Cosine Precision@5
142
  - type: cosine_precision@10
143
- value: 0.07400000000000001
144
  name: Cosine Precision@10
145
  - type: cosine_recall@1
146
- value: 0.32
147
  name: Cosine Recall@1
148
  - type: cosine_recall@3
149
- value: 0.46
150
  name: Cosine Recall@3
151
  - type: cosine_recall@5
152
- value: 0.57
153
  name: Cosine Recall@5
154
  - type: cosine_recall@10
155
- value: 0.67
156
  name: Cosine Recall@10
157
  - type: cosine_ndcg@10
158
- value: 0.4959822522649102
159
  name: Cosine Ndcg@10
160
  - type: cosine_mrr@10
161
- value: 0.447095238095238
162
  name: Cosine Mrr@10
163
  - type: cosine_map@100
164
- value: 0.4450391558194697
165
  name: Cosine Map@100
166
  - task:
167
  type: nano-beir
@@ -171,61 +171,61 @@ model-index:
171
  type: NanoBEIR_mean
172
  metrics:
173
  - type: cosine_accuracy@1
174
- value: 0.30000000000000004
175
  name: Cosine Accuracy@1
176
  - type: cosine_accuracy@3
177
- value: 0.49
178
  name: Cosine Accuracy@3
179
  - type: cosine_accuracy@5
180
- value: 0.6
181
  name: Cosine Accuracy@5
182
  - type: cosine_accuracy@10
183
- value: 0.71
184
  name: Cosine Accuracy@10
185
  - type: cosine_precision@1
186
- value: 0.30000000000000004
187
  name: Cosine Precision@1
188
  - type: cosine_precision@3
189
- value: 0.16666666666666666
190
  name: Cosine Precision@3
191
  - type: cosine_precision@5
192
- value: 0.122
193
  name: Cosine Precision@5
194
  - type: cosine_precision@10
195
- value: 0.07400000000000001
196
  name: Cosine Precision@10
197
  - type: cosine_recall@1
198
- value: 0.29000000000000004
199
  name: Cosine Recall@1
200
  - type: cosine_recall@3
201
- value: 0.48
202
  name: Cosine Recall@3
203
  - type: cosine_recall@5
204
- value: 0.585
205
  name: Cosine Recall@5
206
  - type: cosine_recall@10
207
- value: 0.7050000000000001
208
  name: Cosine Recall@10
209
  - type: cosine_ndcg@10
210
- value: 0.49186611930028923
211
  name: Cosine Ndcg@10
212
  - type: cosine_mrr@10
213
- value: 0.42821428571428566
214
  name: Cosine Mrr@10
215
  - type: cosine_map@100
216
- value: 0.4347874618426309
217
  name: Cosine Map@100
218
  ---
219
 
220
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
221
 
222
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
223
 
224
  ## Model Details
225
 
226
  ### Model Description
227
  - **Model Type:** Sentence Transformer
228
- - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
229
  - **Maximum Sequence Length:** 128 tokens
230
  - **Output Dimensionality:** 384 dimensions
231
  - **Similarity Function:** Cosine Similarity
@@ -278,9 +278,9 @@ print(embeddings.shape)
278
  # Get the similarity scores for the embeddings
279
  similarities = model.similarity(embeddings, embeddings)
280
  print(similarities)
281
- # tensor([[ 1.0000, 0.7596, 0.0004],
282
- # [ 0.7596, 1.0000, -0.0846],
283
- # [ 0.0004, -0.0846, 1.0000]])
284
  ```
285
 
286
  <!--
@@ -316,23 +316,23 @@ You can finetune this model on your own dataset.
316
  * Datasets: `NanoMSMARCO` and `NanoNQ`
317
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
318
 
319
- | Metric | NanoMSMARCO | NanoNQ |
320
- |:--------------------|:------------|:----------|
321
- | cosine_accuracy@1 | 0.26 | 0.34 |
322
- | cosine_accuracy@3 | 0.5 | 0.48 |
323
- | cosine_accuracy@5 | 0.6 | 0.6 |
324
- | cosine_accuracy@10 | 0.74 | 0.68 |
325
- | cosine_precision@1 | 0.26 | 0.34 |
326
- | cosine_precision@3 | 0.1667 | 0.1667 |
327
- | cosine_precision@5 | 0.12 | 0.124 |
328
- | cosine_precision@10 | 0.074 | 0.074 |
329
- | cosine_recall@1 | 0.26 | 0.32 |
330
- | cosine_recall@3 | 0.5 | 0.46 |
331
- | cosine_recall@5 | 0.6 | 0.57 |
332
- | cosine_recall@10 | 0.74 | 0.67 |
333
- | **cosine_ndcg@10** | **0.4877** | **0.496** |
334
- | cosine_mrr@10 | 0.4093 | 0.4471 |
335
- | cosine_map@100 | 0.4245 | 0.445 |
336
 
337
  #### Nano BEIR
338
 
@@ -348,23 +348,23 @@ You can finetune this model on your own dataset.
348
  }
349
  ```
350
 
351
- | Metric | Value |
352
- |:--------------------|:-----------|
353
- | cosine_accuracy@1 | 0.3 |
354
- | cosine_accuracy@3 | 0.49 |
355
- | cosine_accuracy@5 | 0.6 |
356
- | cosine_accuracy@10 | 0.71 |
357
- | cosine_precision@1 | 0.3 |
358
- | cosine_precision@3 | 0.1667 |
359
- | cosine_precision@5 | 0.122 |
360
- | cosine_precision@10 | 0.074 |
361
- | cosine_recall@1 | 0.29 |
362
- | cosine_recall@3 | 0.48 |
363
- | cosine_recall@5 | 0.585 |
364
- | cosine_recall@10 | 0.705 |
365
- | **cosine_ndcg@10** | **0.4919** |
366
- | cosine_mrr@10 | 0.4282 |
367
- | cosine_map@100 | 0.4348 |
368
 
369
  <!--
370
  ## Bias, Risks and Limitations
@@ -581,19 +581,19 @@ You can finetune this model on your own dataset.
581
  ### Training Logs
582
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
583
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
584
- | 0 | 0 | - | 0.5439 | 0.5540 | 0.5931 | 0.5735 |
585
- | 0.3556 | 250 | 0.61 | 0.4258 | 0.5310 | 0.5623 | 0.5466 |
586
- | 0.7112 | 500 | 0.5484 | 0.4127 | 0.5289 | 0.5387 | 0.5338 |
587
- | 1.0669 | 750 | 0.5286 | 0.4054 | 0.5110 | 0.5322 | 0.5216 |
588
- | 1.4225 | 1000 | 0.5138 | 0.4005 | 0.5065 | 0.5266 | 0.5165 |
589
- | 1.7781 | 1250 | 0.508 | 0.3972 | 0.4863 | 0.5172 | 0.5018 |
590
- | 2.1337 | 1500 | 0.4986 | 0.3955 | 0.4837 | 0.5191 | 0.5014 |
591
- | 2.4893 | 1750 | 0.4936 | 0.3933 | 0.4908 | 0.5175 | 0.5041 |
592
- | 2.8450 | 2000 | 0.4896 | 0.3920 | 0.4867 | 0.4974 | 0.4920 |
593
- | 3.2006 | 2250 | 0.486 | 0.3910 | 0.4820 | 0.4963 | 0.4891 |
594
- | 3.5562 | 2500 | 0.482 | 0.3903 | 0.4814 | 0.4961 | 0.4887 |
595
- | 3.9118 | 2750 | 0.481 | 0.3897 | 0.4877 | 0.4956 | 0.4917 |
596
- | 4.2674 | 3000 | 0.4798 | 0.3894 | 0.4877 | 0.4960 | 0.4919 |
597
 
598
 
599
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:89998
9
  - loss:MultipleNegativesRankingLoss
10
+ base_model: sentence-transformers/all-MiniLM-L12-v2
11
  widget:
12
  - source_sentence: Indian university which follow" international education "type system?
13
  sentences:
 
57
  - cosine_mrr@10
58
  - cosine_map@100
59
  model-index:
60
+ - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
61
  results:
62
  - task:
63
  type: information-retrieval
 
67
  type: NanoMSMARCO
68
  metrics:
69
  - type: cosine_accuracy@1
70
+ value: 0.32
71
  name: Cosine Accuracy@1
72
  - type: cosine_accuracy@3
73
+ value: 0.56
74
  name: Cosine Accuracy@3
75
  - type: cosine_accuracy@5
76
+ value: 0.72
77
  name: Cosine Accuracy@5
78
  - type: cosine_accuracy@10
79
+ value: 0.82
80
  name: Cosine Accuracy@10
81
  - type: cosine_precision@1
82
+ value: 0.32
83
  name: Cosine Precision@1
84
  - type: cosine_precision@3
85
+ value: 0.18666666666666665
86
  name: Cosine Precision@3
87
  - type: cosine_precision@5
88
+ value: 0.14400000000000002
89
  name: Cosine Precision@5
90
  - type: cosine_precision@10
91
+ value: 0.08199999999999999
92
  name: Cosine Precision@10
93
  - type: cosine_recall@1
94
+ value: 0.32
95
  name: Cosine Recall@1
96
  - type: cosine_recall@3
97
+ value: 0.56
98
  name: Cosine Recall@3
99
  - type: cosine_recall@5
100
+ value: 0.72
101
  name: Cosine Recall@5
102
  - type: cosine_recall@10
103
+ value: 0.82
104
  name: Cosine Recall@10
105
  - type: cosine_ndcg@10
106
+ value: 0.5574382679738011
107
  name: Cosine Ndcg@10
108
  - type: cosine_mrr@10
109
+ value: 0.4747460317460317
110
  name: Cosine Mrr@10
111
  - type: cosine_map@100
112
+ value: 0.4820380014583416
113
  name: Cosine Map@100
114
  - task:
115
  type: information-retrieval
 
119
  type: NanoNQ
120
  metrics:
121
  - type: cosine_accuracy@1
122
+ value: 0.32
123
  name: Cosine Accuracy@1
124
  - type: cosine_accuracy@3
125
+ value: 0.54
126
  name: Cosine Accuracy@3
127
  - type: cosine_accuracy@5
128
+ value: 0.62
129
  name: Cosine Accuracy@5
130
  - type: cosine_accuracy@10
131
  value: 0.68
132
  name: Cosine Accuracy@10
133
  - type: cosine_precision@1
134
+ value: 0.32
135
  name: Cosine Precision@1
136
  - type: cosine_precision@3
137
+ value: 0.19333333333333333
138
  name: Cosine Precision@3
139
  - type: cosine_precision@5
140
+ value: 0.132
141
  name: Cosine Precision@5
142
  - type: cosine_precision@10
143
+ value: 0.07200000000000001
144
  name: Cosine Precision@10
145
  - type: cosine_recall@1
146
+ value: 0.31
147
  name: Cosine Recall@1
148
  - type: cosine_recall@3
149
+ value: 0.53
150
  name: Cosine Recall@3
151
  - type: cosine_recall@5
152
+ value: 0.6
153
  name: Cosine Recall@5
154
  - type: cosine_recall@10
155
+ value: 0.66
156
  name: Cosine Recall@10
157
  - type: cosine_ndcg@10
158
+ value: 0.492580214786822
159
  name: Cosine Ndcg@10
160
  - type: cosine_mrr@10
161
+ value: 0.4418809523809524
162
  name: Cosine Mrr@10
163
  - type: cosine_map@100
164
+ value: 0.4462290155539738
165
  name: Cosine Map@100
166
  - task:
167
  type: nano-beir
 
171
  type: NanoBEIR_mean
172
  metrics:
173
  - type: cosine_accuracy@1
174
+ value: 0.32
175
  name: Cosine Accuracy@1
176
  - type: cosine_accuracy@3
177
+ value: 0.55
178
  name: Cosine Accuracy@3
179
  - type: cosine_accuracy@5
180
+ value: 0.6699999999999999
181
  name: Cosine Accuracy@5
182
  - type: cosine_accuracy@10
183
+ value: 0.75
184
  name: Cosine Accuracy@10
185
  - type: cosine_precision@1
186
+ value: 0.32
187
  name: Cosine Precision@1
188
  - type: cosine_precision@3
189
+ value: 0.19
190
  name: Cosine Precision@3
191
  - type: cosine_precision@5
192
+ value: 0.138
193
  name: Cosine Precision@5
194
  - type: cosine_precision@10
195
+ value: 0.077
196
  name: Cosine Precision@10
197
  - type: cosine_recall@1
198
+ value: 0.315
199
  name: Cosine Recall@1
200
  - type: cosine_recall@3
201
+ value: 0.545
202
  name: Cosine Recall@3
203
  - type: cosine_recall@5
204
+ value: 0.6599999999999999
205
  name: Cosine Recall@5
206
  - type: cosine_recall@10
207
+ value: 0.74
208
  name: Cosine Recall@10
209
  - type: cosine_ndcg@10
210
+ value: 0.5250092413803116
211
  name: Cosine Ndcg@10
212
  - type: cosine_mrr@10
213
+ value: 0.45831349206349203
214
  name: Cosine Mrr@10
215
  - type: cosine_map@100
216
+ value: 0.4641335085061577
217
  name: Cosine Map@100
218
  ---
219
 
220
+ # SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
221
 
222
+ This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
223
 
224
  ## Model Details
225
 
226
  ### Model Description
227
  - **Model Type:** Sentence Transformer
228
+ - **Base model:** [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2) <!-- at revision 936af83a2ecce5fe87a09109ff5cbcefe073173a -->
229
  - **Maximum Sequence Length:** 128 tokens
230
  - **Output Dimensionality:** 384 dimensions
231
  - **Similarity Function:** Cosine Similarity
 
278
  # Get the similarity scores for the embeddings
279
  similarities = model.similarity(embeddings, embeddings)
280
  print(similarities)
281
+ # tensor([[ 1.0000, 0.8663, 0.0078],
282
+ # [ 0.8663, 1.0000, -0.0501],
283
+ # [ 0.0078, -0.0501, 1.0000]])
284
  ```
285
 
286
  <!--
 
316
  * Datasets: `NanoMSMARCO` and `NanoNQ`
317
  * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
318
 
319
+ | Metric | NanoMSMARCO | NanoNQ |
320
+ |:--------------------|:------------|:-----------|
321
+ | cosine_accuracy@1 | 0.32 | 0.32 |
322
+ | cosine_accuracy@3 | 0.56 | 0.54 |
323
+ | cosine_accuracy@5 | 0.72 | 0.62 |
324
+ | cosine_accuracy@10 | 0.82 | 0.68 |
325
+ | cosine_precision@1 | 0.32 | 0.32 |
326
+ | cosine_precision@3 | 0.1867 | 0.1933 |
327
+ | cosine_precision@5 | 0.144 | 0.132 |
328
+ | cosine_precision@10 | 0.082 | 0.072 |
329
+ | cosine_recall@1 | 0.32 | 0.31 |
330
+ | cosine_recall@3 | 0.56 | 0.53 |
331
+ | cosine_recall@5 | 0.72 | 0.6 |
332
+ | cosine_recall@10 | 0.82 | 0.66 |
333
+ | **cosine_ndcg@10** | **0.5574** | **0.4926** |
334
+ | cosine_mrr@10 | 0.4747 | 0.4419 |
335
+ | cosine_map@100 | 0.482 | 0.4462 |
336
 
337
  #### Nano BEIR
338
 
 
348
  }
349
  ```
350
 
351
+ | Metric | Value |
352
+ |:--------------------|:----------|
353
+ | cosine_accuracy@1 | 0.32 |
354
+ | cosine_accuracy@3 | 0.55 |
355
+ | cosine_accuracy@5 | 0.67 |
356
+ | cosine_accuracy@10 | 0.75 |
357
+ | cosine_precision@1 | 0.32 |
358
+ | cosine_precision@3 | 0.19 |
359
+ | cosine_precision@5 | 0.138 |
360
+ | cosine_precision@10 | 0.077 |
361
+ | cosine_recall@1 | 0.315 |
362
+ | cosine_recall@3 | 0.545 |
363
+ | cosine_recall@5 | 0.66 |
364
+ | cosine_recall@10 | 0.74 |
365
+ | **cosine_ndcg@10** | **0.525** |
366
+ | cosine_mrr@10 | 0.4583 |
367
+ | cosine_map@100 | 0.4641 |
368
 
369
  <!--
370
  ## Bias, Risks and Limitations
 
581
  ### Training Logs
582
  | Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
583
  |:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
584
+ | 0 | 0 | - | 0.5972 | 0.5887 | 0.5786 | 0.5836 |
585
+ | 0.3556 | 250 | 0.5902 | 0.4140 | 0.5596 | 0.5395 | 0.5495 |
586
+ | 0.7112 | 500 | 0.5168 | 0.4000 | 0.5798 | 0.5206 | 0.5502 |
587
+ | 1.0669 | 750 | 0.4977 | 0.3934 | 0.5722 | 0.5079 | 0.5401 |
588
+ | 1.4225 | 1000 | 0.4825 | 0.3875 | 0.5612 | 0.5129 | 0.5370 |
589
+ | 1.7781 | 1250 | 0.4764 | 0.3843 | 0.5734 | 0.5179 | 0.5457 |
590
+ | 2.1337 | 1500 | 0.4672 | 0.3821 | 0.5740 | 0.5065 | 0.5402 |
591
+ | 2.4893 | 1750 | 0.4612 | 0.3804 | 0.5721 | 0.4950 | 0.5335 |
592
+ | 2.8450 | 2000 | 0.4576 | 0.3791 | 0.5588 | 0.4836 | 0.5212 |
593
+ | 3.2006 | 2250 | 0.4533 | 0.3775 | 0.5550 | 0.5005 | 0.5278 |
594
+ | 3.5562 | 2500 | 0.4491 | 0.3770 | 0.5604 | 0.4919 | 0.5262 |
595
+ | 3.9118 | 2750 | 0.4483 | 0.3763 | 0.5569 | 0.4897 | 0.5233 |
596
+ | 4.2674 | 3000 | 0.446 | 0.3760 | 0.5574 | 0.4926 | 0.5250 |
597
 
598
 
599
  ### Framework Versions