Sentence Similarity
sentence-transformers
Safetensors
bert
feature-extraction
dense
Generated from Trainer
dataset_size:111470
loss:MultipleNegativesRankingLoss
Eval Results (legacy)
text-embeddings-inference
Instructions to use redis/model-b-structured with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- sentence-transformers
How to use redis/model-b-structured with sentence-transformers:
from sentence_transformers import SentenceTransformer model = SentenceTransformer("redis/model-b-structured") sentences = [ "when was the first elephant brought to america", "Old Bet The first elephant brought to the United States was in 1796, aboard the America which set sail from Calcutta for New York on December 3, 1795.[4] However, it is not certain that this was Old Bet.[2] The first references to Old Bet start in 1804 in Boston as part of a menagerie.[1] In 1808, while residing in Somers, New York, Hachaliah Bailey purchased the menagerie elephant for $1,000 and named it \"Old Bet\".[5][6]", "Cronus Rhea secretly gave birth to Zeus in Crete, and handed Cronus a stone wrapped in swaddling clothes, also known as the Omphalos Stone, which he promptly swallowed, thinking that it was his son.", "Renal artery One or two accessory renal arteries are frequently found, especially on the left side since they usually arise from the aorta, and may come off above (more common) or below the main artery. Instead of entering the kidney at the hilus, they usually pierce the upper or lower part of the organ." ] embeddings = model.encode(sentences) similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [4, 4] - Notebooks
- Google Colab
- Kaggle
Add new SentenceTransformer model
Browse files
README.md
CHANGED
|
@@ -7,7 +7,7 @@ tags:
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:111470
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
-
base_model: sentence-transformers/all-MiniLM-
|
| 11 |
widget:
|
| 12 |
- source_sentence: when was the first elephant brought to america
|
| 13 |
sentences:
|
|
@@ -132,7 +132,7 @@ metrics:
|
|
| 132 |
- cosine_mrr@10
|
| 133 |
- cosine_map@100
|
| 134 |
model-index:
|
| 135 |
-
- name: SentenceTransformer based on sentence-transformers/all-MiniLM-
|
| 136 |
results:
|
| 137 |
- task:
|
| 138 |
type: information-retrieval
|
|
@@ -142,49 +142,49 @@ model-index:
|
|
| 142 |
type: NanoMSMARCO
|
| 143 |
metrics:
|
| 144 |
- type: cosine_accuracy@1
|
| 145 |
-
value: 0.
|
| 146 |
name: Cosine Accuracy@1
|
| 147 |
- type: cosine_accuracy@3
|
| 148 |
value: 0.5
|
| 149 |
name: Cosine Accuracy@3
|
| 150 |
- type: cosine_accuracy@5
|
| 151 |
-
value: 0.
|
| 152 |
name: Cosine Accuracy@5
|
| 153 |
- type: cosine_accuracy@10
|
| 154 |
-
value: 0.
|
| 155 |
name: Cosine Accuracy@10
|
| 156 |
- type: cosine_precision@1
|
| 157 |
-
value: 0.
|
| 158 |
name: Cosine Precision@1
|
| 159 |
- type: cosine_precision@3
|
| 160 |
value: 0.16666666666666663
|
| 161 |
name: Cosine Precision@3
|
| 162 |
- type: cosine_precision@5
|
| 163 |
-
value: 0.
|
| 164 |
name: Cosine Precision@5
|
| 165 |
- type: cosine_precision@10
|
| 166 |
-
value: 0.
|
| 167 |
name: Cosine Precision@10
|
| 168 |
- type: cosine_recall@1
|
| 169 |
-
value: 0.
|
| 170 |
name: Cosine Recall@1
|
| 171 |
- type: cosine_recall@3
|
| 172 |
value: 0.5
|
| 173 |
name: Cosine Recall@3
|
| 174 |
- type: cosine_recall@5
|
| 175 |
-
value: 0.
|
| 176 |
name: Cosine Recall@5
|
| 177 |
- type: cosine_recall@10
|
| 178 |
-
value: 0.
|
| 179 |
name: Cosine Recall@10
|
| 180 |
- type: cosine_ndcg@10
|
| 181 |
-
value: 0.
|
| 182 |
name: Cosine Ndcg@10
|
| 183 |
- type: cosine_mrr@10
|
| 184 |
-
value: 0.
|
| 185 |
name: Cosine Mrr@10
|
| 186 |
- type: cosine_map@100
|
| 187 |
-
value: 0.
|
| 188 |
name: Cosine Map@100
|
| 189 |
- task:
|
| 190 |
type: information-retrieval
|
|
@@ -194,49 +194,49 @@ model-index:
|
|
| 194 |
type: NanoNQ
|
| 195 |
metrics:
|
| 196 |
- type: cosine_accuracy@1
|
| 197 |
-
value: 0.
|
| 198 |
name: Cosine Accuracy@1
|
| 199 |
- type: cosine_accuracy@3
|
| 200 |
value: 0.66
|
| 201 |
name: Cosine Accuracy@3
|
| 202 |
- type: cosine_accuracy@5
|
| 203 |
-
value: 0.
|
| 204 |
name: Cosine Accuracy@5
|
| 205 |
- type: cosine_accuracy@10
|
| 206 |
-
value: 0.
|
| 207 |
name: Cosine Accuracy@10
|
| 208 |
- type: cosine_precision@1
|
| 209 |
-
value: 0.
|
| 210 |
name: Cosine Precision@1
|
| 211 |
- type: cosine_precision@3
|
| 212 |
-
value: 0.
|
| 213 |
name: Cosine Precision@3
|
| 214 |
- type: cosine_precision@5
|
| 215 |
-
value: 0.
|
| 216 |
name: Cosine Precision@5
|
| 217 |
- type: cosine_precision@10
|
| 218 |
-
value: 0.
|
| 219 |
name: Cosine Precision@10
|
| 220 |
- type: cosine_recall@1
|
| 221 |
-
value: 0.
|
| 222 |
name: Cosine Recall@1
|
| 223 |
- type: cosine_recall@3
|
| 224 |
-
value: 0.
|
| 225 |
name: Cosine Recall@3
|
| 226 |
- type: cosine_recall@5
|
| 227 |
-
value: 0.
|
| 228 |
name: Cosine Recall@5
|
| 229 |
- type: cosine_recall@10
|
| 230 |
-
value: 0.
|
| 231 |
name: Cosine Recall@10
|
| 232 |
- type: cosine_ndcg@10
|
| 233 |
-
value: 0.
|
| 234 |
name: Cosine Ndcg@10
|
| 235 |
- type: cosine_mrr@10
|
| 236 |
-
value: 0.
|
| 237 |
name: Cosine Mrr@10
|
| 238 |
- type: cosine_map@100
|
| 239 |
-
value: 0.
|
| 240 |
name: Cosine Map@100
|
| 241 |
- task:
|
| 242 |
type: nano-beir
|
|
@@ -246,61 +246,61 @@ model-index:
|
|
| 246 |
type: NanoBEIR_mean
|
| 247 |
metrics:
|
| 248 |
- type: cosine_accuracy@1
|
| 249 |
-
value: 0.
|
| 250 |
name: Cosine Accuracy@1
|
| 251 |
- type: cosine_accuracy@3
|
| 252 |
value: 0.5800000000000001
|
| 253 |
name: Cosine Accuracy@3
|
| 254 |
- type: cosine_accuracy@5
|
| 255 |
-
value: 0.
|
| 256 |
name: Cosine Accuracy@5
|
| 257 |
- type: cosine_accuracy@10
|
| 258 |
-
value: 0.
|
| 259 |
name: Cosine Accuracy@10
|
| 260 |
- type: cosine_precision@1
|
| 261 |
-
value: 0.
|
| 262 |
name: Cosine Precision@1
|
| 263 |
- type: cosine_precision@3
|
| 264 |
-
value: 0.
|
| 265 |
name: Cosine Precision@3
|
| 266 |
- type: cosine_precision@5
|
| 267 |
-
value: 0.
|
| 268 |
name: Cosine Precision@5
|
| 269 |
- type: cosine_precision@10
|
| 270 |
-
value: 0.
|
| 271 |
name: Cosine Precision@10
|
| 272 |
- type: cosine_recall@1
|
| 273 |
-
value: 0.
|
| 274 |
name: Cosine Recall@1
|
| 275 |
- type: cosine_recall@3
|
| 276 |
-
value: 0.
|
| 277 |
name: Cosine Recall@3
|
| 278 |
- type: cosine_recall@5
|
| 279 |
-
value: 0.
|
| 280 |
name: Cosine Recall@5
|
| 281 |
- type: cosine_recall@10
|
| 282 |
-
value: 0.
|
| 283 |
name: Cosine Recall@10
|
| 284 |
- type: cosine_ndcg@10
|
| 285 |
-
value: 0.
|
| 286 |
name: Cosine Ndcg@10
|
| 287 |
- type: cosine_mrr@10
|
| 288 |
-
value: 0.
|
| 289 |
name: Cosine Mrr@10
|
| 290 |
- type: cosine_map@100
|
| 291 |
-
value: 0.
|
| 292 |
name: Cosine Map@100
|
| 293 |
---
|
| 294 |
|
| 295 |
-
# SentenceTransformer based on sentence-transformers/all-MiniLM-
|
| 296 |
|
| 297 |
-
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-
|
| 298 |
|
| 299 |
## Model Details
|
| 300 |
|
| 301 |
### Model Description
|
| 302 |
- **Model Type:** Sentence Transformer
|
| 303 |
-
- **Base model:** [sentence-transformers/all-MiniLM-
|
| 304 |
- **Maximum Sequence Length:** 128 tokens
|
| 305 |
- **Output Dimensionality:** 384 dimensions
|
| 306 |
- **Similarity Function:** Cosine Similarity
|
|
@@ -353,9 +353,9 @@ print(embeddings.shape)
|
|
| 353 |
# Get the similarity scores for the embeddings
|
| 354 |
similarities = model.similarity(embeddings, embeddings)
|
| 355 |
print(similarities)
|
| 356 |
-
# tensor([[1.0000, 1.0000, 0.
|
| 357 |
-
# [1.0000, 1.0000, 0.
|
| 358 |
-
# [0.
|
| 359 |
```
|
| 360 |
|
| 361 |
<!--
|
|
@@ -393,21 +393,21 @@ You can finetune this model on your own dataset.
|
|
| 393 |
|
| 394 |
| Metric | NanoMSMARCO | NanoNQ |
|
| 395 |
|:--------------------|:------------|:-----------|
|
| 396 |
-
| cosine_accuracy@1 | 0.
|
| 397 |
| cosine_accuracy@3 | 0.5 | 0.66 |
|
| 398 |
-
| cosine_accuracy@5 | 0.
|
| 399 |
-
| cosine_accuracy@10 | 0.
|
| 400 |
-
| cosine_precision@1 | 0.
|
| 401 |
-
| cosine_precision@3 | 0.1667 | 0.
|
| 402 |
-
| cosine_precision@5 | 0.
|
| 403 |
-
| cosine_precision@10 | 0.
|
| 404 |
-
| cosine_recall@1 | 0.
|
| 405 |
-
| cosine_recall@3 | 0.5 | 0.
|
| 406 |
-
| cosine_recall@5 | 0.
|
| 407 |
-
| cosine_recall@10 | 0.
|
| 408 |
-
| **cosine_ndcg@10** | **0.
|
| 409 |
-
| cosine_mrr@10 | 0.
|
| 410 |
-
| cosine_map@100 | 0.
|
| 411 |
|
| 412 |
#### Nano BEIR
|
| 413 |
|
|
@@ -425,21 +425,21 @@ You can finetune this model on your own dataset.
|
|
| 425 |
|
| 426 |
| Metric | Value |
|
| 427 |
|:--------------------|:-----------|
|
| 428 |
-
| cosine_accuracy@1 | 0.
|
| 429 |
| cosine_accuracy@3 | 0.58 |
|
| 430 |
-
| cosine_accuracy@5 | 0.
|
| 431 |
-
| cosine_accuracy@10 | 0.
|
| 432 |
-
| cosine_precision@1 | 0.
|
| 433 |
-
| cosine_precision@3 | 0.
|
| 434 |
-
| cosine_precision@5 | 0.
|
| 435 |
-
| cosine_precision@10 | 0.
|
| 436 |
-
| cosine_recall@1 | 0.
|
| 437 |
-
| cosine_recall@3 | 0.
|
| 438 |
-
| cosine_recall@5 | 0.
|
| 439 |
-
| cosine_recall@10 | 0.
|
| 440 |
-
| **cosine_ndcg@10** | **0.
|
| 441 |
-
| cosine_mrr@10 | 0.
|
| 442 |
-
| cosine_map@100 | 0.
|
| 443 |
|
| 444 |
<!--
|
| 445 |
## Bias, Risks and Limitations
|
|
@@ -513,9 +513,9 @@ You can finetune this model on your own dataset.
|
|
| 513 |
- `eval_strategy`: steps
|
| 514 |
- `per_device_train_batch_size`: 128
|
| 515 |
- `per_device_eval_batch_size`: 128
|
| 516 |
-
- `learning_rate`:
|
| 517 |
-
- `weight_decay`: 0.
|
| 518 |
-
- `max_steps`:
|
| 519 |
- `warmup_ratio`: 0.1
|
| 520 |
- `fp16`: True
|
| 521 |
- `dataloader_drop_last`: True
|
|
@@ -542,14 +542,14 @@ You can finetune this model on your own dataset.
|
|
| 542 |
- `gradient_accumulation_steps`: 1
|
| 543 |
- `eval_accumulation_steps`: None
|
| 544 |
- `torch_empty_cache_steps`: None
|
| 545 |
-
- `learning_rate`:
|
| 546 |
-
- `weight_decay`: 0.
|
| 547 |
- `adam_beta1`: 0.9
|
| 548 |
- `adam_beta2`: 0.999
|
| 549 |
- `adam_epsilon`: 1e-08
|
| 550 |
- `max_grad_norm`: 1.0
|
| 551 |
- `num_train_epochs`: 3.0
|
| 552 |
-
- `max_steps`:
|
| 553 |
- `lr_scheduler_type`: linear
|
| 554 |
- `lr_scheduler_kwargs`: {}
|
| 555 |
- `warmup_ratio`: 0.1
|
|
@@ -656,13 +656,11 @@ You can finetune this model on your own dataset.
|
|
| 656 |
### Training Logs
|
| 657 |
| Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
|
| 658 |
|:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
|
| 659 |
-
| 0 | 0 | - | 0.
|
| 660 |
-
| 0.2874 | 250 | 0.
|
| 661 |
-
| 0.5747 | 500 | 0.
|
| 662 |
-
| 0.8621 | 750 | 0.
|
| 663 |
-
| 1.1494 | 1000 | 0.
|
| 664 |
-
| 1.4368 | 1250 | 0.0422 | 0.0537 | 0.5300 | 0.6107 | 0.5704 |
|
| 665 |
-
| 1.7241 | 1500 | 0.0402 | 0.0514 | 0.5174 | 0.6172 | 0.5673 |
|
| 666 |
|
| 667 |
|
| 668 |
### Framework Versions
|
|
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:111470
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
+
base_model: sentence-transformers/all-MiniLM-L12-v2
|
| 11 |
widget:
|
| 12 |
- source_sentence: when was the first elephant brought to america
|
| 13 |
sentences:
|
|
|
|
| 132 |
- cosine_mrr@10
|
| 133 |
- cosine_map@100
|
| 134 |
model-index:
|
| 135 |
+
- name: SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
|
| 136 |
results:
|
| 137 |
- task:
|
| 138 |
type: information-retrieval
|
|
|
|
| 142 |
type: NanoMSMARCO
|
| 143 |
metrics:
|
| 144 |
- type: cosine_accuracy@1
|
| 145 |
+
value: 0.34
|
| 146 |
name: Cosine Accuracy@1
|
| 147 |
- type: cosine_accuracy@3
|
| 148 |
value: 0.5
|
| 149 |
name: Cosine Accuracy@3
|
| 150 |
- type: cosine_accuracy@5
|
| 151 |
+
value: 0.66
|
| 152 |
name: Cosine Accuracy@5
|
| 153 |
- type: cosine_accuracy@10
|
| 154 |
+
value: 0.78
|
| 155 |
name: Cosine Accuracy@10
|
| 156 |
- type: cosine_precision@1
|
| 157 |
+
value: 0.34
|
| 158 |
name: Cosine Precision@1
|
| 159 |
- type: cosine_precision@3
|
| 160 |
value: 0.16666666666666663
|
| 161 |
name: Cosine Precision@3
|
| 162 |
- type: cosine_precision@5
|
| 163 |
+
value: 0.132
|
| 164 |
name: Cosine Precision@5
|
| 165 |
- type: cosine_precision@10
|
| 166 |
+
value: 0.078
|
| 167 |
name: Cosine Precision@10
|
| 168 |
- type: cosine_recall@1
|
| 169 |
+
value: 0.34
|
| 170 |
name: Cosine Recall@1
|
| 171 |
- type: cosine_recall@3
|
| 172 |
value: 0.5
|
| 173 |
name: Cosine Recall@3
|
| 174 |
- type: cosine_recall@5
|
| 175 |
+
value: 0.66
|
| 176 |
name: Cosine Recall@5
|
| 177 |
- type: cosine_recall@10
|
| 178 |
+
value: 0.78
|
| 179 |
name: Cosine Recall@10
|
| 180 |
- type: cosine_ndcg@10
|
| 181 |
+
value: 0.5446770528863051
|
| 182 |
name: Cosine Ndcg@10
|
| 183 |
- type: cosine_mrr@10
|
| 184 |
+
value: 0.4708571428571428
|
| 185 |
name: Cosine Mrr@10
|
| 186 |
- type: cosine_map@100
|
| 187 |
+
value: 0.47884258431632043
|
| 188 |
name: Cosine Map@100
|
| 189 |
- task:
|
| 190 |
type: information-retrieval
|
|
|
|
| 194 |
type: NanoNQ
|
| 195 |
metrics:
|
| 196 |
- type: cosine_accuracy@1
|
| 197 |
+
value: 0.5
|
| 198 |
name: Cosine Accuracy@1
|
| 199 |
- type: cosine_accuracy@3
|
| 200 |
value: 0.66
|
| 201 |
name: Cosine Accuracy@3
|
| 202 |
- type: cosine_accuracy@5
|
| 203 |
+
value: 0.7
|
| 204 |
name: Cosine Accuracy@5
|
| 205 |
- type: cosine_accuracy@10
|
| 206 |
+
value: 0.78
|
| 207 |
name: Cosine Accuracy@10
|
| 208 |
- type: cosine_precision@1
|
| 209 |
+
value: 0.5
|
| 210 |
name: Cosine Precision@1
|
| 211 |
- type: cosine_precision@3
|
| 212 |
+
value: 0.22666666666666668
|
| 213 |
name: Cosine Precision@3
|
| 214 |
- type: cosine_precision@5
|
| 215 |
+
value: 0.14400000000000002
|
| 216 |
name: Cosine Precision@5
|
| 217 |
- type: cosine_precision@10
|
| 218 |
+
value: 0.08199999999999999
|
| 219 |
name: Cosine Precision@10
|
| 220 |
- type: cosine_recall@1
|
| 221 |
+
value: 0.48
|
| 222 |
name: Cosine Recall@1
|
| 223 |
- type: cosine_recall@3
|
| 224 |
+
value: 0.64
|
| 225 |
name: Cosine Recall@3
|
| 226 |
- type: cosine_recall@5
|
| 227 |
+
value: 0.67
|
| 228 |
name: Cosine Recall@5
|
| 229 |
- type: cosine_recall@10
|
| 230 |
+
value: 0.74
|
| 231 |
name: Cosine Recall@10
|
| 232 |
- type: cosine_ndcg@10
|
| 233 |
+
value: 0.6136402968638738
|
| 234 |
name: Cosine Ndcg@10
|
| 235 |
- type: cosine_mrr@10
|
| 236 |
+
value: 0.5821666666666667
|
| 237 |
name: Cosine Mrr@10
|
| 238 |
- type: cosine_map@100
|
| 239 |
+
value: 0.5768526974820034
|
| 240 |
name: Cosine Map@100
|
| 241 |
- task:
|
| 242 |
type: nano-beir
|
|
|
|
| 246 |
type: NanoBEIR_mean
|
| 247 |
metrics:
|
| 248 |
- type: cosine_accuracy@1
|
| 249 |
+
value: 0.42000000000000004
|
| 250 |
name: Cosine Accuracy@1
|
| 251 |
- type: cosine_accuracy@3
|
| 252 |
value: 0.5800000000000001
|
| 253 |
name: Cosine Accuracy@3
|
| 254 |
- type: cosine_accuracy@5
|
| 255 |
+
value: 0.6799999999999999
|
| 256 |
name: Cosine Accuracy@5
|
| 257 |
- type: cosine_accuracy@10
|
| 258 |
+
value: 0.78
|
| 259 |
name: Cosine Accuracy@10
|
| 260 |
- type: cosine_precision@1
|
| 261 |
+
value: 0.42000000000000004
|
| 262 |
name: Cosine Precision@1
|
| 263 |
- type: cosine_precision@3
|
| 264 |
+
value: 0.19666666666666666
|
| 265 |
name: Cosine Precision@3
|
| 266 |
- type: cosine_precision@5
|
| 267 |
+
value: 0.138
|
| 268 |
name: Cosine Precision@5
|
| 269 |
- type: cosine_precision@10
|
| 270 |
+
value: 0.07999999999999999
|
| 271 |
name: Cosine Precision@10
|
| 272 |
- type: cosine_recall@1
|
| 273 |
+
value: 0.41000000000000003
|
| 274 |
name: Cosine Recall@1
|
| 275 |
- type: cosine_recall@3
|
| 276 |
+
value: 0.5700000000000001
|
| 277 |
name: Cosine Recall@3
|
| 278 |
- type: cosine_recall@5
|
| 279 |
+
value: 0.665
|
| 280 |
name: Cosine Recall@5
|
| 281 |
- type: cosine_recall@10
|
| 282 |
+
value: 0.76
|
| 283 |
name: Cosine Recall@10
|
| 284 |
- type: cosine_ndcg@10
|
| 285 |
+
value: 0.5791586748750894
|
| 286 |
name: Cosine Ndcg@10
|
| 287 |
- type: cosine_mrr@10
|
| 288 |
+
value: 0.5265119047619048
|
| 289 |
name: Cosine Mrr@10
|
| 290 |
- type: cosine_map@100
|
| 291 |
+
value: 0.5278476408991619
|
| 292 |
name: Cosine Map@100
|
| 293 |
---
|
| 294 |
|
| 295 |
+
# SentenceTransformer based on sentence-transformers/all-MiniLM-L12-v2
|
| 296 |
|
| 297 |
+
This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 298 |
|
| 299 |
## Model Details
|
| 300 |
|
| 301 |
### Model Description
|
| 302 |
- **Model Type:** Sentence Transformer
|
| 303 |
+
- **Base model:** [sentence-transformers/all-MiniLM-L12-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2) <!-- at revision 936af83a2ecce5fe87a09109ff5cbcefe073173a -->
|
| 304 |
- **Maximum Sequence Length:** 128 tokens
|
| 305 |
- **Output Dimensionality:** 384 dimensions
|
| 306 |
- **Similarity Function:** Cosine Similarity
|
|
|
|
| 353 |
# Get the similarity scores for the embeddings
|
| 354 |
similarities = model.similarity(embeddings, embeddings)
|
| 355 |
print(similarities)
|
| 356 |
+
# tensor([[1.0000, 1.0000, 0.8845],
|
| 357 |
+
# [1.0000, 1.0000, 0.8845],
|
| 358 |
+
# [0.8845, 0.8845, 1.0000]])
|
| 359 |
```
|
| 360 |
|
| 361 |
<!--
|
|
|
|
| 393 |
|
| 394 |
| Metric | NanoMSMARCO | NanoNQ |
|
| 395 |
|:--------------------|:------------|:-----------|
|
| 396 |
+
| cosine_accuracy@1 | 0.34 | 0.5 |
|
| 397 |
| cosine_accuracy@3 | 0.5 | 0.66 |
|
| 398 |
+
| cosine_accuracy@5 | 0.66 | 0.7 |
|
| 399 |
+
| cosine_accuracy@10 | 0.78 | 0.78 |
|
| 400 |
+
| cosine_precision@1 | 0.34 | 0.5 |
|
| 401 |
+
| cosine_precision@3 | 0.1667 | 0.2267 |
|
| 402 |
+
| cosine_precision@5 | 0.132 | 0.144 |
|
| 403 |
+
| cosine_precision@10 | 0.078 | 0.082 |
|
| 404 |
+
| cosine_recall@1 | 0.34 | 0.48 |
|
| 405 |
+
| cosine_recall@3 | 0.5 | 0.64 |
|
| 406 |
+
| cosine_recall@5 | 0.66 | 0.67 |
|
| 407 |
+
| cosine_recall@10 | 0.78 | 0.74 |
|
| 408 |
+
| **cosine_ndcg@10** | **0.5447** | **0.6136** |
|
| 409 |
+
| cosine_mrr@10 | 0.4709 | 0.5822 |
|
| 410 |
+
| cosine_map@100 | 0.4788 | 0.5769 |
|
| 411 |
|
| 412 |
#### Nano BEIR
|
| 413 |
|
|
|
|
| 425 |
|
| 426 |
| Metric | Value |
|
| 427 |
|:--------------------|:-----------|
|
| 428 |
+
| cosine_accuracy@1 | 0.42 |
|
| 429 |
| cosine_accuracy@3 | 0.58 |
|
| 430 |
+
| cosine_accuracy@5 | 0.68 |
|
| 431 |
+
| cosine_accuracy@10 | 0.78 |
|
| 432 |
+
| cosine_precision@1 | 0.42 |
|
| 433 |
+
| cosine_precision@3 | 0.1967 |
|
| 434 |
+
| cosine_precision@5 | 0.138 |
|
| 435 |
+
| cosine_precision@10 | 0.08 |
|
| 436 |
+
| cosine_recall@1 | 0.41 |
|
| 437 |
+
| cosine_recall@3 | 0.57 |
|
| 438 |
+
| cosine_recall@5 | 0.665 |
|
| 439 |
+
| cosine_recall@10 | 0.76 |
|
| 440 |
+
| **cosine_ndcg@10** | **0.5792** |
|
| 441 |
+
| cosine_mrr@10 | 0.5265 |
|
| 442 |
+
| cosine_map@100 | 0.5278 |
|
| 443 |
|
| 444 |
<!--
|
| 445 |
## Bias, Risks and Limitations
|
|
|
|
| 513 |
- `eval_strategy`: steps
|
| 514 |
- `per_device_train_batch_size`: 128
|
| 515 |
- `per_device_eval_batch_size`: 128
|
| 516 |
+
- `learning_rate`: 8e-05
|
| 517 |
+
- `weight_decay`: 0.005
|
| 518 |
+
- `max_steps`: 1125
|
| 519 |
- `warmup_ratio`: 0.1
|
| 520 |
- `fp16`: True
|
| 521 |
- `dataloader_drop_last`: True
|
|
|
|
| 542 |
- `gradient_accumulation_steps`: 1
|
| 543 |
- `eval_accumulation_steps`: None
|
| 544 |
- `torch_empty_cache_steps`: None
|
| 545 |
+
- `learning_rate`: 8e-05
|
| 546 |
+
- `weight_decay`: 0.005
|
| 547 |
- `adam_beta1`: 0.9
|
| 548 |
- `adam_beta2`: 0.999
|
| 549 |
- `adam_epsilon`: 1e-08
|
| 550 |
- `max_grad_norm`: 1.0
|
| 551 |
- `num_train_epochs`: 3.0
|
| 552 |
+
- `max_steps`: 1125
|
| 553 |
- `lr_scheduler_type`: linear
|
| 554 |
- `lr_scheduler_kwargs`: {}
|
| 555 |
- `warmup_ratio`: 0.1
|
|
|
|
| 656 |
### Training Logs
|
| 657 |
| Epoch | Step | Training Loss | Validation Loss | NanoMSMARCO_cosine_ndcg@10 | NanoNQ_cosine_ndcg@10 | NanoBEIR_mean_cosine_ndcg@10 |
|
| 658 |
|:------:|:----:|:-------------:|:---------------:|:--------------------------:|:---------------------:|:----------------------------:|
|
| 659 |
+
| 0 | 0 | - | 0.1203 | 0.5887 | 0.5786 | 0.5836 |
|
| 660 |
+
| 0.2874 | 250 | 0.094 | 0.0631 | 0.5536 | 0.5611 | 0.5574 |
|
| 661 |
+
| 0.5747 | 500 | 0.0766 | 0.0586 | 0.5317 | 0.5724 | 0.5521 |
|
| 662 |
+
| 0.8621 | 750 | 0.0674 | 0.0494 | 0.5357 | 0.5675 | 0.5516 |
|
| 663 |
+
| 1.1494 | 1000 | 0.0491 | 0.0468 | 0.5447 | 0.6136 | 0.5792 |
|
|
|
|
|
|
|
| 664 |
|
| 665 |
|
| 666 |
### Framework Versions
|