Add new SentenceTransformer model
Browse files- README.md +44 -7
- model.safetensors +1 -1
README.md
CHANGED
|
@@ -462,6 +462,21 @@ widget:
|
|
| 462 |
JUAN MARÍA ABURTO RIQUE.'
|
| 463 |
pipeline_tag: sentence-similarity
|
| 464 |
library_name: sentence-transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 465 |
---
|
| 466 |
|
| 467 |
# SentenceTransformer based on intfloat/multilingual-e5-large
|
|
@@ -525,9 +540,9 @@ print(embeddings.shape)
|
|
| 525 |
# Get the similarity scores for the embeddings
|
| 526 |
similarities = model.similarity(embeddings, embeddings)
|
| 527 |
print(similarities)
|
| 528 |
-
# tensor([[1.0000,
|
| 529 |
-
# [
|
| 530 |
-
# [
|
| 531 |
```
|
| 532 |
|
| 533 |
<!--
|
|
@@ -554,6 +569,19 @@ You can finetune this model on your own dataset.
|
|
| 554 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 555 |
-->
|
| 556 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 557 |
<!--
|
| 558 |
## Bias, Risks and Limitations
|
| 559 |
|
|
@@ -622,8 +650,6 @@ You can finetune this model on your own dataset.
|
|
| 622 |
#### Non-Default Hyperparameters
|
| 623 |
|
| 624 |
- `eval_strategy`: steps
|
| 625 |
-
- `per_device_train_batch_size`: 16
|
| 626 |
-
- `per_device_eval_batch_size`: 16
|
| 627 |
- `warmup_ratio`: 0.1
|
| 628 |
- `fp16`: True
|
| 629 |
- `batch_sampler`: no_duplicates
|
|
@@ -635,8 +661,8 @@ You can finetune this model on your own dataset.
|
|
| 635 |
- `do_predict`: False
|
| 636 |
- `eval_strategy`: steps
|
| 637 |
- `prediction_loss_only`: True
|
| 638 |
-
- `per_device_train_batch_size`:
|
| 639 |
-
- `per_device_eval_batch_size`:
|
| 640 |
- `per_gpu_train_batch_size`: None
|
| 641 |
- `per_gpu_eval_batch_size`: None
|
| 642 |
- `gradient_accumulation_steps`: 1
|
|
@@ -749,6 +775,17 @@ You can finetune this model on your own dataset.
|
|
| 749 |
|
| 750 |
</details>
|
| 751 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 752 |
### Framework Versions
|
| 753 |
- Python: 3.9.7
|
| 754 |
- Sentence Transformers: 5.0.0
|
|
|
|
| 462 |
JUAN MARÍA ABURTO RIQUE.'
|
| 463 |
pipeline_tag: sentence-similarity
|
| 464 |
library_name: sentence-transformers
|
| 465 |
+
metrics:
|
| 466 |
+
- cosine_accuracy
|
| 467 |
+
model-index:
|
| 468 |
+
- name: SentenceTransformer based on intfloat/multilingual-e5-large
|
| 469 |
+
results:
|
| 470 |
+
- task:
|
| 471 |
+
type: triplet
|
| 472 |
+
name: Triplet
|
| 473 |
+
dataset:
|
| 474 |
+
name: multilingual e5 large
|
| 475 |
+
type: multilingual-e5-large
|
| 476 |
+
metrics:
|
| 477 |
+
- type: cosine_accuracy
|
| 478 |
+
value: 0.2175000011920929
|
| 479 |
+
name: Cosine Accuracy
|
| 480 |
---
|
| 481 |
|
| 482 |
# SentenceTransformer based on intfloat/multilingual-e5-large
|
|
|
|
| 540 |
# Get the similarity scores for the embeddings
|
| 541 |
similarities = model.similarity(embeddings, embeddings)
|
| 542 |
print(similarities)
|
| 543 |
+
# tensor([[ 1.0000, -0.9967, -0.9967],
|
| 544 |
+
# [-0.9967, 1.0000, 1.0000],
|
| 545 |
+
# [-0.9967, 1.0000, 1.0000]])
|
| 546 |
```
|
| 547 |
|
| 548 |
<!--
|
|
|
|
| 569 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 570 |
-->
|
| 571 |
|
| 572 |
+
## Evaluation
|
| 573 |
+
|
| 574 |
+
### Metrics
|
| 575 |
+
|
| 576 |
+
#### Triplet
|
| 577 |
+
|
| 578 |
+
* Dataset: `multilingual-e5-large`
|
| 579 |
+
* Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator)
|
| 580 |
+
|
| 581 |
+
| Metric | Value |
|
| 582 |
+
|:--------------------|:-----------|
|
| 583 |
+
| **cosine_accuracy** | **0.2175** |
|
| 584 |
+
|
| 585 |
<!--
|
| 586 |
## Bias, Risks and Limitations
|
| 587 |
|
|
|
|
| 650 |
#### Non-Default Hyperparameters
|
| 651 |
|
| 652 |
- `eval_strategy`: steps
|
|
|
|
|
|
|
| 653 |
- `warmup_ratio`: 0.1
|
| 654 |
- `fp16`: True
|
| 655 |
- `batch_sampler`: no_duplicates
|
|
|
|
| 661 |
- `do_predict`: False
|
| 662 |
- `eval_strategy`: steps
|
| 663 |
- `prediction_loss_only`: True
|
| 664 |
+
- `per_device_train_batch_size`: 8
|
| 665 |
+
- `per_device_eval_batch_size`: 8
|
| 666 |
- `per_gpu_train_batch_size`: None
|
| 667 |
- `per_gpu_eval_batch_size`: None
|
| 668 |
- `gradient_accumulation_steps`: 1
|
|
|
|
| 775 |
|
| 776 |
</details>
|
| 777 |
|
| 778 |
+
### Training Logs
|
| 779 |
+
| Epoch | Step | Training Loss | Validation Loss | multilingual-e5-large_cosine_accuracy |
|
| 780 |
+
|:-----:|:----:|:-------------:|:---------------:|:-------------------------------------:|
|
| 781 |
+
| 0.5 | 100 | 4.1212 | 3.8693 | 0.2325 |
|
| 782 |
+
| 1.0 | 200 | 3.8258 | 3.8681 | 0.2225 |
|
| 783 |
+
| 1.5 | 300 | 3.7783 | 3.8678 | 0.2150 |
|
| 784 |
+
| 2.0 | 400 | 3.8344 | 3.8676 | 0.2325 |
|
| 785 |
+
| 2.5 | 500 | 3.7929 | 3.8674 | 0.2175 |
|
| 786 |
+
| 3.0 | 600 | 3.8121 | 3.8674 | 0.2175 |
|
| 787 |
+
|
| 788 |
+
|
| 789 |
### Framework Versions
|
| 790 |
- Python: 3.9.7
|
| 791 |
- Sentence Transformers: 5.0.0
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 2239607176
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:71c616df6c24ba89358947095e881f0cba7430855683767390326414888d35b6
|
| 3 |
size 2239607176
|