Updating model weights
Browse files
README.md
CHANGED
|
@@ -7,7 +7,6 @@ tags:
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:291522
|
| 9 |
- loss:MultipleNegativesSymmetricRankingLoss
|
| 10 |
-
base_model: sentence-transformers/all-MiniLM-L6-v2
|
| 11 |
widget:
|
| 12 |
- source_sentence: cream 21 baby oil with almond oil
|
| 13 |
sentences:
|
|
@@ -41,7 +40,7 @@ library_name: sentence-transformers
|
|
| 41 |
metrics:
|
| 42 |
- cosine_accuracy
|
| 43 |
model-index:
|
| 44 |
-
- name: SentenceTransformer
|
| 45 |
results:
|
| 46 |
- task:
|
| 47 |
type: triplet
|
|
@@ -51,19 +50,19 @@ model-index:
|
|
| 51 |
type: unknown
|
| 52 |
metrics:
|
| 53 |
- type: cosine_accuracy
|
| 54 |
-
value: 0.
|
| 55 |
name: Cosine Accuracy
|
| 56 |
---
|
| 57 |
|
| 58 |
-
# SentenceTransformer
|
| 59 |
|
| 60 |
-
This is a [sentence-transformers](https://www.SBERT.net) model
|
| 61 |
|
| 62 |
## Model Details
|
| 63 |
|
| 64 |
### Model Description
|
| 65 |
- **Model Type:** Sentence Transformer
|
| 66 |
-
- **Base model:** [
|
| 67 |
- **Maximum Sequence Length:** 256 tokens
|
| 68 |
- **Output Dimensionality:** 384 dimensions
|
| 69 |
- **Similarity Function:** Cosine Similarity
|
|
@@ -116,9 +115,9 @@ print(embeddings.shape)
|
|
| 116 |
# Get the similarity scores for the embeddings
|
| 117 |
similarities = model.similarity(embeddings, embeddings)
|
| 118 |
print(similarities)
|
| 119 |
-
# tensor([[1.0000, 0.
|
| 120 |
-
# [0.
|
| 121 |
-
# [0.
|
| 122 |
```
|
| 123 |
|
| 124 |
<!--
|
|
@@ -155,7 +154,7 @@ You can finetune this model on your own dataset.
|
|
| 155 |
|
| 156 |
| Metric | Value |
|
| 157 |
|:--------------------|:-----------|
|
| 158 |
-
| **cosine_accuracy** | **0.
|
| 159 |
|
| 160 |
<!--
|
| 161 |
## Bias, Risks and Limitations
|
|
@@ -230,6 +229,7 @@ You can finetune this model on your own dataset.
|
|
| 230 |
- `per_device_train_batch_size`: 256
|
| 231 |
- `per_device_eval_batch_size`: 256
|
| 232 |
- `weight_decay`: 0.001
|
|
|
|
| 233 |
- `warmup_steps`: 1138
|
| 234 |
- `fp16`: True
|
| 235 |
- `dataloader_num_workers`: 4
|
|
@@ -260,7 +260,7 @@ You can finetune this model on your own dataset.
|
|
| 260 |
- `adam_beta2`: 0.999
|
| 261 |
- `adam_epsilon`: 1e-08
|
| 262 |
- `max_grad_norm`: 1.0
|
| 263 |
-
- `num_train_epochs`:
|
| 264 |
- `max_steps`: -1
|
| 265 |
- `lr_scheduler_type`: linear
|
| 266 |
- `lr_scheduler_kwargs`: {}
|
|
@@ -364,13 +364,10 @@ You can finetune this model on your own dataset.
|
|
| 364 |
</details>
|
| 365 |
|
| 366 |
### Training Logs
|
| 367 |
-
| Epoch
|
| 368 |
-
|
| 369 |
-
|
|
| 370 |
-
| 0
|
| 371 |
-
| 1.0 | 1139 | 3.0136 | 0.8482 | 0.9113 |
|
| 372 |
-
| 2.0 | 2278 | 2.2096 | 0.7465 | 0.9241 |
|
| 373 |
-
| 3.0 | 3417 | 1.966 | 0.6980 | 0.9337 |
|
| 374 |
|
| 375 |
|
| 376 |
### Framework Versions
|
|
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:291522
|
| 9 |
- loss:MultipleNegativesSymmetricRankingLoss
|
|
|
|
| 10 |
widget:
|
| 11 |
- source_sentence: cream 21 baby oil with almond oil
|
| 12 |
sentences:
|
|
|
|
| 40 |
metrics:
|
| 41 |
- cosine_accuracy
|
| 42 |
model-index:
|
| 43 |
+
- name: SentenceTransformer
|
| 44 |
results:
|
| 45 |
- task:
|
| 46 |
type: triplet
|
|
|
|
| 50 |
type: unknown
|
| 51 |
metrics:
|
| 52 |
- type: cosine_accuracy
|
| 53 |
+
value: 0.9403471946716309
|
| 54 |
name: Cosine Accuracy
|
| 55 |
---
|
| 56 |
|
| 57 |
+
# SentenceTransformer
|
| 58 |
|
| 59 |
+
This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 60 |
|
| 61 |
## Model Details
|
| 62 |
|
| 63 |
### Model Description
|
| 64 |
- **Model Type:** Sentence Transformer
|
| 65 |
+
<!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
|
| 66 |
- **Maximum Sequence Length:** 256 tokens
|
| 67 |
- **Output Dimensionality:** 384 dimensions
|
| 68 |
- **Similarity Function:** Cosine Similarity
|
|
|
|
| 115 |
# Get the similarity scores for the embeddings
|
| 116 |
similarities = model.similarity(embeddings, embeddings)
|
| 117 |
print(similarities)
|
| 118 |
+
# tensor([[1.0000, 0.7730, 0.3475],
|
| 119 |
+
# [0.7730, 1.0000, 0.3615],
|
| 120 |
+
# [0.3475, 0.3615, 1.0000]])
|
| 121 |
```
|
| 122 |
|
| 123 |
<!--
|
|
|
|
| 154 |
|
| 155 |
| Metric | Value |
|
| 156 |
|:--------------------|:-----------|
|
| 157 |
+
| **cosine_accuracy** | **0.9403** |
|
| 158 |
|
| 159 |
<!--
|
| 160 |
## Bias, Risks and Limitations
|
|
|
|
| 229 |
- `per_device_train_batch_size`: 256
|
| 230 |
- `per_device_eval_batch_size`: 256
|
| 231 |
- `weight_decay`: 0.001
|
| 232 |
+
- `num_train_epochs`: 5
|
| 233 |
- `warmup_steps`: 1138
|
| 234 |
- `fp16`: True
|
| 235 |
- `dataloader_num_workers`: 4
|
|
|
|
| 260 |
- `adam_beta2`: 0.999
|
| 261 |
- `adam_epsilon`: 1e-08
|
| 262 |
- `max_grad_norm`: 1.0
|
| 263 |
+
- `num_train_epochs`: 5
|
| 264 |
- `max_steps`: -1
|
| 265 |
- `lr_scheduler_type`: linear
|
| 266 |
- `lr_scheduler_kwargs`: {}
|
|
|
|
| 364 |
</details>
|
| 365 |
|
| 366 |
### Training Logs
|
| 367 |
+
| Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
|
| 368 |
+
|:-----:|:----:|:-------------:|:---------------:|:---------------:|
|
| 369 |
+
| 4.0 | 4556 | 1.8731 | 0.7003 | 0.9331 |
|
| 370 |
+
| 5.0 | 5695 | 1.7998 | 0.6516 | 0.9403 |
|
|
|
|
|
|
|
|
|
|
| 371 |
|
| 372 |
|
| 373 |
### Framework Versions
|