Updating model weights
Browse files
README.md
CHANGED
|
@@ -7,7 +7,6 @@ tags:
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:291522
|
| 9 |
- loss:MultipleNegativesSymmetricRankingLoss
|
| 10 |
-
base_model: sentence-transformers/all-MiniLM-L6-v2
|
| 11 |
widget:
|
| 12 |
- source_sentence: cream 21 baby oil with almond oil
|
| 13 |
sentences:
|
|
@@ -38,32 +37,17 @@ widget:
|
|
| 38 |
- kids game
|
| 39 |
pipeline_tag: sentence-similarity
|
| 40 |
library_name: sentence-transformers
|
| 41 |
-
metrics:
|
| 42 |
-
- cosine_accuracy
|
| 43 |
-
model-index:
|
| 44 |
-
- name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
|
| 45 |
-
results:
|
| 46 |
-
- task:
|
| 47 |
-
type: triplet
|
| 48 |
-
name: Triplet
|
| 49 |
-
dataset:
|
| 50 |
-
name: Unknown
|
| 51 |
-
type: unknown
|
| 52 |
-
metrics:
|
| 53 |
-
- type: cosine_accuracy
|
| 54 |
-
value: 0.9412940740585327
|
| 55 |
-
name: Cosine Accuracy
|
| 56 |
---
|
| 57 |
|
| 58 |
-
# SentenceTransformer
|
| 59 |
|
| 60 |
-
This is a [sentence-transformers](https://www.SBERT.net) model
|
| 61 |
|
| 62 |
## Model Details
|
| 63 |
|
| 64 |
### Model Description
|
| 65 |
- **Model Type:** Sentence Transformer
|
| 66 |
-
- **Base model:** [
|
| 67 |
- **Maximum Sequence Length:** 256 tokens
|
| 68 |
- **Output Dimensionality:** 384 dimensions
|
| 69 |
- **Similarity Function:** Cosine Similarity
|
|
@@ -145,18 +129,6 @@ You can finetune this model on your own dataset.
|
|
| 145 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 146 |
-->
|
| 147 |
|
| 148 |
-
## Evaluation
|
| 149 |
-
|
| 150 |
-
### Metrics
|
| 151 |
-
|
| 152 |
-
#### Triplet
|
| 153 |
-
|
| 154 |
-
* Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator)
|
| 155 |
-
|
| 156 |
-
| Metric | Value |
|
| 157 |
-
|:--------------------|:-----------|
|
| 158 |
-
| **cosine_accuracy** | **0.9413** |
|
| 159 |
-
|
| 160 |
<!--
|
| 161 |
## Bias, Risks and Limitations
|
| 162 |
|
|
@@ -226,7 +198,7 @@ You can finetune this model on your own dataset.
|
|
| 226 |
### Training Hyperparameters
|
| 227 |
#### Non-Default Hyperparameters
|
| 228 |
|
| 229 |
-
- `eval_strategy`:
|
| 230 |
- `per_device_train_batch_size`: 128
|
| 231 |
- `per_device_eval_batch_size`: 128
|
| 232 |
- `weight_decay`: 0.001
|
|
@@ -245,7 +217,7 @@ You can finetune this model on your own dataset.
|
|
| 245 |
|
| 246 |
- `overwrite_output_dir`: False
|
| 247 |
- `do_predict`: False
|
| 248 |
-
- `eval_strategy`:
|
| 249 |
- `prediction_loss_only`: True
|
| 250 |
- `per_device_train_batch_size`: 128
|
| 251 |
- `per_device_eval_batch_size`: 128
|
|
@@ -363,13 +335,6 @@ You can finetune this model on your own dataset.
|
|
| 363 |
|
| 364 |
</details>
|
| 365 |
|
| 366 |
-
### Training Logs
|
| 367 |
-
| Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
|
| 368 |
-
|:------:|:----:|:-------------:|:---------------:|:---------------:|
|
| 369 |
-
| 0.0004 | 1 | 5.3655 | - | - |
|
| 370 |
-
| 2.1949 | 5000 | 2.1423 | 0.7694 | 0.9413 |
|
| 371 |
-
|
| 372 |
-
|
| 373 |
### Framework Versions
|
| 374 |
- Python: 3.11.13
|
| 375 |
- Sentence Transformers: 5.1.2
|
|
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:291522
|
| 9 |
- loss:MultipleNegativesSymmetricRankingLoss
|
|
|
|
| 10 |
widget:
|
| 11 |
- source_sentence: cream 21 baby oil with almond oil
|
| 12 |
sentences:
|
|
|
|
| 37 |
- kids game
|
| 38 |
pipeline_tag: sentence-similarity
|
| 39 |
library_name: sentence-transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
---
|
| 41 |
|
| 42 |
+
# SentenceTransformer
|
| 43 |
|
| 44 |
+
This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 45 |
|
| 46 |
## Model Details
|
| 47 |
|
| 48 |
### Model Description
|
| 49 |
- **Model Type:** Sentence Transformer
|
| 50 |
+
<!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
|
| 51 |
- **Maximum Sequence Length:** 256 tokens
|
| 52 |
- **Output Dimensionality:** 384 dimensions
|
| 53 |
- **Similarity Function:** Cosine Similarity
|
|
|
|
| 129 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 130 |
-->
|
| 131 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 132 |
<!--
|
| 133 |
## Bias, Risks and Limitations
|
| 134 |
|
|
|
|
| 198 |
### Training Hyperparameters
|
| 199 |
#### Non-Default Hyperparameters
|
| 200 |
|
| 201 |
+
- `eval_strategy`: epoch
|
| 202 |
- `per_device_train_batch_size`: 128
|
| 203 |
- `per_device_eval_batch_size`: 128
|
| 204 |
- `weight_decay`: 0.001
|
|
|
|
| 217 |
|
| 218 |
- `overwrite_output_dir`: False
|
| 219 |
- `do_predict`: False
|
| 220 |
+
- `eval_strategy`: epoch
|
| 221 |
- `prediction_loss_only`: True
|
| 222 |
- `per_device_train_batch_size`: 128
|
| 223 |
- `per_device_eval_batch_size`: 128
|
|
|
|
| 335 |
|
| 336 |
</details>
|
| 337 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 338 |
### Framework Versions
|
| 339 |
- Python: 3.11.13
|
| 340 |
- Sentence Transformers: 5.1.2
|