Updating model weights
Browse files
README.md
CHANGED
|
@@ -7,7 +7,6 @@ tags:
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:556626
|
| 9 |
- loss:MultipleNegativesSymmetricRankingLoss
|
| 10 |
-
base_model: sentence-transformers/all-MiniLM-L6-v2
|
| 11 |
widget:
|
| 12 |
- source_sentence: dimlaj orchid printed finest durable glass terkish tea set
|
| 13 |
sentences:
|
|
@@ -39,7 +38,7 @@ library_name: sentence-transformers
|
|
| 39 |
metrics:
|
| 40 |
- cosine_accuracy
|
| 41 |
model-index:
|
| 42 |
-
- name: SentenceTransformer
|
| 43 |
results:
|
| 44 |
- task:
|
| 45 |
type: triplet
|
|
@@ -49,19 +48,19 @@ model-index:
|
|
| 49 |
type: unknown
|
| 50 |
metrics:
|
| 51 |
- type: cosine_accuracy
|
| 52 |
-
value: 0.
|
| 53 |
name: Cosine Accuracy
|
| 54 |
---
|
| 55 |
|
| 56 |
-
# SentenceTransformer
|
| 57 |
|
| 58 |
-
This is a [sentence-transformers](https://www.SBERT.net) model
|
| 59 |
|
| 60 |
## Model Details
|
| 61 |
|
| 62 |
### Model Description
|
| 63 |
- **Model Type:** Sentence Transformer
|
| 64 |
-
- **Base model:** [
|
| 65 |
- **Maximum Sequence Length:** 256 tokens
|
| 66 |
- **Output Dimensionality:** 384 dimensions
|
| 67 |
- **Similarity Function:** Cosine Similarity
|
|
@@ -114,9 +113,9 @@ print(embeddings.shape)
|
|
| 114 |
# Get the similarity scores for the embeddings
|
| 115 |
similarities = model.similarity(embeddings, embeddings)
|
| 116 |
print(similarities)
|
| 117 |
-
# tensor([[1.0000, 0.
|
| 118 |
-
# [0.
|
| 119 |
-
# [0.
|
| 120 |
```
|
| 121 |
|
| 122 |
<!--
|
|
@@ -153,7 +152,7 @@ You can finetune this model on your own dataset.
|
|
| 153 |
|
| 154 |
| Metric | Value |
|
| 155 |
|:--------------------|:-----------|
|
| 156 |
-
| **cosine_accuracy** | **0.
|
| 157 |
|
| 158 |
<!--
|
| 159 |
## Bias, Risks and Limitations
|
|
@@ -228,6 +227,7 @@ You can finetune this model on your own dataset.
|
|
| 228 |
- `per_device_train_batch_size`: 128
|
| 229 |
- `per_device_eval_batch_size`: 128
|
| 230 |
- `weight_decay`: 0.001
|
|
|
|
| 231 |
- `warmup_steps`: 6956
|
| 232 |
- `fp16`: True
|
| 233 |
- `dataloader_num_workers`: 2
|
|
@@ -258,7 +258,7 @@ You can finetune this model on your own dataset.
|
|
| 258 |
- `adam_beta2`: 0.999
|
| 259 |
- `adam_epsilon`: 1e-08
|
| 260 |
- `max_grad_norm`: 1.0
|
| 261 |
-
- `num_train_epochs`:
|
| 262 |
- `max_steps`: -1
|
| 263 |
- `lr_scheduler_type`: linear
|
| 264 |
- `lr_scheduler_kwargs`: {}
|
|
@@ -362,12 +362,11 @@ You can finetune this model on your own dataset.
|
|
| 362 |
</details>
|
| 363 |
|
| 364 |
### Training Logs
|
| 365 |
-
| Epoch
|
| 366 |
-
|
| 367 |
-
| 0
|
| 368 |
-
|
|
| 369 |
-
|
|
| 370 |
-
| 3.0 | 13047 | 1.4655 | 1.3219 | 0.9608 |
|
| 371 |
|
| 372 |
|
| 373 |
### Framework Versions
|
|
|
|
| 7 |
- generated_from_trainer
|
| 8 |
- dataset_size:556626
|
| 9 |
- loss:MultipleNegativesSymmetricRankingLoss
|
|
|
|
| 10 |
widget:
|
| 11 |
- source_sentence: dimlaj orchid printed finest durable glass terkish tea set
|
| 12 |
sentences:
|
|
|
|
| 38 |
metrics:
|
| 39 |
- cosine_accuracy
|
| 40 |
model-index:
|
| 41 |
+
- name: SentenceTransformer
|
| 42 |
results:
|
| 43 |
- task:
|
| 44 |
type: triplet
|
|
|
|
| 48 |
type: unknown
|
| 49 |
metrics:
|
| 50 |
- type: cosine_accuracy
|
| 51 |
+
value: 0.9618095755577087
|
| 52 |
name: Cosine Accuracy
|
| 53 |
---
|
| 54 |
|
| 55 |
+
# SentenceTransformer
|
| 56 |
|
| 57 |
+
This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
|
| 58 |
|
| 59 |
## Model Details
|
| 60 |
|
| 61 |
### Model Description
|
| 62 |
- **Model Type:** Sentence Transformer
|
| 63 |
+
<!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
|
| 64 |
- **Maximum Sequence Length:** 256 tokens
|
| 65 |
- **Output Dimensionality:** 384 dimensions
|
| 66 |
- **Similarity Function:** Cosine Similarity
|
|
|
|
| 113 |
# Get the similarity scores for the embeddings
|
| 114 |
similarities = model.similarity(embeddings, embeddings)
|
| 115 |
print(similarities)
|
| 116 |
+
# tensor([[1.0000, 0.4517, 0.3474],
|
| 117 |
+
# [0.4517, 1.0000, 0.3222],
|
| 118 |
+
# [0.3474, 0.3222, 1.0000]])
|
| 119 |
```
|
| 120 |
|
| 121 |
<!--
|
|
|
|
| 152 |
|
| 153 |
| Metric | Value |
|
| 154 |
|:--------------------|:-----------|
|
| 155 |
+
| **cosine_accuracy** | **0.9618** |
|
| 156 |
|
| 157 |
<!--
|
| 158 |
## Bias, Risks and Limitations
|
|
|
|
| 227 |
- `per_device_train_batch_size`: 128
|
| 228 |
- `per_device_eval_batch_size`: 128
|
| 229 |
- `weight_decay`: 0.001
|
| 230 |
+
- `num_train_epochs`: 6
|
| 231 |
- `warmup_steps`: 6956
|
| 232 |
- `fp16`: True
|
| 233 |
- `dataloader_num_workers`: 2
|
|
|
|
| 258 |
- `adam_beta2`: 0.999
|
| 259 |
- `adam_epsilon`: 1e-08
|
| 260 |
- `max_grad_norm`: 1.0
|
| 261 |
+
- `num_train_epochs`: 6
|
| 262 |
- `max_steps`: -1
|
| 263 |
- `lr_scheduler_type`: linear
|
| 264 |
- `lr_scheduler_kwargs`: {}
|
|
|
|
| 362 |
</details>
|
| 363 |
|
| 364 |
### Training Logs
|
| 365 |
+
| Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
|
| 366 |
+
|:-----:|:-----:|:-------------:|:---------------:|:---------------:|
|
| 367 |
+
| 4.0 | 17396 | 1.3564 | 1.3029 | 0.9600 |
|
| 368 |
+
| 5.0 | 21745 | 1.2501 | 1.3017 | 0.9622 |
|
| 369 |
+
| 6.0 | 26094 | 1.1858 | 1.2925 | 0.9618 |
|
|
|
|
| 370 |
|
| 371 |
|
| 372 |
### Framework Versions
|