LamaDiab commited on
Commit
53e5948
·
verified ·
1 Parent(s): fe594d0

Updating model weights

Browse files
Files changed (1) hide show
  1. README.md +15 -18
README.md CHANGED
@@ -7,7 +7,6 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:291522
9
  - loss:MultipleNegativesSymmetricRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
  - source_sentence: cream 21 baby oil with almond oil
13
  sentences:
@@ -41,7 +40,7 @@ library_name: sentence-transformers
41
  metrics:
42
  - cosine_accuracy
43
  model-index:
44
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
45
  results:
46
  - task:
47
  type: triplet
@@ -51,19 +50,19 @@ model-index:
51
  type: unknown
52
  metrics:
53
  - type: cosine_accuracy
54
- value: 0.9337190985679626
55
  name: Cosine Accuracy
56
  ---
57
 
58
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
59
 
60
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
61
 
62
  ## Model Details
63
 
64
  ### Model Description
65
  - **Model Type:** Sentence Transformer
66
- - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
67
  - **Maximum Sequence Length:** 256 tokens
68
  - **Output Dimensionality:** 384 dimensions
69
  - **Similarity Function:** Cosine Similarity
@@ -116,9 +115,9 @@ print(embeddings.shape)
116
  # Get the similarity scores for the embeddings
117
  similarities = model.similarity(embeddings, embeddings)
118
  print(similarities)
119
- # tensor([[1.0000, 0.7955, 0.3891],
120
- # [0.7955, 1.0000, 0.4118],
121
- # [0.3891, 0.4118, 1.0000]])
122
  ```
123
 
124
  <!--
@@ -155,7 +154,7 @@ You can finetune this model on your own dataset.
155
 
156
  | Metric | Value |
157
  |:--------------------|:-----------|
158
- | **cosine_accuracy** | **0.9337** |
159
 
160
  <!--
161
  ## Bias, Risks and Limitations
@@ -230,6 +229,7 @@ You can finetune this model on your own dataset.
230
  - `per_device_train_batch_size`: 256
231
  - `per_device_eval_batch_size`: 256
232
  - `weight_decay`: 0.001
 
233
  - `warmup_steps`: 1138
234
  - `fp16`: True
235
  - `dataloader_num_workers`: 4
@@ -260,7 +260,7 @@ You can finetune this model on your own dataset.
260
  - `adam_beta2`: 0.999
261
  - `adam_epsilon`: 1e-08
262
  - `max_grad_norm`: 1.0
263
- - `num_train_epochs`: 3
264
  - `max_steps`: -1
265
  - `lr_scheduler_type`: linear
266
  - `lr_scheduler_kwargs`: {}
@@ -364,13 +364,10 @@ You can finetune this model on your own dataset.
364
  </details>
365
 
366
  ### Training Logs
367
- | Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
368
- |:------:|:----:|:-------------:|:---------------:|:---------------:|
369
- | -1 | -1 | - | - | 0.8861 |
370
- | 0.0009 | 1 | 5.8495 | - | - |
371
- | 1.0 | 1139 | 3.0136 | 0.8482 | 0.9113 |
372
- | 2.0 | 2278 | 2.2096 | 0.7465 | 0.9241 |
373
- | 3.0 | 3417 | 1.966 | 0.6980 | 0.9337 |
374
 
375
 
376
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:291522
9
  - loss:MultipleNegativesSymmetricRankingLoss
 
10
  widget:
11
  - source_sentence: cream 21 baby oil with almond oil
12
  sentences:
 
40
  metrics:
41
  - cosine_accuracy
42
  model-index:
43
+ - name: SentenceTransformer
44
  results:
45
  - task:
46
  type: triplet
 
50
  type: unknown
51
  metrics:
52
  - type: cosine_accuracy
53
+ value: 0.9403471946716309
54
  name: Cosine Accuracy
55
  ---
56
 
57
+ # SentenceTransformer
58
 
59
+ This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
60
 
61
  ## Model Details
62
 
63
  ### Model Description
64
  - **Model Type:** Sentence Transformer
65
+ <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
66
  - **Maximum Sequence Length:** 256 tokens
67
  - **Output Dimensionality:** 384 dimensions
68
  - **Similarity Function:** Cosine Similarity
 
115
  # Get the similarity scores for the embeddings
116
  similarities = model.similarity(embeddings, embeddings)
117
  print(similarities)
118
+ # tensor([[1.0000, 0.7730, 0.3475],
119
+ # [0.7730, 1.0000, 0.3615],
120
+ # [0.3475, 0.3615, 1.0000]])
121
  ```
122
 
123
  <!--
 
154
 
155
  | Metric | Value |
156
  |:--------------------|:-----------|
157
+ | **cosine_accuracy** | **0.9403** |
158
 
159
  <!--
160
  ## Bias, Risks and Limitations
 
229
  - `per_device_train_batch_size`: 256
230
  - `per_device_eval_batch_size`: 256
231
  - `weight_decay`: 0.001
232
+ - `num_train_epochs`: 5
233
  - `warmup_steps`: 1138
234
  - `fp16`: True
235
  - `dataloader_num_workers`: 4
 
260
  - `adam_beta2`: 0.999
261
  - `adam_epsilon`: 1e-08
262
  - `max_grad_norm`: 1.0
263
+ - `num_train_epochs`: 5
264
  - `max_steps`: -1
265
  - `lr_scheduler_type`: linear
266
  - `lr_scheduler_kwargs`: {}
 
364
  </details>
365
 
366
  ### Training Logs
367
+ | Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
368
+ |:-----:|:----:|:-------------:|:---------------:|:---------------:|
369
+ | 4.0 | 4556 | 1.8731 | 0.7003 | 0.9331 |
370
+ | 5.0 | 5695 | 1.7998 | 0.6516 | 0.9403 |
 
 
 
371
 
372
 
373
  ### Framework Versions