LamaDiab commited on
Commit
82a7a1f
·
verified ·
1 Parent(s): c138d35

Updating model weights

Browse files
Files changed (1) hide show
  1. README.md +5 -40
README.md CHANGED
@@ -7,7 +7,6 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:291522
9
  - loss:MultipleNegativesSymmetricRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
  - source_sentence: cream 21 baby oil with almond oil
13
  sentences:
@@ -38,32 +37,17 @@ widget:
38
  - kids game
39
  pipeline_tag: sentence-similarity
40
  library_name: sentence-transformers
41
- metrics:
42
- - cosine_accuracy
43
- model-index:
44
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
45
- results:
46
- - task:
47
- type: triplet
48
- name: Triplet
49
- dataset:
50
- name: Unknown
51
- type: unknown
52
- metrics:
53
- - type: cosine_accuracy
54
- value: 0.9412940740585327
55
- name: Cosine Accuracy
56
  ---
57
 
58
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
59
 
60
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
61
 
62
  ## Model Details
63
 
64
  ### Model Description
65
  - **Model Type:** Sentence Transformer
66
- - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
67
  - **Maximum Sequence Length:** 256 tokens
68
  - **Output Dimensionality:** 384 dimensions
69
  - **Similarity Function:** Cosine Similarity
@@ -145,18 +129,6 @@ You can finetune this model on your own dataset.
145
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
146
  -->
147
 
148
- ## Evaluation
149
-
150
- ### Metrics
151
-
152
- #### Triplet
153
-
154
- * Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator)
155
-
156
- | Metric | Value |
157
- |:--------------------|:-----------|
158
- | **cosine_accuracy** | **0.9413** |
159
-
160
  <!--
161
  ## Bias, Risks and Limitations
162
 
@@ -226,7 +198,7 @@ You can finetune this model on your own dataset.
226
  ### Training Hyperparameters
227
  #### Non-Default Hyperparameters
228
 
229
- - `eval_strategy`: steps
230
  - `per_device_train_batch_size`: 128
231
  - `per_device_eval_batch_size`: 128
232
  - `weight_decay`: 0.001
@@ -245,7 +217,7 @@ You can finetune this model on your own dataset.
245
 
246
  - `overwrite_output_dir`: False
247
  - `do_predict`: False
248
- - `eval_strategy`: steps
249
  - `prediction_loss_only`: True
250
  - `per_device_train_batch_size`: 128
251
  - `per_device_eval_batch_size`: 128
@@ -363,13 +335,6 @@ You can finetune this model on your own dataset.
363
 
364
  </details>
365
 
366
- ### Training Logs
367
- | Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
368
- |:------:|:----:|:-------------:|:---------------:|:---------------:|
369
- | 0.0004 | 1 | 5.3655 | - | - |
370
- | 2.1949 | 5000 | 2.1423 | 0.7694 | 0.9413 |
371
-
372
-
373
  ### Framework Versions
374
  - Python: 3.11.13
375
  - Sentence Transformers: 5.1.2
 
7
  - generated_from_trainer
8
  - dataset_size:291522
9
  - loss:MultipleNegativesSymmetricRankingLoss
 
10
  widget:
11
  - source_sentence: cream 21 baby oil with almond oil
12
  sentences:
 
37
  - kids game
38
  pipeline_tag: sentence-similarity
39
  library_name: sentence-transformers
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ---
41
 
42
+ # SentenceTransformer
43
 
44
+ This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
45
 
46
  ## Model Details
47
 
48
  ### Model Description
49
  - **Model Type:** Sentence Transformer
50
+ <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
51
  - **Maximum Sequence Length:** 256 tokens
52
  - **Output Dimensionality:** 384 dimensions
53
  - **Similarity Function:** Cosine Similarity
 
129
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
130
  -->
131
 
 
 
 
 
 
 
 
 
 
 
 
 
132
  <!--
133
  ## Bias, Risks and Limitations
134
 
 
198
  ### Training Hyperparameters
199
  #### Non-Default Hyperparameters
200
 
201
+ - `eval_strategy`: epoch
202
  - `per_device_train_batch_size`: 128
203
  - `per_device_eval_batch_size`: 128
204
  - `weight_decay`: 0.001
 
217
 
218
  - `overwrite_output_dir`: False
219
  - `do_predict`: False
220
+ - `eval_strategy`: epoch
221
  - `prediction_loss_only`: True
222
  - `per_device_train_batch_size`: 128
223
  - `per_device_eval_batch_size`: 128
 
335
 
336
  </details>
337
 
 
 
 
 
 
 
 
338
  ### Framework Versions
339
  - Python: 3.11.13
340
  - Sentence Transformers: 5.1.2