LamaDiab commited on
Commit
1a4bf84
·
verified ·
1 Parent(s): 38e7084

Updating model weights

Browse files
Files changed (1) hide show
  1. README.md +16 -17
README.md CHANGED
@@ -7,7 +7,6 @@ tags:
7
  - generated_from_trainer
8
  - dataset_size:556626
9
  - loss:MultipleNegativesSymmetricRankingLoss
10
- base_model: sentence-transformers/all-MiniLM-L6-v2
11
  widget:
12
  - source_sentence: dimlaj orchid printed finest durable glass terkish tea set
13
  sentences:
@@ -39,7 +38,7 @@ library_name: sentence-transformers
39
  metrics:
40
  - cosine_accuracy
41
  model-index:
42
- - name: SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
43
  results:
44
  - task:
45
  type: triplet
@@ -49,19 +48,19 @@ model-index:
49
  type: unknown
50
  metrics:
51
  - type: cosine_accuracy
52
- value: 0.9607574939727783
53
  name: Cosine Accuracy
54
  ---
55
 
56
- # SentenceTransformer based on sentence-transformers/all-MiniLM-L6-v2
57
 
58
- This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2). It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
59
 
60
  ## Model Details
61
 
62
  ### Model Description
63
  - **Model Type:** Sentence Transformer
64
- - **Base model:** [sentence-transformers/all-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2) <!-- at revision c9745ed1d9f207416be6d2e6f8de32d1f16199bf -->
65
  - **Maximum Sequence Length:** 256 tokens
66
  - **Output Dimensionality:** 384 dimensions
67
  - **Similarity Function:** Cosine Similarity
@@ -114,9 +113,9 @@ print(embeddings.shape)
114
  # Get the similarity scores for the embeddings
115
  similarities = model.similarity(embeddings, embeddings)
116
  print(similarities)
117
- # tensor([[1.0000, 0.5010, 0.3796],
118
- # [0.5010, 1.0000, 0.3538],
119
- # [0.3796, 0.3538, 1.0000]])
120
  ```
121
 
122
  <!--
@@ -153,7 +152,7 @@ You can finetune this model on your own dataset.
153
 
154
  | Metric | Value |
155
  |:--------------------|:-----------|
156
- | **cosine_accuracy** | **0.9608** |
157
 
158
  <!--
159
  ## Bias, Risks and Limitations
@@ -228,6 +227,7 @@ You can finetune this model on your own dataset.
228
  - `per_device_train_batch_size`: 128
229
  - `per_device_eval_batch_size`: 128
230
  - `weight_decay`: 0.001
 
231
  - `warmup_steps`: 6956
232
  - `fp16`: True
233
  - `dataloader_num_workers`: 2
@@ -258,7 +258,7 @@ You can finetune this model on your own dataset.
258
  - `adam_beta2`: 0.999
259
  - `adam_epsilon`: 1e-08
260
  - `max_grad_norm`: 1.0
261
- - `num_train_epochs`: 3
262
  - `max_steps`: -1
263
  - `lr_scheduler_type`: linear
264
  - `lr_scheduler_kwargs`: {}
@@ -362,12 +362,11 @@ You can finetune this model on your own dataset.
362
  </details>
363
 
364
  ### Training Logs
365
- | Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
366
- |:------:|:-----:|:-------------:|:---------------:|:---------------:|
367
- | 0.0002 | 1 | 5.3185 | - | - |
368
- | 1.0 | 4349 | 2.6529 | 1.4502 | 0.9492 |
369
- | 2.0 | 8698 | 1.8024 | 1.3993 | 0.9600 |
370
- | 3.0 | 13047 | 1.4655 | 1.3219 | 0.9608 |
371
 
372
 
373
  ### Framework Versions
 
7
  - generated_from_trainer
8
  - dataset_size:556626
9
  - loss:MultipleNegativesSymmetricRankingLoss
 
10
  widget:
11
  - source_sentence: dimlaj orchid printed finest durable glass terkish tea set
12
  sentences:
 
38
  metrics:
39
  - cosine_accuracy
40
  model-index:
41
+ - name: SentenceTransformer
42
  results:
43
  - task:
44
  type: triplet
 
48
  type: unknown
49
  metrics:
50
  - type: cosine_accuracy
51
+ value: 0.9618095755577087
52
  name: Cosine Accuracy
53
  ---
54
 
55
+ # SentenceTransformer
56
 
57
+ This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 384-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more.
58
 
59
  ## Model Details
60
 
61
  ### Model Description
62
  - **Model Type:** Sentence Transformer
63
+ <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) -->
64
  - **Maximum Sequence Length:** 256 tokens
65
  - **Output Dimensionality:** 384 dimensions
66
  - **Similarity Function:** Cosine Similarity
 
113
  # Get the similarity scores for the embeddings
114
  similarities = model.similarity(embeddings, embeddings)
115
  print(similarities)
116
+ # tensor([[1.0000, 0.4517, 0.3474],
117
+ # [0.4517, 1.0000, 0.3222],
118
+ # [0.3474, 0.3222, 1.0000]])
119
  ```
120
 
121
  <!--
 
152
 
153
  | Metric | Value |
154
  |:--------------------|:-----------|
155
+ | **cosine_accuracy** | **0.9618** |
156
 
157
  <!--
158
  ## Bias, Risks and Limitations
 
227
  - `per_device_train_batch_size`: 128
228
  - `per_device_eval_batch_size`: 128
229
  - `weight_decay`: 0.001
230
+ - `num_train_epochs`: 6
231
  - `warmup_steps`: 6956
232
  - `fp16`: True
233
  - `dataloader_num_workers`: 2
 
258
  - `adam_beta2`: 0.999
259
  - `adam_epsilon`: 1e-08
260
  - `max_grad_norm`: 1.0
261
+ - `num_train_epochs`: 6
262
  - `max_steps`: -1
263
  - `lr_scheduler_type`: linear
264
  - `lr_scheduler_kwargs`: {}
 
362
  </details>
363
 
364
  ### Training Logs
365
+ | Epoch | Step | Training Loss | Validation Loss | cosine_accuracy |
366
+ |:-----:|:-----:|:-------------:|:---------------:|:---------------:|
367
+ | 4.0 | 17396 | 1.3564 | 1.3029 | 0.9600 |
368
+ | 5.0 | 21745 | 1.2501 | 1.3017 | 0.9622 |
369
+ | 6.0 | 26094 | 1.1858 | 1.2925 | 0.9618 |
 
370
 
371
 
372
  ### Framework Versions