radoslavralev commited on
Commit
370f794
·
verified ·
1 Parent(s): f35fedd

Training in progress, step 5000

Browse files
Information-Retrieval_evaluation_val_results.csv CHANGED
@@ -3,3 +3,4 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
3
  -1,-1,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402433333333332,0.9416250793650793,0.9545809774353143,0.9420576026548708
4
  -1,-1,0.8281,0.9026,0.93105,0.8281,0.8281,0.3008666666666666,0.9026,0.18621000000000004,0.93105,0.8281,0.8677437499999962,0.8721381249999942,0.8942437004811851,0.874246358340888
5
  -1,-1,0.82925,0.903025,0.931175,0.82925,0.82925,0.3010083333333333,0.903025,0.186235,0.931175,0.82925,0.8687345833333282,0.8731489384920591,0.8950131360828151,0.8752091976044037
 
 
3
  -1,-1,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402433333333332,0.9416250793650793,0.9545809774353143,0.9420576026548708
4
  -1,-1,0.8281,0.9026,0.93105,0.8281,0.8281,0.3008666666666666,0.9026,0.18621000000000004,0.93105,0.8281,0.8677437499999962,0.8721381249999942,0.8942437004811851,0.874246358340888
5
  -1,-1,0.82925,0.903025,0.931175,0.82925,0.82925,0.3010083333333333,0.903025,0.186235,0.931175,0.82925,0.8687345833333282,0.8731489384920591,0.8950131360828151,0.8752091976044037
6
+ -1,-1,0.7614,0.82615,0.850775,0.7614,0.7614,0.2753833333333333,0.82615,0.170155,0.850775,0.7614,0.7960862499999959,0.8003843253968239,0.8201550154419872,0.8038332983359062
README.md CHANGED
@@ -5,110 +5,38 @@ tags:
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
- - dataset_size:359999
9
  - loss:MultipleNegativesRankingLoss
10
  base_model: prajjwal1/bert-small
11
  widget:
12
- - source_sentence: Someone blocked me on Instagram. How do I unblock myself from their
13
- account?
14
  sentences:
15
- - Someone blocked me on Instagram. How do I unblock myself from their account?
16
- - Someone blocked me on Instagram. How do myself unblock Ifrom their account?
17
- - What are some good tips for dealing with a very easily frustrated 1 year old?
18
- - source_sentence: Do you love the life you live?
19
  sentences:
20
- - What is Jakob Nowell, Bradley Nowell's son, up to and will he pursue a career
21
- in music?
22
- - Do you love the life you're living?
23
- - Do you love not the life you live ?
24
- - source_sentence: I had sex on the 9th and my period started on the 11th. Could I
25
- still get pregnant?
26
  sentences:
27
- - How can I earn money easily online?
28
- - If I have sex on the day of my ovulation and I get my period two weeks later,
29
- can I still be pregnant?
30
- - I did not have sex on the 9th and my period started on the 11th . Could I still
31
- get pregnant ?
32
- - source_sentence: Would you read book at your office?
33
  sentences:
34
- - Would book read youat your office?
35
- - I am a married woman and I'm in love with married man. what should I do?
36
- - Would you read book at your office?
37
- - source_sentence: How do you earn money on Quora?
38
  sentences:
39
- - Ordered food on Swiggy 3 days ago.After accepting my money, said no more on Menu!
40
- When if ever will I atleast get refund in cr card a/c?
41
- - How do you earn not money on Quora ?
42
- - What is the best way to make money on Quora?
43
  pipeline_tag: sentence-similarity
44
  library_name: sentence-transformers
45
- metrics:
46
- - cosine_accuracy@1
47
- - cosine_accuracy@3
48
- - cosine_accuracy@5
49
- - cosine_precision@1
50
- - cosine_precision@3
51
- - cosine_precision@5
52
- - cosine_recall@1
53
- - cosine_recall@3
54
- - cosine_recall@5
55
- - cosine_ndcg@10
56
- - cosine_mrr@1
57
- - cosine_mrr@5
58
- - cosine_mrr@10
59
- - cosine_map@100
60
- model-index:
61
- - name: SentenceTransformer based on prajjwal1/bert-small
62
- results:
63
- - task:
64
- type: information-retrieval
65
- name: Information Retrieval
66
- dataset:
67
- name: val
68
- type: val
69
- metrics:
70
- - type: cosine_accuracy@1
71
- value: 0.761525
72
- name: Cosine Accuracy@1
73
- - type: cosine_accuracy@3
74
- value: 0.826125
75
- name: Cosine Accuracy@3
76
- - type: cosine_accuracy@5
77
- value: 0.85095
78
- name: Cosine Accuracy@5
79
- - type: cosine_precision@1
80
- value: 0.761525
81
- name: Cosine Precision@1
82
- - type: cosine_precision@3
83
- value: 0.275375
84
- name: Cosine Precision@3
85
- - type: cosine_precision@5
86
- value: 0.17019000000000004
87
- name: Cosine Precision@5
88
- - type: cosine_recall@1
89
- value: 0.761525
90
- name: Cosine Recall@1
91
- - type: cosine_recall@3
92
- value: 0.826125
93
- name: Cosine Recall@3
94
- - type: cosine_recall@5
95
- value: 0.85095
96
- name: Cosine Recall@5
97
- - type: cosine_ndcg@10
98
- value: 0.8202534934281767
99
- name: Cosine Ndcg@10
100
- - type: cosine_mrr@1
101
- value: 0.761525
102
- name: Cosine Mrr@1
103
- - type: cosine_mrr@5
104
- value: 0.7961479166666627
105
- name: Cosine Mrr@5
106
- - type: cosine_mrr@10
107
- value: 0.8004402281746008
108
- name: Cosine Mrr@10
109
- - type: cosine_map@100
110
- value: 0.8038638243708912
111
- name: Cosine Map@100
112
  ---
113
 
114
  # SentenceTransformer based on prajjwal1/bert-small
@@ -157,12 +85,12 @@ Then you can load this model and run inference.
157
  from sentence_transformers import SentenceTransformer
158
 
159
  # Download from the 🤗 Hub
160
- model = SentenceTransformer("redis/model-b-structured")
161
  # Run inference
162
  sentences = [
163
- 'How do you earn money on Quora?',
164
- 'What is the best way to make money on Quora?',
165
- 'How do you earn not money on Quora ?',
166
  ]
167
  embeddings = model.encode(sentences)
168
  print(embeddings.shape)
@@ -171,9 +99,9 @@ print(embeddings.shape)
171
  # Get the similarity scores for the embeddings
172
  similarities = model.similarity(embeddings, embeddings)
173
  print(similarities)
174
- # tensor([[ 1.0001, 0.9996, -0.3001],
175
- # [ 0.9996, 1.0000, -0.3007],
176
- # [-0.3001, -0.3007, 1.0003]])
177
  ```
178
 
179
  <!--
@@ -200,32 +128,6 @@ You can finetune this model on your own dataset.
200
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
201
  -->
202
 
203
- ## Evaluation
204
-
205
- ### Metrics
206
-
207
- #### Information Retrieval
208
-
209
- * Dataset: `val`
210
- * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
211
-
212
- | Metric | Value |
213
- |:-------------------|:-----------|
214
- | cosine_accuracy@1 | 0.7615 |
215
- | cosine_accuracy@3 | 0.8261 |
216
- | cosine_accuracy@5 | 0.8509 |
217
- | cosine_precision@1 | 0.7615 |
218
- | cosine_precision@3 | 0.2754 |
219
- | cosine_precision@5 | 0.1702 |
220
- | cosine_recall@1 | 0.7615 |
221
- | cosine_recall@3 | 0.8261 |
222
- | cosine_recall@5 | 0.8509 |
223
- | **cosine_ndcg@10** | **0.8203** |
224
- | cosine_mrr@1 | 0.7615 |
225
- | cosine_mrr@5 | 0.7961 |
226
- | cosine_mrr@10 | 0.8004 |
227
- | cosine_map@100 | 0.8039 |
228
-
229
  <!--
230
  ## Bias, Risks and Limitations
231
 
@@ -244,49 +146,23 @@ You can finetune this model on your own dataset.
244
 
245
  #### Unnamed Dataset
246
 
247
- * Size: 359,999 training samples
248
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
249
  * Approximate statistics based on the first 1000 samples:
250
- | | anchor | positive | negative |
251
- |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
252
- | type | string | string | string |
253
- | details | <ul><li>min: 6 tokens</li><li>mean: 15.4 tokens</li><li>max: 60 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.45 tokens</li><li>max: 78 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.07 tokens</li><li>max: 62 tokens</li></ul> |
254
- * Samples:
255
- | anchor | positive | negative |
256
- |:--------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------|
257
- | <code>Shall I upgrade my iPhone 5s to iOS 10 final version?</code> | <code>Should I upgrade an iPhone 5s to iOS 10?</code> | <code>Shall I upgrade not my iPhone 5s to iOS 10 final version ?</code> |
258
- | <code>Do Census Bureau income figures count sources of unearned income, or do they just count earned income?</code> | <code>Do Census Bureau income figures count sources of unearned income, or do they just count earned income?</code> | <code>Do Census Bureau income figures count sources of unearned income, or do income just count earned they?</code> |
259
- | <code>Who has the highest IQ?</code> | <code>Who has the highest IQ?</code> | <code>the highest IQ has Who?</code> |
260
- * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
261
- ```json
262
- {
263
- "scale": 1.0,
264
- "similarity_fct": "cos_sim",
265
- "gather_across_devices": false
266
- }
267
- ```
268
-
269
- ### Evaluation Dataset
270
-
271
- #### Unnamed Dataset
272
-
273
- * Size: 40,000 evaluation samples
274
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
275
- * Approximate statistics based on the first 1000 samples:
276
- | | anchor | positive | negative |
277
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
278
  | type | string | string | string |
279
- | details | <ul><li>min: 6 tokens</li><li>mean: 15.86 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.94 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.46 tokens</li><li>max: 66 tokens</li></ul> |
280
  * Samples:
281
- | anchor | positive | negative |
282
- |:------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------------------------------------------|
283
- | <code>What are some mind-blowing Iphone gadgets and tools that most people don't know about?</code> | <code>What are some mind-blowing iphone tools that most people don't know about?</code> | <code>most people are some mind-blowing Iphone gadgets and tools that Whatdon't know about?</code> |
284
- | <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>If FOX News is not the conservative news station , which cable news network is for liberals / progressives ?</code> |
285
- | <code>How can guys last longer during sex?</code> | <code>How do I last longer in sex?</code> | <code>How can guys last not longer during sex ?</code> |
286
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
287
  ```json
288
  {
289
- "scale": 1.0,
290
  "similarity_fct": "cos_sim",
291
  "gather_across_devices": false
292
  }
@@ -295,49 +171,36 @@ You can finetune this model on your own dataset.
295
  ### Training Hyperparameters
296
  #### Non-Default Hyperparameters
297
 
298
- - `eval_strategy`: steps
299
- - `per_device_train_batch_size`: 256
300
- - `per_device_eval_batch_size`: 256
301
- - `learning_rate`: 2e-05
302
- - `weight_decay`: 0.001
303
- - `max_steps`: 14060
304
- - `warmup_ratio`: 0.1
305
  - `fp16`: True
306
- - `dataloader_drop_last`: True
307
- - `dataloader_num_workers`: 1
308
- - `dataloader_prefetch_factor`: 1
309
- - `load_best_model_at_end`: True
310
- - `optim`: adamw_torch
311
- - `ddp_find_unused_parameters`: False
312
- - `push_to_hub`: True
313
- - `hub_model_id`: redis/model-b-structured
314
- - `eval_on_start`: True
315
 
316
  #### All Hyperparameters
317
  <details><summary>Click to expand</summary>
318
 
319
  - `overwrite_output_dir`: False
320
  - `do_predict`: False
321
- - `eval_strategy`: steps
322
  - `prediction_loss_only`: True
323
- - `per_device_train_batch_size`: 256
324
- - `per_device_eval_batch_size`: 256
325
  - `per_gpu_train_batch_size`: None
326
  - `per_gpu_eval_batch_size`: None
327
  - `gradient_accumulation_steps`: 1
328
  - `eval_accumulation_steps`: None
329
  - `torch_empty_cache_steps`: None
330
- - `learning_rate`: 2e-05
331
- - `weight_decay`: 0.001
332
  - `adam_beta1`: 0.9
333
  - `adam_beta2`: 0.999
334
  - `adam_epsilon`: 1e-08
335
- - `max_grad_norm`: 1.0
336
- - `num_train_epochs`: 3.0
337
- - `max_steps`: 14060
338
  - `lr_scheduler_type`: linear
339
  - `lr_scheduler_kwargs`: {}
340
- - `warmup_ratio`: 0.1
341
  - `warmup_steps`: 0
342
  - `log_level`: passive
343
  - `log_level_replica`: warning
@@ -365,14 +228,14 @@ You can finetune this model on your own dataset.
365
  - `tpu_num_cores`: None
366
  - `tpu_metrics_debug`: False
367
  - `debug`: []
368
- - `dataloader_drop_last`: True
369
- - `dataloader_num_workers`: 1
370
- - `dataloader_prefetch_factor`: 1
371
  - `past_index`: -1
372
  - `disable_tqdm`: False
373
  - `remove_unused_columns`: True
374
  - `label_names`: None
375
- - `load_best_model_at_end`: True
376
  - `ignore_data_skip`: False
377
  - `fsdp`: []
378
  - `fsdp_min_num_params`: 0
@@ -382,23 +245,23 @@ You can finetune this model on your own dataset.
382
  - `parallelism_config`: None
383
  - `deepspeed`: None
384
  - `label_smoothing_factor`: 0.0
385
- - `optim`: adamw_torch
386
  - `optim_args`: None
387
  - `adafactor`: False
388
  - `group_by_length`: False
389
  - `length_column_name`: length
390
  - `project`: huggingface
391
  - `trackio_space_id`: trackio
392
- - `ddp_find_unused_parameters`: False
393
  - `ddp_bucket_cap_mb`: None
394
  - `ddp_broadcast_buffers`: False
395
  - `dataloader_pin_memory`: True
396
  - `dataloader_persistent_workers`: False
397
  - `skip_memory_metrics`: True
398
  - `use_legacy_prediction_loop`: False
399
- - `push_to_hub`: True
400
  - `resume_from_checkpoint`: None
401
- - `hub_model_id`: redis/model-b-structured
402
  - `hub_strategy`: every_save
403
  - `hub_private_repo`: None
404
  - `hub_always_push`: False
@@ -425,167 +288,32 @@ You can finetune this model on your own dataset.
425
  - `neftune_noise_alpha`: None
426
  - `optim_target_modules`: None
427
  - `batch_eval_metrics`: False
428
- - `eval_on_start`: True
429
  - `use_liger_kernel`: False
430
  - `liger_kernel_config`: None
431
  - `eval_use_gather_object`: False
432
  - `average_tokens_across_devices`: True
433
  - `prompts`: None
434
  - `batch_sampler`: batch_sampler
435
- - `multi_dataset_batch_sampler`: proportional
436
  - `router_mapping`: {}
437
  - `learning_rate_mapping`: {}
438
 
439
  </details>
440
 
441
  ### Training Logs
442
- <details><summary>Click to expand</summary>
 
 
 
 
 
 
 
 
 
 
443
 
444
- | Epoch | Step | Training Loss | Validation Loss | val_cosine_ndcg@10 |
445
- |:------:|:-----:|:-------------:|:---------------:|:------------------:|
446
- | 0 | 0 | - | 5.9531 | 0.7603 |
447
- | 0.0711 | 100 | 5.9694 | 5.7072 | 0.7792 |
448
- | 0.1422 | 200 | 5.7181 | 5.4263 | 0.7865 |
449
- | 0.2134 | 300 | 5.5628 | 5.3443 | 0.7829 |
450
- | 0.2845 | 400 | 5.4947 | 5.3221 | 0.7774 |
451
- | 0.3556 | 500 | 5.4597 | 5.3180 | 0.7741 |
452
- | 0.4267 | 600 | 5.4387 | 5.3158 | 0.7737 |
453
- | 0.4979 | 700 | 5.423 | 5.3141 | 0.7751 |
454
- | 0.5690 | 800 | 5.4108 | 5.3109 | 0.7848 |
455
- | 0.6401 | 900 | 5.397 | 5.2923 | 0.8008 |
456
- | 0.7112 | 1000 | 5.3724 | 5.2839 | 0.8004 |
457
- | 0.7824 | 1100 | 5.3625 | 5.2804 | 0.8007 |
458
- | 0.8535 | 1200 | 5.355 | 5.2777 | 0.8013 |
459
- | 0.9246 | 1300 | 5.3499 | 5.2748 | 0.8030 |
460
- | 0.9957 | 1400 | 5.3442 | 5.2729 | 0.8067 |
461
- | 1.0669 | 1500 | 5.3382 | 5.2624 | 0.8103 |
462
- | 1.1380 | 1600 | 5.3254 | 5.2557 | 0.8138 |
463
- | 1.2091 | 1700 | 5.3159 | 5.2441 | 0.8163 |
464
- | 1.2802 | 1800 | 5.3035 | 5.2350 | 0.8180 |
465
- | 1.3514 | 1900 | 5.295 | 5.2303 | 0.8179 |
466
- | 1.4225 | 2000 | 5.2925 | 5.2292 | 0.8182 |
467
- | 1.4936 | 2100 | 5.2881 | 5.2271 | 0.8187 |
468
- | 1.5647 | 2200 | 5.2854 | 5.2258 | 0.8187 |
469
- | 1.6358 | 2300 | 5.2831 | 5.2258 | 0.8189 |
470
- | 1.7070 | 2400 | 5.2805 | 5.2247 | 0.8192 |
471
- | 1.7781 | 2500 | 5.278 | 5.2247 | 0.8186 |
472
- | 1.8492 | 2600 | 5.2761 | 5.2230 | 0.8184 |
473
- | 1.9203 | 2700 | 5.2754 | 5.2221 | 0.8185 |
474
- | 1.9915 | 2800 | 5.274 | 5.2228 | 0.8185 |
475
- | 2.0626 | 2900 | 5.2722 | 5.2209 | 0.8175 |
476
- | 2.1337 | 3000 | 5.2708 | 5.2206 | 0.8182 |
477
- | 2.2048 | 3100 | 5.2686 | 5.2211 | 0.8194 |
478
- | 2.2760 | 3200 | 5.2666 | 5.2204 | 0.8186 |
479
- | 2.3471 | 3300 | 5.2671 | 5.2192 | 0.8191 |
480
- | 2.4182 | 3400 | 5.2657 | 5.2200 | 0.8188 |
481
- | 2.4893 | 3500 | 5.2638 | 5.2188 | 0.8184 |
482
- | 2.5605 | 3600 | 5.2635 | 5.2189 | 0.8188 |
483
- | 2.6316 | 3700 | 5.2624 | 5.2187 | 0.8192 |
484
- | 2.7027 | 3800 | 5.262 | 5.2178 | 0.8182 |
485
- | 2.7738 | 3900 | 5.2608 | 5.2175 | 0.8188 |
486
- | 2.8450 | 4000 | 5.2595 | 5.2179 | 0.8189 |
487
- | 2.9161 | 4100 | 5.2599 | 5.2163 | 0.8191 |
488
- | 2.9872 | 4200 | 5.2587 | 5.2162 | 0.8184 |
489
- | 3.0583 | 4300 | 5.2574 | 5.2168 | 0.8193 |
490
- | 3.1294 | 4400 | 5.256 | 5.2165 | 0.8197 |
491
- | 3.2006 | 4500 | 5.2551 | 5.2158 | 0.8188 |
492
- | 3.2717 | 4600 | 5.2552 | 5.2159 | 0.8188 |
493
- | 3.3428 | 4700 | 5.2549 | 5.2157 | 0.8192 |
494
- | 3.4139 | 4800 | 5.2531 | 5.2154 | 0.8192 |
495
- | 3.4851 | 4900 | 5.2534 | 5.2152 | 0.8191 |
496
- | 3.5562 | 5000 | 5.2528 | 5.2146 | 0.8197 |
497
- | 3.6273 | 5100 | 5.2521 | 5.2149 | 0.8193 |
498
- | 3.6984 | 5200 | 5.2509 | 5.2145 | 0.8199 |
499
- | 3.7696 | 5300 | 5.2509 | 5.2144 | 0.8189 |
500
- | 3.8407 | 5400 | 5.2495 | 5.2139 | 0.8195 |
501
- | 3.9118 | 5500 | 5.2496 | 5.2140 | 0.8195 |
502
- | 3.9829 | 5600 | 5.2505 | 5.2135 | 0.8193 |
503
- | 4.0541 | 5700 | 5.2478 | 5.2140 | 0.8197 |
504
- | 4.1252 | 5800 | 5.2476 | 5.2136 | 0.8196 |
505
- | 4.1963 | 5900 | 5.248 | 5.2130 | 0.8199 |
506
- | 4.2674 | 6000 | 5.2482 | 5.2129 | 0.8196 |
507
- | 4.3385 | 6100 | 5.2466 | 5.2135 | 0.8196 |
508
- | 4.4097 | 6200 | 5.2461 | 5.2126 | 0.8196 |
509
- | 4.4808 | 6300 | 5.2453 | 5.2124 | 0.8196 |
510
- | 4.5519 | 6400 | 5.2448 | 5.2128 | 0.8197 |
511
- | 4.6230 | 6500 | 5.2439 | 5.2124 | 0.8193 |
512
- | 4.6942 | 6600 | 5.244 | 5.2123 | 0.8192 |
513
- | 4.7653 | 6700 | 5.2428 | 5.2114 | 0.8192 |
514
- | 4.8364 | 6800 | 5.2433 | 5.2112 | 0.8197 |
515
- | 4.9075 | 6900 | 5.2439 | 5.2117 | 0.8194 |
516
- | 4.9787 | 7000 | 5.2422 | 5.2121 | 0.8204 |
517
- | 5.0498 | 7100 | 5.2425 | 5.2114 | 0.8198 |
518
- | 5.1209 | 7200 | 5.2418 | 5.2113 | 0.8201 |
519
- | 5.1920 | 7300 | 5.2416 | 5.2113 | 0.8200 |
520
- | 5.2632 | 7400 | 5.2405 | 5.2109 | 0.8199 |
521
- | 5.3343 | 7500 | 5.242 | 5.2106 | 0.8197 |
522
- | 5.4054 | 7600 | 5.2402 | 5.2105 | 0.8199 |
523
- | 5.4765 | 7700 | 5.2393 | 5.2108 | 0.8203 |
524
- | 5.5477 | 7800 | 5.24 | 5.2104 | 0.8198 |
525
- | 5.6188 | 7900 | 5.2395 | 5.2103 | 0.8201 |
526
- | 5.6899 | 8000 | 5.2381 | 5.2102 | 0.8198 |
527
- | 5.7610 | 8100 | 5.2399 | 5.2102 | 0.8195 |
528
- | 5.8321 | 8200 | 5.2395 | 5.2100 | 0.8195 |
529
- | 5.9033 | 8300 | 5.2377 | 5.2100 | 0.8197 |
530
- | 5.9744 | 8400 | 5.238 | 5.2097 | 0.8198 |
531
- | 6.0455 | 8500 | 5.2372 | 5.2097 | 0.8200 |
532
- | 6.1166 | 8600 | 5.2368 | 5.2095 | 0.8200 |
533
- | 6.1878 | 8700 | 5.2378 | 5.2096 | 0.8201 |
534
- | 6.2589 | 8800 | 5.2372 | 5.2097 | 0.8197 |
535
- | 6.3300 | 8900 | 5.2365 | 5.2098 | 0.8197 |
536
- | 6.4011 | 9000 | 5.2367 | 5.2092 | 0.8199 |
537
- | 6.4723 | 9100 | 5.2364 | 5.2093 | 0.8197 |
538
- | 6.5434 | 9200 | 5.2362 | 5.2095 | 0.8202 |
539
- | 6.6145 | 9300 | 5.2359 | 5.2096 | 0.8199 |
540
- | 6.6856 | 9400 | 5.2345 | 5.2095 | 0.8203 |
541
- | 6.7568 | 9500 | 5.2362 | 5.2090 | 0.8202 |
542
- | 6.8279 | 9600 | 5.2353 | 5.2089 | 0.8201 |
543
- | 6.8990 | 9700 | 5.2346 | 5.2090 | 0.8203 |
544
- | 6.9701 | 9800 | 5.2354 | 5.2090 | 0.8202 |
545
- | 7.0413 | 9900 | 5.234 | 5.2089 | 0.8202 |
546
- | 7.1124 | 10000 | 5.2334 | 5.2087 | 0.8202 |
547
- | 7.1835 | 10100 | 5.2342 | 5.2089 | 0.8204 |
548
- | 7.2546 | 10200 | 5.2342 | 5.2089 | 0.8204 |
549
- | 7.3257 | 10300 | 5.2336 | 5.2085 | 0.8203 |
550
- | 7.3969 | 10400 | 5.2347 | 5.2086 | 0.8206 |
551
- | 7.4680 | 10500 | 5.2326 | 5.2086 | 0.8203 |
552
- | 7.5391 | 10600 | 5.2336 | 5.2082 | 0.8201 |
553
- | 7.6102 | 10700 | 5.2328 | 5.2084 | 0.8202 |
554
- | 7.6814 | 10800 | 5.2328 | 5.2085 | 0.8203 |
555
- | 7.7525 | 10900 | 5.2321 | 5.2083 | 0.8201 |
556
- | 7.8236 | 11000 | 5.2332 | 5.2082 | 0.8202 |
557
- | 7.8947 | 11100 | 5.2325 | 5.2082 | 0.8202 |
558
- | 7.9659 | 11200 | 5.2331 | 5.2082 | 0.8200 |
559
- | 8.0370 | 11300 | 5.2322 | 5.2081 | 0.8202 |
560
- | 8.1081 | 11400 | 5.2324 | 5.2082 | 0.8206 |
561
- | 8.1792 | 11500 | 5.2318 | 5.2080 | 0.8200 |
562
- | 8.2504 | 11600 | 5.2314 | 5.2082 | 0.8202 |
563
- | 8.3215 | 11700 | 5.2318 | 5.2082 | 0.8202 |
564
- | 8.3926 | 11800 | 5.2317 | 5.2078 | 0.8203 |
565
- | 8.4637 | 11900 | 5.2312 | 5.2078 | 0.8202 |
566
- | 8.5349 | 12000 | 5.2327 | 5.2079 | 0.8201 |
567
- | 8.6060 | 12100 | 5.2316 | 5.2077 | 0.8203 |
568
- | 8.6771 | 12200 | 5.2317 | 5.2078 | 0.8204 |
569
- | 8.7482 | 12300 | 5.2301 | 5.2079 | 0.8202 |
570
- | 8.8193 | 12400 | 5.2308 | 5.2077 | 0.8201 |
571
- | 8.8905 | 12500 | 5.2306 | 5.2078 | 0.8200 |
572
- | 8.9616 | 12600 | 5.231 | 5.2077 | 0.8200 |
573
- | 9.0327 | 12700 | 5.2307 | 5.2076 | 0.8199 |
574
- | 9.1038 | 12800 | 5.2309 | 5.2076 | 0.8201 |
575
- | 9.1750 | 12900 | 5.2301 | 5.2076 | 0.8200 |
576
- | 9.2461 | 13000 | 5.231 | 5.2076 | 0.8202 |
577
- | 9.3172 | 13100 | 5.2312 | 5.2075 | 0.8201 |
578
- | 9.3883 | 13200 | 5.2304 | 5.2077 | 0.8204 |
579
- | 9.4595 | 13300 | 5.2304 | 5.2075 | 0.8202 |
580
- | 9.5306 | 13400 | 5.2312 | 5.2076 | 0.8203 |
581
- | 9.6017 | 13500 | 5.2304 | 5.2076 | 0.8204 |
582
- | 9.6728 | 13600 | 5.2309 | 5.2076 | 0.8203 |
583
- | 9.7440 | 13700 | 5.23 | 5.2075 | 0.8202 |
584
- | 9.8151 | 13800 | 5.2301 | 5.2075 | 0.8201 |
585
- | 9.8862 | 13900 | 5.231 | 5.2075 | 0.8203 |
586
- | 9.9573 | 14000 | 5.2303 | 5.2075 | 0.8203 |
587
-
588
- </details>
589
 
590
  ### Framework Versions
591
  - Python: 3.10.18
 
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
+ - dataset_size:100000
9
  - loss:MultipleNegativesRankingLoss
10
  base_model: prajjwal1/bert-small
11
  widget:
12
+ - source_sentence: How do I calculate IQ?
 
13
  sentences:
14
+ - What is the easiest way to know my IQ?
15
+ - How do I calculate not IQ ?
16
+ - What are some creative and innovative business ideas with less investment in India?
17
+ - source_sentence: How can I learn martial arts in my home?
18
  sentences:
19
+ - How can I learn martial arts by myself?
20
+ - What are the advantages and disadvantages of investing in gold?
21
+ - Can people see that I have looked at their pictures on instagram if I am not following
22
+ them?
23
+ - source_sentence: When Enterprise picks you up do you have to take them back?
 
24
  sentences:
25
+ - Are there any software Training institute in Tuticorin?
26
+ - When Enterprise picks you up do you have to take them back?
27
+ - When Enterprise picks you up do them have to take youback?
28
+ - source_sentence: What are some non-capital goods?
 
 
29
  sentences:
30
+ - What are capital goods?
31
+ - How is the value of [math]\pi[/math] calculated?
32
+ - What are some non-capital goods?
33
+ - source_sentence: What is the QuickBooks technical support phone number in New York?
34
  sentences:
35
+ - What caused the Great Depression?
36
+ - Can I apply for PR in Canada?
37
+ - Which is the best QuickBooks Hosting Support Number in New York?
 
38
  pipeline_tag: sentence-similarity
39
  library_name: sentence-transformers
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ---
41
 
42
  # SentenceTransformer based on prajjwal1/bert-small
 
85
  from sentence_transformers import SentenceTransformer
86
 
87
  # Download from the 🤗 Hub
88
+ model = SentenceTransformer("sentence_transformers_model_id")
89
  # Run inference
90
  sentences = [
91
+ 'What is the QuickBooks technical support phone number in New York?',
92
+ 'Which is the best QuickBooks Hosting Support Number in New York?',
93
+ 'Can I apply for PR in Canada?',
94
  ]
95
  embeddings = model.encode(sentences)
96
  print(embeddings.shape)
 
99
  # Get the similarity scores for the embeddings
100
  similarities = model.similarity(embeddings, embeddings)
101
  print(similarities)
102
+ # tensor([[1.0000, 0.8563, 0.0594],
103
+ # [0.8563, 1.0000, 0.1245],
104
+ # [0.0594, 0.1245, 1.0000]])
105
  ```
106
 
107
  <!--
 
128
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
129
  -->
130
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  <!--
132
  ## Bias, Risks and Limitations
133
 
 
146
 
147
  #### Unnamed Dataset
148
 
149
+ * Size: 100,000 training samples
150
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
151
  * Approximate statistics based on the first 1000 samples:
152
+ | | sentence_0 | sentence_1 | sentence_2 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
153
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
154
  | type | string | string | string |
155
+ | details | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.68 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 16.37 tokens</li><li>max: 67 tokens</li></ul> |
156
  * Samples:
157
+ | sentence_0 | sentence_1 | sentence_2 |
158
+ |:-----------------------------------------------------------------|:-----------------------------------------------------------------|:----------------------------------------------------------------------------------|
159
+ | <code>Is masturbating bad for boys?</code> | <code>Is masturbating bad for boys?</code> | <code>How harmful or unhealthy is masturbation?</code> |
160
+ | <code>Does a train engine move in reverse?</code> | <code>Does a train engine move in reverse?</code> | <code>Time moves forward, not in reverse. Doesn't that make time a vector?</code> |
161
+ | <code>What is the most badass thing anyone has ever done?</code> | <code>What is the most badass thing anyone has ever done?</code> | <code>anyone is the most badass thing Whathas ever done?</code> |
162
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
163
  ```json
164
  {
165
+ "scale": 20.0,
166
  "similarity_fct": "cos_sim",
167
  "gather_across_devices": false
168
  }
 
171
  ### Training Hyperparameters
172
  #### Non-Default Hyperparameters
173
 
174
+ - `per_device_train_batch_size`: 64
175
+ - `per_device_eval_batch_size`: 64
 
 
 
 
 
176
  - `fp16`: True
177
+ - `multi_dataset_batch_sampler`: round_robin
 
 
 
 
 
 
 
 
178
 
179
  #### All Hyperparameters
180
  <details><summary>Click to expand</summary>
181
 
182
  - `overwrite_output_dir`: False
183
  - `do_predict`: False
184
+ - `eval_strategy`: no
185
  - `prediction_loss_only`: True
186
+ - `per_device_train_batch_size`: 64
187
+ - `per_device_eval_batch_size`: 64
188
  - `per_gpu_train_batch_size`: None
189
  - `per_gpu_eval_batch_size`: None
190
  - `gradient_accumulation_steps`: 1
191
  - `eval_accumulation_steps`: None
192
  - `torch_empty_cache_steps`: None
193
+ - `learning_rate`: 5e-05
194
+ - `weight_decay`: 0.0
195
  - `adam_beta1`: 0.9
196
  - `adam_beta2`: 0.999
197
  - `adam_epsilon`: 1e-08
198
+ - `max_grad_norm`: 1
199
+ - `num_train_epochs`: 3
200
+ - `max_steps`: -1
201
  - `lr_scheduler_type`: linear
202
  - `lr_scheduler_kwargs`: {}
203
+ - `warmup_ratio`: 0.0
204
  - `warmup_steps`: 0
205
  - `log_level`: passive
206
  - `log_level_replica`: warning
 
228
  - `tpu_num_cores`: None
229
  - `tpu_metrics_debug`: False
230
  - `debug`: []
231
+ - `dataloader_drop_last`: False
232
+ - `dataloader_num_workers`: 0
233
+ - `dataloader_prefetch_factor`: None
234
  - `past_index`: -1
235
  - `disable_tqdm`: False
236
  - `remove_unused_columns`: True
237
  - `label_names`: None
238
+ - `load_best_model_at_end`: False
239
  - `ignore_data_skip`: False
240
  - `fsdp`: []
241
  - `fsdp_min_num_params`: 0
 
245
  - `parallelism_config`: None
246
  - `deepspeed`: None
247
  - `label_smoothing_factor`: 0.0
248
+ - `optim`: adamw_torch_fused
249
  - `optim_args`: None
250
  - `adafactor`: False
251
  - `group_by_length`: False
252
  - `length_column_name`: length
253
  - `project`: huggingface
254
  - `trackio_space_id`: trackio
255
+ - `ddp_find_unused_parameters`: None
256
  - `ddp_bucket_cap_mb`: None
257
  - `ddp_broadcast_buffers`: False
258
  - `dataloader_pin_memory`: True
259
  - `dataloader_persistent_workers`: False
260
  - `skip_memory_metrics`: True
261
  - `use_legacy_prediction_loop`: False
262
+ - `push_to_hub`: False
263
  - `resume_from_checkpoint`: None
264
+ - `hub_model_id`: None
265
  - `hub_strategy`: every_save
266
  - `hub_private_repo`: None
267
  - `hub_always_push`: False
 
288
  - `neftune_noise_alpha`: None
289
  - `optim_target_modules`: None
290
  - `batch_eval_metrics`: False
291
+ - `eval_on_start`: False
292
  - `use_liger_kernel`: False
293
  - `liger_kernel_config`: None
294
  - `eval_use_gather_object`: False
295
  - `average_tokens_across_devices`: True
296
  - `prompts`: None
297
  - `batch_sampler`: batch_sampler
298
+ - `multi_dataset_batch_sampler`: round_robin
299
  - `router_mapping`: {}
300
  - `learning_rate_mapping`: {}
301
 
302
  </details>
303
 
304
  ### Training Logs
305
+ | Epoch | Step | Training Loss |
306
+ |:------:|:----:|:-------------:|
307
+ | 0.3199 | 500 | 0.4294 |
308
+ | 0.6398 | 1000 | 0.1268 |
309
+ | 0.9597 | 1500 | 0.1 |
310
+ | 1.2796 | 2000 | 0.0792 |
311
+ | 1.5995 | 2500 | 0.0706 |
312
+ | 1.9194 | 3000 | 0.0687 |
313
+ | 2.2393 | 3500 | 0.0584 |
314
+ | 2.5592 | 4000 | 0.057 |
315
+ | 2.8791 | 4500 | 0.0581 |
316
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
317
 
318
  ### Framework Versions
319
  - Python: 3.10.18
eval/Information-Retrieval_evaluation_val_results.csv CHANGED
@@ -461,3 +461,54 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
461
  9.815078236130867,13800,0.761775,0.825825,0.85075,0.761775,0.761775,0.275275,0.825825,0.17015,0.85075,0.761775,0.7961812499999961,0.8004442063492047,0.8201302361968327,0.8038924254433807
462
  9.88620199146515,13900,0.76165,0.8262,0.8508,0.76165,0.76165,0.2754,0.8262,0.17016,0.8508,0.76165,0.7961833333333295,0.8005036607142841,0.8202859573683243,0.8039239466787359
463
  9.95732574679943,14000,0.761525,0.826125,0.85095,0.761525,0.761525,0.275375,0.826125,0.17019000000000004,0.85095,0.761525,0.7961479166666627,0.8004402281746008,0.8202534934281767,0.8038638243708912
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
461
  9.815078236130867,13800,0.761775,0.825825,0.85075,0.761775,0.761775,0.275275,0.825825,0.17015,0.85075,0.761775,0.7961812499999961,0.8004442063492047,0.8201302361968327,0.8038924254433807
462
  9.88620199146515,13900,0.76165,0.8262,0.8508,0.76165,0.76165,0.2754,0.8262,0.17016,0.8508,0.76165,0.7961833333333295,0.8005036607142841,0.8202859573683243,0.8039239466787359
463
  9.95732574679943,14000,0.761525,0.826125,0.85095,0.761525,0.761525,0.275375,0.826125,0.17019000000000004,0.85095,0.761525,0.7961479166666627,0.8004402281746008,0.8202534934281767,0.8038638243708912
464
+ 0,0,0.70285,0.79605,0.8218,0.70285,0.70285,0.26535,0.79605,0.16436,0.8218,0.70285,0.7512891666666622,0.7555783333333311,0.7794223255349312,0.7588558044000856
465
+ 0.14347202295552366,100,0.719,0.84285,0.872675,0.719,0.719,0.28095,0.84285,0.174535,0.872675,0.719,0.7830449999999936,0.7875145634920601,0.8164610131170594,0.7906443623737947
466
+ 0.28694404591104733,200,0.720475,0.8441,0.8731,0.720475,0.720475,0.2813666666666666,0.8441,0.17462000000000003,0.8731,0.720475,0.7841733333333266,0.7884872916666616,0.8171180249207256,0.7915054648946475
467
+ 0.430416068866571,300,0.734225,0.847,0.87485,0.734225,0.734225,0.2823333333333333,0.847,0.17497000000000001,0.87485,0.734225,0.792364583333327,0.7966418452380913,0.823543683794261,0.7997161682861882
468
+ 0.5738880918220947,400,0.767825,0.85415,0.879475,0.767825,0.767825,0.2847166666666667,0.85415,0.17589500000000002,0.879475,0.767825,0.8127879166666606,0.8170799999999967,0.8399432792184387,0.8200855451322794
469
+ 0.7173601147776184,500,0.7684,0.85535,0.881075,0.7684,0.7684,0.28511666666666663,0.85535,0.176215,0.881075,0.7684,0.813672499999994,0.8180551190476166,0.8411684249227638,0.8211524512569452
470
+ 0.860832137733142,600,0.770075,0.856575,0.883375,0.770075,0.770075,0.285525,0.856575,0.176675,0.883375,0.770075,0.8152941666666604,0.8196073908730117,0.8428450304728525,0.8226640279051873
471
+ 1.0043041606886658,700,0.77065,0.8581,0.884225,0.77065,0.77065,0.2860333333333333,0.8581,0.176845,0.884225,0.77065,0.8159979166666597,0.8204329067460283,0.8438585312774513,0.8234774368034475
472
+ 1.1477761836441893,800,0.772,0.8595,0.885475,0.772,0.772,0.2864999999999999,0.8595,0.177095,0.885475,0.772,0.8172545833333261,0.8216476984126937,0.8450234064355687,0.8246727126545477
473
+ 1.291248206599713,900,0.772025,0.859125,0.8859,0.772025,0.772025,0.28637499999999994,0.859125,0.17718000000000003,0.8859,0.772025,0.8173566666666607,0.8217971825396798,0.8453118351395867,0.8248254459650096
474
+ 1.4347202295552366,1000,0.7729,0.859775,0.886875,0.7729,0.7729,0.28659166666666663,0.859775,0.17737500000000003,0.886875,0.7729,0.8180929166666607,0.8224405753968238,0.8458532846321035,0.825523835584324
475
+ 1.5781922525107603,1100,0.773575,0.860275,0.88775,0.773575,0.773575,0.2867583333333333,0.860275,0.17755,0.88775,0.773575,0.8187566666666616,0.823100605158728,0.8465896111780518,0.826152915497014
476
+ 1.721664275466284,1200,0.77435,0.8607,0.88835,0.77435,0.77435,0.28689999999999993,0.8607,0.17767,0.88835,0.77435,0.819414999999994,0.8237554067460292,0.8471635033699103,0.8268239553737456
477
+ 1.8651362984218078,1300,0.774725,0.8618,0.8893,0.774725,0.774725,0.2872666666666666,0.8618,0.17786000000000002,0.8893,0.774725,0.8200458333333279,0.8243714087301552,0.8478746522446807,0.8274045280642844
478
+ 2.0086083213773316,1400,0.77535,0.862225,0.88985,0.77535,0.77535,0.2874083333333333,0.862225,0.17797000000000002,0.88985,0.77535,0.8206862499999943,0.8250188888888853,0.8485109476637894,0.8280391911329523
479
+ 2.152080344332855,1500,0.77595,0.862725,0.89,0.77595,0.77595,0.287575,0.862725,0.17800000000000002,0.89,0.77595,0.8211174999999945,0.8255692956349164,0.8491781941787744,0.8285374285017794
480
+ 2.2955523672883786,1600,0.7767,0.863225,0.8902,0.7767,0.7767,0.2877416666666667,0.863225,0.17804000000000003,0.8902,0.7767,0.8216608333333275,0.8260917757936465,0.8496076724497309,0.8290711273746223
481
+ 2.4390243902439024,1700,0.777225,0.863475,0.891225,0.777225,0.777225,0.28782499999999994,0.863475,0.17824500000000001,0.891225,0.777225,0.8222145833333278,0.8265672817460276,0.8500954337191586,0.8295570818284077
482
+ 2.582496413199426,1800,0.77805,0.8639,0.8913,0.77805,0.77805,0.2879666666666666,0.8639,0.17826,0.8913,0.77805,0.8227304166666612,0.827145585317457,0.850644080505664,0.8301130910406217
483
+ 2.72596843615495,1900,0.778275,0.86375,0.89145,0.778275,0.778275,0.28791666666666665,0.86375,0.17829000000000003,0.89145,0.778275,0.8229012499999945,0.8273325496031723,0.8508192713538839,0.8302895273161849
484
+ 2.869440459110473,2000,0.77855,0.864725,0.892125,0.77855,0.77855,0.2882416666666666,0.864725,0.17842500000000003,0.892125,0.77855,0.8233962499999948,0.8278009722222194,0.8512696523745088,0.8307845745575778
485
+ 3.012912482065997,2100,0.779125,0.8652,0.892625,0.779125,0.779125,0.2884,0.8652,0.17852500000000004,0.892625,0.779125,0.8240054166666613,0.8283730357142838,0.8517705008752033,0.8313728112076162
486
+ 3.1563845050215207,2200,0.78015,0.865575,0.8929,0.78015,0.78015,0.288525,0.865575,0.17858000000000002,0.8929,0.78015,0.824600833333328,0.8290193452380925,0.8524261174442127,0.8319848134194368
487
+ 3.2998565279770444,2300,0.78065,0.86615,0.892975,0.78065,0.78065,0.2887166666666667,0.86615,0.178595,0.892975,0.78065,0.8250170833333279,0.8294674404761879,0.8528264662625589,0.8324378340650243
488
+ 3.443328550932568,2400,0.78115,0.8657,0.89325,0.78115,0.78115,0.2885666666666666,0.8657,0.17865000000000006,0.89325,0.78115,0.8253199999999944,0.829710376984124,0.8529684135498555,0.8327162907465414
489
+ 3.586800573888092,2500,0.782425,0.867,0.893425,0.782425,0.782425,0.2889999999999999,0.867,0.17868499999999998,0.893425,0.782425,0.8262479166666608,0.8307461111111079,0.8539413492468615,0.833704350343306
490
+ 3.7302725968436157,2600,0.785025,0.86815,0.8947,0.785025,0.785025,0.2893833333333333,0.86815,0.17894000000000002,0.8947,0.785025,0.8282516666666608,0.8326595138888846,0.8555509282711606,0.835614215395537
491
+ 3.8737446197991394,2700,0.78835,0.869425,0.8953,0.78835,0.78835,0.2898083333333333,0.869425,0.17906000000000002,0.8953,0.78835,0.8303924999999946,0.8348711805555517,0.8574317937994774,0.8378084083914124
492
+ 4.017216642754663,2800,0.789625,0.869975,0.896375,0.789625,0.789625,0.2899916666666666,0.869975,0.17927500000000002,0.896375,0.789625,0.8315524999999954,0.8359549603174564,0.8584329706633668,0.8388745159081235
493
+ 4.160688665710187,2900,0.7904,0.870225,0.8966,0.7904,0.7904,0.29007499999999997,0.870225,0.17932000000000003,0.8966,0.7904,0.8321333333333286,0.8365512301587261,0.8589791279951865,0.8394417496177948
494
+ 4.30416068866571,3000,0.791,0.87075,0.896925,0.791,0.791,0.29025,0.87075,0.17938500000000004,0.896925,0.791,0.8326162499999953,0.8370420039682492,0.8594495308982596,0.8399103794062596
495
+ 4.4476327116212335,3100,0.7909,0.8707,0.89715,0.7909,0.7909,0.2902333333333333,0.8707,0.17943,0.89715,0.7909,0.8325645833333286,0.8370284523809475,0.859566366154815,0.8398713536662339
496
+ 4.591104734576757,3200,0.791,0.8714,0.8973,0.791,0.791,0.29046666666666665,0.8714,0.17946000000000004,0.8973,0.791,0.8327691666666618,0.8372275099206298,0.8597537767314556,0.8400765245602044
497
+ 4.734576757532281,3300,0.7913,0.87155,0.89775,0.7913,0.7913,0.2905166666666666,0.87155,0.17955,0.89775,0.7913,0.8331179166666621,0.8375202876984075,0.8599949949176163,0.8403721064349383
498
+ 4.878048780487805,3400,0.791675,0.87175,0.898,0.791675,0.791675,0.2905833333333333,0.87175,0.1796,0.898,0.791675,0.8334462499999957,0.8378570833333279,0.8603339594646136,0.8406951566811449
499
+ 5.0215208034433285,3500,0.791975,0.872325,0.898175,0.791975,0.791975,0.2907749999999999,0.872325,0.17963500000000002,0.898175,0.791975,0.8337024999999952,0.8381229563492009,0.8605866702778303,0.8409624835010298
500
+ 5.164992826398852,3600,0.792875,0.872425,0.898625,0.792875,0.792875,0.2908083333333333,0.872425,0.17972500000000002,0.898625,0.792875,0.8343720833333287,0.8387630853174556,0.8611154131232631,0.8416006979310702
501
+ 5.308464849354376,3700,0.795075,0.873425,0.900075,0.795075,0.795075,0.29114166666666663,0.873425,0.18001500000000004,0.900075,0.795075,0.8360816666666618,0.8403687202380905,0.862515115440112,0.8431846364719889
502
+ 5.4519368723099,3800,0.795475,0.8736,0.9,0.795475,0.795475,0.2912,0.8736,0.18,0.9,0.795475,0.8363358333333288,0.8406914682539638,0.8628528597721815,0.8434900769937785
503
+ 5.5954088952654235,3900,0.79585,0.873925,0.9002,0.79585,0.79585,0.2913083333333333,0.873925,0.18004000000000003,0.9002,0.79585,0.8367083333333285,0.8410683035714231,0.8632126094245476,0.843847672580781
504
+ 5.738880918220947,4000,0.795925,0.87395,0.9003,0.795925,0.795925,0.2913166666666666,0.87395,0.18006,0.9003,0.795925,0.8368241666666619,0.8411607341269797,0.8632362511657556,0.8439728103363221
505
+ 5.882352941176471,4100,0.796225,0.873875,0.900375,0.796225,0.796225,0.2912916666666666,0.873875,0.180075,0.900375,0.796225,0.836939999999995,0.8412853472222173,0.8633881265956905,0.8440847141806891
506
+ 6.025824964131994,4200,0.7963,0.87395,0.900275,0.7963,0.7963,0.29131666666666667,0.87395,0.180055,0.900275,0.7963,0.8370295833333288,0.8413968650793606,0.8634661920710485,0.8442117176658458
507
+ 6.169296987087518,4300,0.796625,0.874375,0.9004,0.796625,0.796625,0.2914583333333333,0.874375,0.18008000000000002,0.9004,0.796625,0.8372995833333283,0.8416971329365032,0.863770166922786,0.8444891316790374
508
+ 6.312769010043041,4400,0.7964,0.874075,0.90055,0.7964,0.7964,0.29135833333333333,0.874075,0.18011000000000002,0.90055,0.7964,0.8371683333333285,0.8415165575396771,0.8636091649093294,0.8443133788931692
509
+ 6.456241032998565,4500,0.796425,0.8742,0.900575,0.796425,0.796425,0.29139999999999994,0.8742,0.18011500000000003,0.900575,0.796425,0.8372233333333284,0.8415766964285661,0.8636591002412395,0.8443833078274336
510
+ 6.599713055954089,4600,0.796725,0.874275,0.900675,0.796725,0.796725,0.291425,0.874275,0.18013500000000002,0.900675,0.796725,0.8373820833333288,0.8417151587301536,0.8637637204512882,0.844523830924215
511
+ 6.743185078909613,4700,0.796575,0.8743,0.90045,0.796575,0.796575,0.2914333333333333,0.8743,0.18009,0.90045,0.796575,0.8372449999999956,0.8416117361111064,0.8636712529829847,0.8444289062161769
512
+ 6.886657101865136,4800,0.79665,0.8745,0.900525,0.79665,0.79665,0.2915,0.8745,0.18010500000000002,0.900525,0.79665,0.8373320833333286,0.8416871428571374,0.863729044462657,0.8445069828327856
513
+ 7.03012912482066,4900,0.79655,0.87425,0.9004,0.79655,0.79655,0.2914166666666666,0.87425,0.18008000000000002,0.9004,0.79655,0.8372299999999956,0.8416038690476145,0.8636646643385855,0.844421583046012
514
+ 7.173601147776184,5000,0.7966,0.87425,0.900575,0.7966,0.7966,0.2914166666666666,0.87425,0.180115,0.900575,0.7966,0.8372962499999956,0.8416481150793601,0.8637140791780538,0.8444611118975183
final_metrics.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "val_cosine_accuracy@1": 0.82925,
3
- "val_cosine_accuracy@3": 0.903025,
4
- "val_cosine_accuracy@5": 0.931175,
5
- "val_cosine_precision@1": 0.82925,
6
- "val_cosine_precision@3": 0.3010083333333333,
7
- "val_cosine_precision@5": 0.186235,
8
- "val_cosine_recall@1": 0.82925,
9
- "val_cosine_recall@3": 0.903025,
10
- "val_cosine_recall@5": 0.931175,
11
- "val_cosine_ndcg@10": 0.8950131360828151,
12
- "val_cosine_mrr@1": 0.82925,
13
- "val_cosine_mrr@5": 0.8687345833333282,
14
- "val_cosine_mrr@10": 0.8731489384920591,
15
- "val_cosine_map@100": 0.8752091976044037
16
  }
 
1
  {
2
+ "val_cosine_accuracy@1": 0.7614,
3
+ "val_cosine_accuracy@3": 0.82615,
4
+ "val_cosine_accuracy@5": 0.850775,
5
+ "val_cosine_precision@1": 0.7614,
6
+ "val_cosine_precision@3": 0.2753833333333333,
7
+ "val_cosine_precision@5": 0.170155,
8
+ "val_cosine_recall@1": 0.7614,
9
+ "val_cosine_recall@3": 0.82615,
10
+ "val_cosine_recall@5": 0.850775,
11
+ "val_cosine_ndcg@10": 0.8201550154419872,
12
+ "val_cosine_mrr@1": 0.7614,
13
+ "val_cosine_mrr@5": 0.7960862499999959,
14
+ "val_cosine_mrr@10": 0.8003843253968239,
15
+ "val_cosine_map@100": 0.8038332983359062
16
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:92f71bc07a682285115d27c1ff96626310f991575725ca37984cfd4651023a66
3
  size 114011616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5db1674c675ac3fecdc903e40be0f70444de31052681e25137b3a8818fa28d06
3
  size 114011616
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:97dfabb4d1b0f4e7472f63fe7d96632afc7ca09080738d20201c498fcf74eb45
3
  size 6161
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d4c139a71ec4d4b0dfbafa31c4950656cd8f9c5e2ab0d688a8a2c2fdc8272c3
3
  size 6161