radoslavralev commited on
Commit
7bd707d
·
verified ·
1 Parent(s): d5e7923

Training in progress, step 14060

Browse files
Information-Retrieval_evaluation_val_results.csv CHANGED
@@ -1,3 +1,4 @@
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.908,0.9684,0.9834,0.908,0.908,0.3228,0.9684,0.19667999999999997,0.9834,0.908,0.9386633333333337,0.9400269841269848,0.9532296698470627,0.9404621256346036
3
  -1,-1,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402433333333332,0.9416250793650793,0.9545809774353143,0.9420576026548708
 
 
1
  epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
2
  -1,-1,0.908,0.9684,0.9834,0.908,0.908,0.3228,0.9684,0.19667999999999997,0.9834,0.908,0.9386633333333337,0.9400269841269848,0.9532296698470627,0.9404621256346036
3
  -1,-1,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402433333333332,0.9416250793650793,0.9545809774353143,0.9420576026548708
4
+ -1,-1,0.8281,0.9026,0.93105,0.8281,0.8281,0.3008666666666666,0.9026,0.18621000000000004,0.93105,0.8281,0.8677437499999962,0.8721381249999942,0.8942437004811851,0.874246358340888
README.md CHANGED
@@ -5,114 +5,38 @@ tags:
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
- - dataset_size:359997
9
  - loss:MultipleNegativesRankingLoss
10
  base_model: prajjwal1/bert-small
11
  widget:
12
- - source_sentence: When do you use Ms. or Mrs.? Is one for a married woman and one
13
- for one that's not married? Which one is for what?
14
  sentences:
15
- - When do you use Ms. or Mrs.? Is one for a married woman and one for one that's
16
- not married? Which one is for what?
17
- - Nations that do/does otherwise? Which one do I use?
18
- - What is the best way to make money on Quora?
19
- - source_sentence: Which ointment is applied to the face of UFC fighters at the commencement
20
- of a bout? What does it do?
21
  sentences:
22
- - Why don't bikes have a gear indicator?
23
- - Which ointment is applied to the face of UFC fighters at the commencement of a
24
- bout? What does it do?
25
- - How do I get the body of a UFC Fighter?
26
- - source_sentence: Do you love the life you live?
27
  sentences:
28
- - Which file formats are compatible with iTunes?
29
- - Do you love the life you're living?
30
- - What is the best way to find a person just using their phone by trying to track
31
- the other persons phone and get a location from it?
32
- - source_sentence: Can I do shoulder and triceps workout on same day? What other combinations
33
- like this can I do?
34
  sentences:
35
- - Can I do shoulder and triceps workout on same day? I can What other combinations
36
- like thisdo?
37
- - How can I save a Snapchat video that others posted?
38
- - Can I do shoulder and triceps workout on same day? What other combinations like
39
- this can I do?
40
- - source_sentence: I am a married woman and I'm in love with married man. what should
41
- I do?
42
  sentences:
43
- - How can I earn money easily online?
44
- - I am not a married woman and I 'm in love with married man . what should I do
45
- ?
46
- - I am a married woman and I'm in love with married man. what should I do?
47
  pipeline_tag: sentence-similarity
48
  library_name: sentence-transformers
49
- metrics:
50
- - cosine_accuracy@1
51
- - cosine_accuracy@3
52
- - cosine_accuracy@5
53
- - cosine_precision@1
54
- - cosine_precision@3
55
- - cosine_precision@5
56
- - cosine_recall@1
57
- - cosine_recall@3
58
- - cosine_recall@5
59
- - cosine_ndcg@10
60
- - cosine_mrr@1
61
- - cosine_mrr@5
62
- - cosine_mrr@10
63
- - cosine_map@100
64
- model-index:
65
- - name: SentenceTransformer based on prajjwal1/bert-small
66
- results:
67
- - task:
68
- type: information-retrieval
69
- name: Information Retrieval
70
- dataset:
71
- name: val
72
- type: val
73
- metrics:
74
- - type: cosine_accuracy@1
75
- value: 0.828025
76
- name: Cosine Accuracy@1
77
- - type: cosine_accuracy@3
78
- value: 0.9027
79
- name: Cosine Accuracy@3
80
- - type: cosine_accuracy@5
81
- value: 0.931025
82
- name: Cosine Accuracy@5
83
- - type: cosine_precision@1
84
- value: 0.828025
85
- name: Cosine Precision@1
86
- - type: cosine_precision@3
87
- value: 0.3008999999999999
88
- name: Cosine Precision@3
89
- - type: cosine_precision@5
90
- value: 0.186205
91
- name: Cosine Precision@5
92
- - type: cosine_recall@1
93
- value: 0.828025
94
- name: Cosine Recall@1
95
- - type: cosine_recall@3
96
- value: 0.9027
97
- name: Cosine Recall@3
98
- - type: cosine_recall@5
99
- value: 0.931025
100
- name: Cosine Recall@5
101
- - type: cosine_ndcg@10
102
- value: 0.8942284691055087
103
- name: Cosine Ndcg@10
104
- - type: cosine_mrr@1
105
- value: 0.828025
106
- name: Cosine Mrr@1
107
- - type: cosine_mrr@5
108
- value: 0.8677179166666629
109
- name: Cosine Mrr@5
110
- - type: cosine_mrr@10
111
- value: 0.8721162896825339
112
- name: Cosine Mrr@10
113
- - type: cosine_map@100
114
- value: 0.8742240723304836
115
- name: Cosine Map@100
116
  ---
117
 
118
  # SentenceTransformer based on prajjwal1/bert-small
@@ -161,12 +85,12 @@ Then you can load this model and run inference.
161
  from sentence_transformers import SentenceTransformer
162
 
163
  # Download from the 🤗 Hub
164
- model = SentenceTransformer("redis/model-b-structured")
165
  # Run inference
166
  sentences = [
167
- "I am a married woman and I'm in love with married man. what should I do?",
168
- "I am a married woman and I'm in love with married man. what should I do?",
169
- "I am not a married woman and I 'm in love with married man . what should I do ?",
170
  ]
171
  embeddings = model.encode(sentences)
172
  print(embeddings.shape)
@@ -175,9 +99,9 @@ print(embeddings.shape)
175
  # Get the similarity scores for the embeddings
176
  similarities = model.similarity(embeddings, embeddings)
177
  print(similarities)
178
- # tensor([[1.0000, 1.0000, 0.4050],
179
- # [1.0000, 1.0000, 0.4050],
180
- # [0.4050, 0.4050, 1.0000]])
181
  ```
182
 
183
  <!--
@@ -204,32 +128,6 @@ You can finetune this model on your own dataset.
204
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
205
  -->
206
 
207
- ## Evaluation
208
-
209
- ### Metrics
210
-
211
- #### Information Retrieval
212
-
213
- * Dataset: `val`
214
- * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
215
-
216
- | Metric | Value |
217
- |:-------------------|:-----------|
218
- | cosine_accuracy@1 | 0.828 |
219
- | cosine_accuracy@3 | 0.9027 |
220
- | cosine_accuracy@5 | 0.931 |
221
- | cosine_precision@1 | 0.828 |
222
- | cosine_precision@3 | 0.3009 |
223
- | cosine_precision@5 | 0.1862 |
224
- | cosine_recall@1 | 0.828 |
225
- | cosine_recall@3 | 0.9027 |
226
- | cosine_recall@5 | 0.931 |
227
- | **cosine_ndcg@10** | **0.8942** |
228
- | cosine_mrr@1 | 0.828 |
229
- | cosine_mrr@5 | 0.8677 |
230
- | cosine_mrr@10 | 0.8721 |
231
- | cosine_map@100 | 0.8742 |
232
-
233
  <!--
234
  ## Bias, Risks and Limitations
235
 
@@ -248,45 +146,19 @@ You can finetune this model on your own dataset.
248
 
249
  #### Unnamed Dataset
250
 
251
- * Size: 359,997 training samples
252
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
253
- * Approximate statistics based on the first 1000 samples:
254
- | | anchor | positive | negative |
255
- |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
256
- | type | string | string | string |
257
- | details | <ul><li>min: 4 tokens</li><li>mean: 15.46 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 15.52 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 16.63 tokens</li><li>max: 59 tokens</li></ul> |
258
- * Samples:
259
- | anchor | positive | negative |
260
- |:--------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
261
- | <code>Shall I upgrade my iPhone 5s to iOS 10 final version?</code> | <code>Should I upgrade an iPhone 5s to iOS 10?</code> | <code>Shall my iPhone 5s upgrade Ito iOS 10 final version?</code> |
262
- | <code>Is Donald Trump really going to be the president of United States?</code> | <code>Do you think Donald Trump could conceivably be the next President of the United States?</code> | <code>Is Donald Trump really going not to be the president of United States ?</code> |
263
- | <code>What are real tips to improve work life balance?</code> | <code>What are the best ways to create a work life balance?</code> | <code>How far is Miami from Fort Lauderdale?</code> |
264
- * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
265
- ```json
266
- {
267
- "scale": 20.0,
268
- "similarity_fct": "cos_sim",
269
- "gather_across_devices": false
270
- }
271
- ```
272
-
273
- ### Evaluation Dataset
274
-
275
- #### Unnamed Dataset
276
-
277
- * Size: 40,000 evaluation samples
278
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
279
  * Approximate statistics based on the first 1000 samples:
280
- | | anchor | positive | negative |
281
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
282
  | type | string | string | string |
283
- | details | <ul><li>min: 6 tokens</li><li>mean: 15.71 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 16.59 tokens</li><li>max: 77 tokens</li></ul> |
284
  * Samples:
285
- | anchor | positive | negative |
286
- |:------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------|
287
- | <code>Why were feathered dinosaur fossils only found in the last 20 years?</code> | <code>Why were feathered dinosaur fossils only found in the last 20 years?</code> | <code>Why are only few people aware that many dinosaurs had feathers?</code> |
288
- | <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>How much did Fox News and conservative leaning media networks stoke the anger that contributed to Donald Trump's popularity?</code> |
289
- | <code>How can guys last longer during sex?</code> | <code>How do I last longer in sex?</code> | <code>Why does economics require calculus?</code> |
290
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
291
  ```json
292
  {
@@ -299,49 +171,36 @@ You can finetune this model on your own dataset.
299
  ### Training Hyperparameters
300
  #### Non-Default Hyperparameters
301
 
302
- - `eval_strategy`: steps
303
- - `per_device_train_batch_size`: 256
304
- - `per_device_eval_batch_size`: 256
305
- - `learning_rate`: 2e-05
306
- - `weight_decay`: 0.001
307
- - `max_steps`: 14060
308
- - `warmup_ratio`: 0.1
309
  - `fp16`: True
310
- - `dataloader_drop_last`: True
311
- - `dataloader_num_workers`: 1
312
- - `dataloader_prefetch_factor`: 1
313
- - `load_best_model_at_end`: True
314
- - `optim`: adamw_torch
315
- - `ddp_find_unused_parameters`: False
316
- - `push_to_hub`: True
317
- - `hub_model_id`: redis/model-b-structured
318
- - `eval_on_start`: True
319
 
320
  #### All Hyperparameters
321
  <details><summary>Click to expand</summary>
322
 
323
  - `overwrite_output_dir`: False
324
  - `do_predict`: False
325
- - `eval_strategy`: steps
326
  - `prediction_loss_only`: True
327
- - `per_device_train_batch_size`: 256
328
- - `per_device_eval_batch_size`: 256
329
  - `per_gpu_train_batch_size`: None
330
  - `per_gpu_eval_batch_size`: None
331
  - `gradient_accumulation_steps`: 1
332
  - `eval_accumulation_steps`: None
333
  - `torch_empty_cache_steps`: None
334
- - `learning_rate`: 2e-05
335
- - `weight_decay`: 0.001
336
  - `adam_beta1`: 0.9
337
  - `adam_beta2`: 0.999
338
  - `adam_epsilon`: 1e-08
339
- - `max_grad_norm`: 1.0
340
- - `num_train_epochs`: 3.0
341
- - `max_steps`: 14060
342
  - `lr_scheduler_type`: linear
343
  - `lr_scheduler_kwargs`: {}
344
- - `warmup_ratio`: 0.1
345
  - `warmup_steps`: 0
346
  - `log_level`: passive
347
  - `log_level_replica`: warning
@@ -369,14 +228,14 @@ You can finetune this model on your own dataset.
369
  - `tpu_num_cores`: None
370
  - `tpu_metrics_debug`: False
371
  - `debug`: []
372
- - `dataloader_drop_last`: True
373
- - `dataloader_num_workers`: 1
374
- - `dataloader_prefetch_factor`: 1
375
  - `past_index`: -1
376
  - `disable_tqdm`: False
377
  - `remove_unused_columns`: True
378
  - `label_names`: None
379
- - `load_best_model_at_end`: True
380
  - `ignore_data_skip`: False
381
  - `fsdp`: []
382
  - `fsdp_min_num_params`: 0
@@ -386,23 +245,23 @@ You can finetune this model on your own dataset.
386
  - `parallelism_config`: None
387
  - `deepspeed`: None
388
  - `label_smoothing_factor`: 0.0
389
- - `optim`: adamw_torch
390
  - `optim_args`: None
391
  - `adafactor`: False
392
  - `group_by_length`: False
393
  - `length_column_name`: length
394
  - `project`: huggingface
395
  - `trackio_space_id`: trackio
396
- - `ddp_find_unused_parameters`: False
397
  - `ddp_bucket_cap_mb`: None
398
  - `ddp_broadcast_buffers`: False
399
  - `dataloader_pin_memory`: True
400
  - `dataloader_persistent_workers`: False
401
  - `skip_memory_metrics`: True
402
  - `use_legacy_prediction_loop`: False
403
- - `push_to_hub`: True
404
  - `resume_from_checkpoint`: None
405
- - `hub_model_id`: redis/model-b-structured
406
  - `hub_strategy`: every_save
407
  - `hub_private_repo`: None
408
  - `hub_always_push`: False
@@ -429,167 +288,32 @@ You can finetune this model on your own dataset.
429
  - `neftune_noise_alpha`: None
430
  - `optim_target_modules`: None
431
  - `batch_eval_metrics`: False
432
- - `eval_on_start`: True
433
  - `use_liger_kernel`: False
434
  - `liger_kernel_config`: None
435
  - `eval_use_gather_object`: False
436
  - `average_tokens_across_devices`: True
437
  - `prompts`: None
438
  - `batch_sampler`: batch_sampler
439
- - `multi_dataset_batch_sampler`: proportional
440
  - `router_mapping`: {}
441
  - `learning_rate_mapping`: {}
442
 
443
  </details>
444
 
445
  ### Training Logs
446
- <details><summary>Click to expand</summary>
447
-
448
- | Epoch | Step | Training Loss | Validation Loss | val_cosine_ndcg@10 |
449
- |:------:|:-----:|:-------------:|:---------------:|:------------------:|
450
- | 0 | 0 | - | 1.7418 | 0.7821 |
451
- | 0.0711 | 100 | 2.0777 | 0.7932 | 0.8130 |
452
- | 0.1422 | 200 | 0.7966 | 0.4005 | 0.8510 |
453
- | 0.2134 | 300 | 0.3991 | 0.2603 | 0.8615 |
454
- | 0.2845 | 400 | 0.3153 | 0.2051 | 0.8652 |
455
- | 0.3556 | 500 | 0.2593 | 0.1740 | 0.8681 |
456
- | 0.4267 | 600 | 0.2231 | 0.1568 | 0.8707 |
457
- | 0.4979 | 700 | 0.2017 | 0.1443 | 0.8727 |
458
- | 0.5690 | 800 | 0.1933 | 0.1322 | 0.8746 |
459
- | 0.6401 | 900 | 0.1818 | 0.1217 | 0.8755 |
460
- | 0.7112 | 1000 | 0.1714 | 0.1141 | 0.8769 |
461
- | 0.7824 | 1100 | 0.157 | 0.1060 | 0.8780 |
462
- | 0.8535 | 1200 | 0.1467 | 0.0998 | 0.8788 |
463
- | 0.9246 | 1300 | 0.1394 | 0.0937 | 0.8805 |
464
- | 0.9957 | 1400 | 0.1343 | 0.0910 | 0.8813 |
465
- | 1.0669 | 1500 | 0.1222 | 0.0853 | 0.8822 |
466
- | 1.1380 | 1600 | 0.1173 | 0.0820 | 0.8821 |
467
- | 1.2091 | 1700 | 0.1082 | 0.0797 | 0.8828 |
468
- | 1.2802 | 1800 | 0.1105 | 0.0777 | 0.8835 |
469
- | 1.3514 | 1900 | 0.1093 | 0.0734 | 0.8833 |
470
- | 1.4225 | 2000 | 0.1034 | 0.0744 | 0.8840 |
471
- | 1.4936 | 2100 | 0.1016 | 0.0713 | 0.8845 |
472
- | 1.5647 | 2200 | 0.0995 | 0.0699 | 0.8851 |
473
- | 1.6358 | 2300 | 0.0994 | 0.0679 | 0.8849 |
474
- | 1.7070 | 2400 | 0.1024 | 0.0667 | 0.8867 |
475
- | 1.7781 | 2500 | 0.0911 | 0.0658 | 0.8868 |
476
- | 1.8492 | 2600 | 0.0907 | 0.0640 | 0.8861 |
477
- | 1.9203 | 2700 | 0.0941 | 0.0632 | 0.8859 |
478
- | 1.9915 | 2800 | 0.093 | 0.0625 | 0.8870 |
479
- | 2.0626 | 2900 | 0.0814 | 0.0618 | 0.8875 |
480
- | 2.1337 | 3000 | 0.0811 | 0.0609 | 0.8868 |
481
- | 2.2048 | 3100 | 0.0773 | 0.0602 | 0.8880 |
482
- | 2.2760 | 3200 | 0.0813 | 0.0590 | 0.8873 |
483
- | 2.3471 | 3300 | 0.0806 | 0.0584 | 0.8876 |
484
- | 2.4182 | 3400 | 0.0765 | 0.0575 | 0.8882 |
485
- | 2.4893 | 3500 | 0.0774 | 0.0581 | 0.8889 |
486
- | 2.5605 | 3600 | 0.0761 | 0.0560 | 0.8883 |
487
- | 2.6316 | 3700 | 0.0735 | 0.0560 | 0.8886 |
488
- | 2.7027 | 3800 | 0.0711 | 0.0555 | 0.8891 |
489
- | 2.7738 | 3900 | 0.0747 | 0.0551 | 0.8889 |
490
- | 2.8450 | 4000 | 0.0731 | 0.0552 | 0.8897 |
491
- | 2.9161 | 4100 | 0.0708 | 0.0543 | 0.8898 |
492
- | 2.9872 | 4200 | 0.0778 | 0.0536 | 0.8901 |
493
- | 3.0583 | 4300 | 0.0697 | 0.0540 | 0.8893 |
494
- | 3.1294 | 4400 | 0.0668 | 0.0533 | 0.8900 |
495
- | 3.2006 | 4500 | 0.0679 | 0.0526 | 0.8893 |
496
- | 3.2717 | 4600 | 0.0652 | 0.0532 | 0.8902 |
497
- | 3.3428 | 4700 | 0.0673 | 0.0520 | 0.8899 |
498
- | 3.4139 | 4800 | 0.0625 | 0.0514 | 0.8903 |
499
- | 3.4851 | 4900 | 0.0669 | 0.0515 | 0.8912 |
500
- | 3.5562 | 5000 | 0.0641 | 0.0515 | 0.8915 |
501
- | 3.6273 | 5100 | 0.0637 | 0.0509 | 0.8909 |
502
- | 3.6984 | 5200 | 0.0635 | 0.0506 | 0.8908 |
503
- | 3.7696 | 5300 | 0.0606 | 0.0499 | 0.8915 |
504
- | 3.8407 | 5400 | 0.0633 | 0.0503 | 0.8917 |
505
- | 3.9118 | 5500 | 0.0656 | 0.0498 | 0.8913 |
506
- | 3.9829 | 5600 | 0.0658 | 0.0492 | 0.8916 |
507
- | 4.0541 | 5700 | 0.0606 | 0.0489 | 0.8917 |
508
- | 4.1252 | 5800 | 0.0585 | 0.0485 | 0.8914 |
509
- | 4.1963 | 5900 | 0.0613 | 0.0490 | 0.8914 |
510
- | 4.2674 | 6000 | 0.0568 | 0.0487 | 0.8909 |
511
- | 4.3385 | 6100 | 0.0576 | 0.0481 | 0.8918 |
512
- | 4.4097 | 6200 | 0.0603 | 0.0481 | 0.8915 |
513
- | 4.4808 | 6300 | 0.0569 | 0.0480 | 0.8918 |
514
- | 4.5519 | 6400 | 0.0553 | 0.0477 | 0.8921 |
515
- | 4.6230 | 6500 | 0.057 | 0.0472 | 0.8918 |
516
- | 4.6942 | 6600 | 0.0602 | 0.0472 | 0.8925 |
517
- | 4.7653 | 6700 | 0.0541 | 0.0468 | 0.8922 |
518
- | 4.8364 | 6800 | 0.0588 | 0.0468 | 0.8917 |
519
- | 4.9075 | 6900 | 0.0588 | 0.0471 | 0.8920 |
520
- | 4.9787 | 7000 | 0.0549 | 0.0469 | 0.8921 |
521
- | 5.0498 | 7100 | 0.0522 | 0.0466 | 0.8920 |
522
- | 5.1209 | 7200 | 0.0527 | 0.0462 | 0.8924 |
523
- | 5.1920 | 7300 | 0.0519 | 0.0461 | 0.8924 |
524
- | 5.2632 | 7400 | 0.0544 | 0.0459 | 0.8927 |
525
- | 5.3343 | 7500 | 0.0549 | 0.0456 | 0.8925 |
526
- | 5.4054 | 7600 | 0.0527 | 0.0460 | 0.8932 |
527
- | 5.4765 | 7700 | 0.0519 | 0.0453 | 0.8920 |
528
- | 5.5477 | 7800 | 0.0528 | 0.0455 | 0.8928 |
529
- | 5.6188 | 7900 | 0.0525 | 0.0451 | 0.8929 |
530
- | 5.6899 | 8000 | 0.0535 | 0.0454 | 0.8931 |
531
- | 5.7610 | 8100 | 0.0526 | 0.0452 | 0.8931 |
532
- | 5.8321 | 8200 | 0.0507 | 0.0454 | 0.8930 |
533
- | 5.9033 | 8300 | 0.0511 | 0.0451 | 0.8932 |
534
- | 5.9744 | 8400 | 0.0489 | 0.0451 | 0.8930 |
535
- | 6.0455 | 8500 | 0.0509 | 0.0451 | 0.8929 |
536
- | 6.1166 | 8600 | 0.0487 | 0.0447 | 0.8931 |
537
- | 6.1878 | 8700 | 0.0494 | 0.0449 | 0.8932 |
538
- | 6.2589 | 8800 | 0.0474 | 0.0444 | 0.8932 |
539
- | 6.3300 | 8900 | 0.049 | 0.0448 | 0.8934 |
540
- | 6.4011 | 9000 | 0.0492 | 0.0446 | 0.8934 |
541
- | 6.4723 | 9100 | 0.0493 | 0.0443 | 0.8931 |
542
- | 6.5434 | 9200 | 0.0517 | 0.0442 | 0.8931 |
543
- | 6.6145 | 9300 | 0.0502 | 0.0445 | 0.8938 |
544
- | 6.6856 | 9400 | 0.0501 | 0.0441 | 0.8935 |
545
- | 6.7568 | 9500 | 0.0484 | 0.0439 | 0.8935 |
546
- | 6.8279 | 9600 | 0.0472 | 0.0437 | 0.8935 |
547
- | 6.8990 | 9700 | 0.0484 | 0.0435 | 0.8936 |
548
- | 6.9701 | 9800 | 0.051 | 0.0433 | 0.8933 |
549
- | 7.0413 | 9900 | 0.0496 | 0.0435 | 0.8935 |
550
- | 7.1124 | 10000 | 0.0469 | 0.0434 | 0.8937 |
551
- | 7.1835 | 10100 | 0.0479 | 0.0432 | 0.8935 |
552
- | 7.2546 | 10200 | 0.0476 | 0.0430 | 0.8937 |
553
- | 7.3257 | 10300 | 0.0454 | 0.0431 | 0.8934 |
554
- | 7.3969 | 10400 | 0.0445 | 0.0430 | 0.8937 |
555
- | 7.4680 | 10500 | 0.0471 | 0.0427 | 0.8936 |
556
- | 7.5391 | 10600 | 0.0441 | 0.0429 | 0.8938 |
557
- | 7.6102 | 10700 | 0.046 | 0.0429 | 0.8932 |
558
- | 7.6814 | 10800 | 0.046 | 0.0428 | 0.8934 |
559
- | 7.7525 | 10900 | 0.049 | 0.0428 | 0.8938 |
560
- | 7.8236 | 11000 | 0.0476 | 0.0427 | 0.8939 |
561
- | 7.8947 | 11100 | 0.0468 | 0.0425 | 0.8938 |
562
- | 7.9659 | 11200 | 0.0465 | 0.0426 | 0.8940 |
563
- | 8.0370 | 11300 | 0.048 | 0.0428 | 0.8938 |
564
- | 8.1081 | 11400 | 0.0448 | 0.0425 | 0.8937 |
565
- | 8.1792 | 11500 | 0.0431 | 0.0424 | 0.8939 |
566
- | 8.2504 | 11600 | 0.0428 | 0.0424 | 0.8935 |
567
- | 8.3215 | 11700 | 0.046 | 0.0424 | 0.8937 |
568
- | 8.3926 | 11800 | 0.0471 | 0.0423 | 0.8938 |
569
- | 8.4637 | 11900 | 0.0466 | 0.0424 | 0.8943 |
570
- | 8.5349 | 12000 | 0.0431 | 0.0421 | 0.8941 |
571
- | 8.6060 | 12100 | 0.0462 | 0.0421 | 0.8938 |
572
- | 8.6771 | 12200 | 0.0425 | 0.0423 | 0.8941 |
573
- | 8.7482 | 12300 | 0.0455 | 0.0421 | 0.8941 |
574
- | 8.8193 | 12400 | 0.0445 | 0.0422 | 0.8940 |
575
- | 8.8905 | 12500 | 0.0455 | 0.0422 | 0.8943 |
576
- | 8.9616 | 12600 | 0.0448 | 0.0421 | 0.8941 |
577
- | 9.0327 | 12700 | 0.0462 | 0.0421 | 0.8940 |
578
- | 9.1038 | 12800 | 0.0429 | 0.0421 | 0.8939 |
579
- | 9.1750 | 12900 | 0.0452 | 0.0421 | 0.8942 |
580
- | 9.2461 | 13000 | 0.0439 | 0.0420 | 0.8943 |
581
- | 9.3172 | 13100 | 0.0472 | 0.0420 | 0.8942 |
582
- | 9.3883 | 13200 | 0.0447 | 0.0420 | 0.8943 |
583
- | 9.4595 | 13300 | 0.0426 | 0.0420 | 0.8942 |
584
- | 9.5306 | 13400 | 0.0445 | 0.0420 | 0.8942 |
585
- | 9.6017 | 13500 | 0.0436 | 0.0419 | 0.8942 |
586
- | 9.6728 | 13600 | 0.0445 | 0.0419 | 0.8943 |
587
- | 9.7440 | 13700 | 0.0477 | 0.0419 | 0.8943 |
588
- | 9.8151 | 13800 | 0.0439 | 0.0419 | 0.8942 |
589
- | 9.8862 | 13900 | 0.0438 | 0.0419 | 0.8942 |
590
- | 9.9573 | 14000 | 0.0468 | 0.0419 | 0.8942 |
591
 
592
- </details>
593
 
594
  ### Framework Versions
595
  - Python: 3.10.18
 
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
+ - dataset_size:100000
9
  - loss:MultipleNegativesRankingLoss
10
  base_model: prajjwal1/bert-small
11
  widget:
12
+ - source_sentence: How do I calculate IQ?
 
13
  sentences:
14
+ - What is the easiest way to know my IQ?
15
+ - How do I calculate not IQ ?
16
+ - What are some creative and innovative business ideas with less investment in India?
17
+ - source_sentence: How can I learn martial arts in my home?
 
 
18
  sentences:
19
+ - How can I learn martial arts by myself?
20
+ - What are the advantages and disadvantages of investing in gold?
21
+ - Can people see that I have looked at their pictures on instagram if I am not following
22
+ them?
23
+ - source_sentence: When Enterprise picks you up do you have to take them back?
24
  sentences:
25
+ - Are there any software Training institute in Tuticorin?
26
+ - When Enterprise picks you up do you have to take them back?
27
+ - When Enterprise picks you up do them have to take youback?
28
+ - source_sentence: What are some non-capital goods?
 
 
29
  sentences:
30
+ - What are capital goods?
31
+ - How is the value of [math]\pi[/math] calculated?
32
+ - What are some non-capital goods?
33
+ - source_sentence: What is the QuickBooks technical support phone number in New York?
 
 
 
34
  sentences:
35
+ - What caused the Great Depression?
36
+ - Can I apply for PR in Canada?
37
+ - Which is the best QuickBooks Hosting Support Number in New York?
 
38
  pipeline_tag: sentence-similarity
39
  library_name: sentence-transformers
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ---
41
 
42
  # SentenceTransformer based on prajjwal1/bert-small
 
85
  from sentence_transformers import SentenceTransformer
86
 
87
  # Download from the 🤗 Hub
88
+ model = SentenceTransformer("sentence_transformers_model_id")
89
  # Run inference
90
  sentences = [
91
+ 'What is the QuickBooks technical support phone number in New York?',
92
+ 'Which is the best QuickBooks Hosting Support Number in New York?',
93
+ 'Can I apply for PR in Canada?',
94
  ]
95
  embeddings = model.encode(sentences)
96
  print(embeddings.shape)
 
99
  # Get the similarity scores for the embeddings
100
  similarities = model.similarity(embeddings, embeddings)
101
  print(similarities)
102
+ # tensor([[1.0000, 0.8563, 0.0594],
103
+ # [0.8563, 1.0000, 0.1245],
104
+ # [0.0594, 0.1245, 1.0000]])
105
  ```
106
 
107
  <!--
 
128
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
129
  -->
130
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  <!--
132
  ## Bias, Risks and Limitations
133
 
 
146
 
147
  #### Unnamed Dataset
148
 
149
+ * Size: 100,000 training samples
150
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
151
  * Approximate statistics based on the first 1000 samples:
152
+ | | sentence_0 | sentence_1 | sentence_2 |
153
  |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
154
  | type | string | string | string |
155
+ | details | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.68 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 16.37 tokens</li><li>max: 67 tokens</li></ul> |
156
  * Samples:
157
+ | sentence_0 | sentence_1 | sentence_2 |
158
+ |:-----------------------------------------------------------------|:-----------------------------------------------------------------|:----------------------------------------------------------------------------------|
159
+ | <code>Is masturbating bad for boys?</code> | <code>Is masturbating bad for boys?</code> | <code>How harmful or unhealthy is masturbation?</code> |
160
+ | <code>Does a train engine move in reverse?</code> | <code>Does a train engine move in reverse?</code> | <code>Time moves forward, not in reverse. Doesn't that make time a vector?</code> |
161
+ | <code>What is the most badass thing anyone has ever done?</code> | <code>What is the most badass thing anyone has ever done?</code> | <code>anyone is the most badass thing Whathas ever done?</code> |
162
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
163
  ```json
164
  {
 
171
  ### Training Hyperparameters
172
  #### Non-Default Hyperparameters
173
 
174
+ - `per_device_train_batch_size`: 64
175
+ - `per_device_eval_batch_size`: 64
 
 
 
 
 
176
  - `fp16`: True
177
+ - `multi_dataset_batch_sampler`: round_robin
 
 
 
 
 
 
 
 
178
 
179
  #### All Hyperparameters
180
  <details><summary>Click to expand</summary>
181
 
182
  - `overwrite_output_dir`: False
183
  - `do_predict`: False
184
+ - `eval_strategy`: no
185
  - `prediction_loss_only`: True
186
+ - `per_device_train_batch_size`: 64
187
+ - `per_device_eval_batch_size`: 64
188
  - `per_gpu_train_batch_size`: None
189
  - `per_gpu_eval_batch_size`: None
190
  - `gradient_accumulation_steps`: 1
191
  - `eval_accumulation_steps`: None
192
  - `torch_empty_cache_steps`: None
193
+ - `learning_rate`: 5e-05
194
+ - `weight_decay`: 0.0
195
  - `adam_beta1`: 0.9
196
  - `adam_beta2`: 0.999
197
  - `adam_epsilon`: 1e-08
198
+ - `max_grad_norm`: 1
199
+ - `num_train_epochs`: 3
200
+ - `max_steps`: -1
201
  - `lr_scheduler_type`: linear
202
  - `lr_scheduler_kwargs`: {}
203
+ - `warmup_ratio`: 0.0
204
  - `warmup_steps`: 0
205
  - `log_level`: passive
206
  - `log_level_replica`: warning
 
228
  - `tpu_num_cores`: None
229
  - `tpu_metrics_debug`: False
230
  - `debug`: []
231
+ - `dataloader_drop_last`: False
232
+ - `dataloader_num_workers`: 0
233
+ - `dataloader_prefetch_factor`: None
234
  - `past_index`: -1
235
  - `disable_tqdm`: False
236
  - `remove_unused_columns`: True
237
  - `label_names`: None
238
+ - `load_best_model_at_end`: False
239
  - `ignore_data_skip`: False
240
  - `fsdp`: []
241
  - `fsdp_min_num_params`: 0
 
245
  - `parallelism_config`: None
246
  - `deepspeed`: None
247
  - `label_smoothing_factor`: 0.0
248
+ - `optim`: adamw_torch_fused
249
  - `optim_args`: None
250
  - `adafactor`: False
251
  - `group_by_length`: False
252
  - `length_column_name`: length
253
  - `project`: huggingface
254
  - `trackio_space_id`: trackio
255
+ - `ddp_find_unused_parameters`: None
256
  - `ddp_bucket_cap_mb`: None
257
  - `ddp_broadcast_buffers`: False
258
  - `dataloader_pin_memory`: True
259
  - `dataloader_persistent_workers`: False
260
  - `skip_memory_metrics`: True
261
  - `use_legacy_prediction_loop`: False
262
+ - `push_to_hub`: False
263
  - `resume_from_checkpoint`: None
264
+ - `hub_model_id`: None
265
  - `hub_strategy`: every_save
266
  - `hub_private_repo`: None
267
  - `hub_always_push`: False
 
288
  - `neftune_noise_alpha`: None
289
  - `optim_target_modules`: None
290
  - `batch_eval_metrics`: False
291
+ - `eval_on_start`: False
292
  - `use_liger_kernel`: False
293
  - `liger_kernel_config`: None
294
  - `eval_use_gather_object`: False
295
  - `average_tokens_across_devices`: True
296
  - `prompts`: None
297
  - `batch_sampler`: batch_sampler
298
+ - `multi_dataset_batch_sampler`: round_robin
299
  - `router_mapping`: {}
300
  - `learning_rate_mapping`: {}
301
 
302
  </details>
303
 
304
  ### Training Logs
305
+ | Epoch | Step | Training Loss |
306
+ |:------:|:----:|:-------------:|
307
+ | 0.3199 | 500 | 0.4294 |
308
+ | 0.6398 | 1000 | 0.1268 |
309
+ | 0.9597 | 1500 | 0.1 |
310
+ | 1.2796 | 2000 | 0.0792 |
311
+ | 1.5995 | 2500 | 0.0706 |
312
+ | 1.9194 | 3000 | 0.0687 |
313
+ | 2.2393 | 3500 | 0.0584 |
314
+ | 2.5592 | 4000 | 0.057 |
315
+ | 2.8791 | 4500 | 0.0581 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
316
 
 
317
 
318
  ### Framework Versions
319
  - Python: 3.10.18
eval/Information-Retrieval_evaluation_val_results.csv CHANGED
@@ -179,3 +179,144 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
179
  9.815078236130867,13800,0.827975,0.90265,0.931,0.827975,0.827975,0.3008833333333333,0.90265,0.18620000000000003,0.931,0.827975,0.8676762499999963,0.8720721230158671,0.894178246968981,0.8741864677767469
180
  9.88620199146515,13900,0.828025,0.9026,0.9311,0.828025,0.828025,0.3008666666666666,0.9026,0.18622000000000002,0.9311,0.828025,0.867727916666663,0.8721113690476133,0.8942189714318465,0.8742215047241502
181
  9.95732574679943,14000,0.828025,0.9027,0.931025,0.828025,0.828025,0.3008999999999999,0.9027,0.186205,0.931025,0.828025,0.8677179166666629,0.8721162896825339,0.8942284691055087,0.8742240723304836
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
179
  9.815078236130867,13800,0.827975,0.90265,0.931,0.827975,0.827975,0.3008833333333333,0.90265,0.18620000000000003,0.931,0.827975,0.8676762499999963,0.8720721230158671,0.894178246968981,0.8741864677767469
180
  9.88620199146515,13900,0.828025,0.9026,0.9311,0.828025,0.828025,0.3008666666666666,0.9026,0.18622000000000002,0.9311,0.828025,0.867727916666663,0.8721113690476133,0.8942189714318465,0.8742215047241502
181
  9.95732574679943,14000,0.828025,0.9027,0.931025,0.828025,0.828025,0.3008999999999999,0.9027,0.186205,0.931025,0.828025,0.8677179166666629,0.8721162896825339,0.8942284691055087,0.8742240723304836
182
+ 0,0,0.660625,0.78855,0.816875,0.660625,0.660625,0.26285,0.78855,0.163375,0.816875,0.660625,0.7265508333333278,0.7310880059523766,0.7604347668434622,0.7344176635988929
183
+ 0.07112375533428165,100,0.65785,0.8221,0.855375,0.65785,0.65785,0.2740333333333333,0.8221,0.171075,0.855375,0.65785,0.7426349999999933,0.747496111111108,0.7830189713908015,0.7509685938326416
184
+ 0.1422475106685633,200,0.76665,0.86435,0.8921,0.76665,0.76665,0.28811666666666663,0.86435,0.17842,0.8921,0.76665,0.8175337499999934,0.8216855654761861,0.8463620941294383,0.8244802043057975
185
+ 0.21337126600284495,300,0.790825,0.87115,0.89655,0.790825,0.790825,0.2903833333333333,0.87115,0.17931,0.89655,0.790825,0.832905833333328,0.8370798809523765,0.8589576487020009,0.8396630602475953
186
+ 0.2844950213371266,400,0.7981,0.8745,0.899575,0.7981,0.7981,0.2915,0.8745,0.17991500000000002,0.899575,0.7981,0.838206249999995,0.8422020833333296,0.8632435713976536,0.8448211843344123
187
+ 0.35561877667140823,500,0.802025,0.878225,0.902325,0.802025,0.802025,0.2927416666666666,0.878225,0.18046500000000004,0.902325,0.802025,0.8418195833333277,0.845981617063489,0.8670208129259122,0.8485604470631575
188
+ 0.4267425320056899,600,0.805275,0.880225,0.90495,0.805275,0.805275,0.2934083333333333,0.880225,0.18099000000000004,0.90495,0.805275,0.8446729166666621,0.848717420634917,0.8695488582496811,0.8513143530847598
189
+ 0.49786628733997157,700,0.807725,0.88225,0.906575,0.807725,0.807725,0.2940833333333333,0.88225,0.181315,0.906575,0.807725,0.84687833333333,0.8510880654761879,0.8720327027175575,0.8536533089445467
190
+ 0.5689900426742532,800,0.809625,0.881675,0.9072,0.809625,0.809625,0.29389166666666666,0.881675,0.18144000000000005,0.9072,0.809625,0.8480374999999966,0.8522258333333291,0.872984794383946,0.854857136239104
191
+ 0.6401137980085349,900,0.810425,0.883875,0.908725,0.810425,0.810425,0.29462499999999997,0.883875,0.18174500000000002,0.908725,0.810425,0.8493383333333296,0.8535709623015844,0.8744516835300677,0.8561838754847994
192
+ 0.7112375533428165,1000,0.812775,0.885475,0.910425,0.812775,0.812775,0.2951583333333333,0.885475,0.18208500000000002,0.910425,0.812775,0.8513895833333308,0.8556871527777754,0.8765193710353214,0.858272041629315
193
+ 0.7823613086770982,1100,0.813225,0.8863,0.9114,0.813225,0.813225,0.2954333333333333,0.8863,0.18228000000000003,0.9114,0.813225,0.8519433333333298,0.8562210515872983,0.8771499843059728,0.8588020045901827
194
+ 0.8534850640113798,1200,0.81475,0.88835,0.91365,0.81475,0.81475,0.2961166666666666,0.88835,0.18273000000000003,0.91365,0.81475,0.85359958333333,0.8578052579365052,0.8787946145827298,0.8603464162431504
195
+ 0.9246088193456614,1300,0.815125,0.88775,0.913825,0.815125,0.815125,0.29591666666666666,0.88775,0.182765,0.913825,0.815125,0.853832916666663,0.8579933333333299,0.8788853951072472,0.8605746130331765
196
+ 0.9957325746799431,1400,0.8174,0.88975,0.916,0.8174,0.8174,0.2965833333333333,0.88975,0.18320000000000003,0.916,0.8174,0.8559345833333298,0.8600880952380925,0.8810021639106901,0.8625984545198154
197
+ 1.0668563300142249,1500,0.816325,0.8888,0.914975,0.816325,0.816325,0.2962666666666666,0.8888,0.18299500000000002,0.914975,0.816325,0.854826249999997,0.8589861408730142,0.8799266262551008,0.8615620862321841
198
+ 1.1379800853485065,1600,0.818025,0.890975,0.91655,0.818025,0.818025,0.2969916666666667,0.890975,0.18331,0.91655,0.818025,0.8567195833333306,0.8609042559523788,0.8818126370189291,0.8634202065126004
199
+ 1.209103840682788,1700,0.8187,0.8904,0.916775,0.8187,0.8187,0.2968,0.8904,0.18335500000000002,0.916775,0.8187,0.8569770833333301,0.861196865079362,0.8821120071786194,0.863703746171397
200
+ 1.2802275960170697,1800,0.819325,0.890925,0.91765,0.819325,0.819325,0.296975,0.890925,0.18353000000000005,0.91765,0.819325,0.8575979166666624,0.8617786904761866,0.8827203211699732,0.8642523489926849
201
+ 1.3513513513513513,1900,0.819575,0.89235,0.918925,0.819575,0.819575,0.29744999999999994,0.89235,0.183785,0.918925,0.819575,0.8583795833333298,0.8625744345238064,0.8836298047121208,0.8650300188501633
202
+ 1.422475106685633,2000,0.82045,0.892125,0.91885,0.82045,0.82045,0.297375,0.892125,0.18377000000000002,0.91885,0.82045,0.8587412499999957,0.8629089980158691,0.8838312959674957,0.8653673262499637
203
+ 1.4935988620199145,2100,0.8213,0.891925,0.91905,0.8213,0.8213,0.29730833333333323,0.891925,0.18381000000000003,0.91905,0.8213,0.8593224999999954,0.8635408134920594,0.8844500604469038,0.8659837345007254
204
+ 1.5647226173541964,2200,0.82035,0.89215,0.919425,0.82035,0.82035,0.2973833333333333,0.89215,0.18388500000000002,0.919425,0.82035,0.8589304166666636,0.8631554464285682,0.884277860760647,0.8656085699448136
205
+ 1.635846372688478,2300,0.820625,0.89235,0.91915,0.820625,0.820625,0.29744999999999994,0.89235,0.18383000000000002,0.91915,0.820625,0.8590399999999964,0.8633189781745999,0.8844012972181756,0.8657824496152007
206
+ 1.7069701280227596,2400,0.8213,0.893175,0.920525,0.8213,0.8213,0.2977249999999999,0.893175,0.18410500000000005,0.920525,0.8213,0.8596737499999956,0.8639137103174555,0.8851440135349451,0.8663259037610367
207
+ 1.7780938833570412,2500,0.82315,0.896175,0.92235,0.82315,0.82315,0.29872499999999996,0.896175,0.18447000000000002,0.92235,0.82315,0.8618333333333296,0.866077976190474,0.8872320902095131,0.8684474128011467
208
+ 1.8492176386913228,2600,0.82145,0.89405,0.920575,0.82145,0.82145,0.2980166666666666,0.89405,0.18411500000000006,0.920575,0.82145,0.8601562499999962,0.8644306547619017,0.885607192551569,0.8668801137114076
209
+ 1.9203413940256047,2700,0.823,0.894925,0.921975,0.823,0.823,0.2983083333333333,0.894925,0.1843950000000001,0.921975,0.823,0.8615154166666628,0.865800605158727,0.8869919207490155,0.868200699005933
210
+ 1.991465149359886,2800,0.822525,0.8952,0.921975,0.822525,0.822525,0.2984,0.8952,0.18439500000000003,0.921975,0.822525,0.8612554166666628,0.8655788888888847,0.8868942891876812,0.8679661063701757
211
+ 2.062588904694168,2900,0.82265,0.8948,0.921475,0.82265,0.82265,0.2982666666666666,0.8948,0.18429500000000001,0.921475,0.82265,0.8611083333333301,0.8654150793650768,0.8866298372806077,0.8678082206661994
212
+ 2.1337126600284497,3000,0.8235,0.895625,0.92215,0.8235,0.8235,0.2985416666666666,0.895625,0.18442999999999998,0.92215,0.8235,0.8619112499999959,0.8662926984126956,0.8875291073406817,0.868677931078467
213
+ 2.204836415362731,3100,0.822925,0.89535,0.9217,0.822925,0.822925,0.29844999999999994,0.89535,0.18434000000000003,0.9217,0.822925,0.8613891666666638,0.8658161507936486,0.887148424379975,0.8681836052644616
214
+ 2.275960170697013,3200,0.824975,0.89685,0.922975,0.824975,0.824975,0.29894999999999994,0.89685,0.18459500000000004,0.922975,0.824975,0.8630554166666626,0.8674809722222193,0.8887076656677213,0.8697873502580414
215
+ 2.3470839260312943,3300,0.824325,0.89665,0.922425,0.824325,0.824325,0.2988833333333333,0.89665,0.18448500000000004,0.922425,0.824325,0.862727916666663,0.8672325099206315,0.8884958531410477,0.8695684400453714
216
+ 2.418207681365576,3400,0.824275,0.896175,0.922325,0.824275,0.824275,0.2987249999999999,0.896175,0.18446500000000005,0.922325,0.824275,0.8625166666666632,0.8669797718253927,0.8882555705244735,0.8693215056646579
217
+ 2.4893314366998576,3500,0.824075,0.8962,0.922325,0.824075,0.824075,0.2987333333333333,0.8962,0.18446500000000005,0.922325,0.824075,0.8624966666666627,0.8669934424603137,0.8883243690177098,0.8693086070714775
218
+ 2.5604551920341394,3600,0.8248,0.896775,0.923875,0.8248,0.8248,0.2989249999999999,0.896775,0.18477500000000002,0.923875,0.8248,0.8632529166666633,0.8675360714285681,0.8887711826852934,0.8698778746887647
219
+ 2.6315789473684212,3700,0.825075,0.897175,0.923625,0.825075,0.825075,0.2990583333333333,0.897175,0.18472500000000003,0.923625,0.825075,0.8634116666666629,0.8677288293650759,0.8889119603422132,0.8701022749841466
220
+ 2.7027027027027026,3800,0.824525,0.89705,0.923225,0.824525,0.824525,0.2990166666666666,0.89705,0.18464500000000003,0.923225,0.824525,0.8629341666666629,0.867378343253966,0.8887312117027028,0.8697003139888475
221
+ 2.7738264580369845,3900,0.825025,0.897425,0.924375,0.825025,0.825025,0.2991416666666666,0.897425,0.184875,0.924375,0.825025,0.8637245833333296,0.8680381746031713,0.8893130999822193,0.8703772079610197
222
+ 2.844950213371266,4000,0.823725,0.89655,0.92275,0.823725,0.823725,0.29885,0.89655,0.18455000000000002,0.92275,0.823725,0.8624545833333299,0.8669824801587267,0.8884952313868706,0.8693080141848417
223
+ 2.9160739687055477,4100,0.8248,0.89735,0.92425,0.8248,0.8248,0.2991166666666666,0.89735,0.18485000000000001,0.92425,0.8248,0.8635024999999963,0.8679394246031711,0.8894314662912053,0.8702279980402775
224
+ 2.987197724039829,4200,0.8259,0.897925,0.924975,0.8259,0.8259,0.2993083333333333,0.897925,0.18499500000000002,0.924975,0.8259,0.8644899999999974,0.8688425595238083,0.8901161301402605,0.8711507401738996
225
+ 3.058321479374111,4300,0.824725,0.8973,0.92515,0.824725,0.824725,0.2991,0.8973,0.18503000000000003,0.92515,0.824725,0.8636933333333314,0.868004523809522,0.8894692161504967,0.8703365016119559
226
+ 3.1294452347083928,4400,0.8249,0.897525,0.924325,0.8249,0.8249,0.2991749999999999,0.897525,0.184865,0.924325,0.8249,0.8636020833333312,0.868028988095237,0.8894864613534786,0.8703489640976624
227
+ 3.200568990042674,4500,0.8258,0.8986,0.92565,0.8258,0.8258,0.2995333333333333,0.8986,0.18513000000000004,0.92565,0.8258,0.8646745833333307,0.8689872420634889,0.890385534258935,0.8712701053994162
228
+ 3.271692745376956,4600,0.8257,0.8983,0.92535,0.8257,0.8257,0.29943333333333333,0.8983,0.18507000000000004,0.92535,0.8257,0.8644454166666634,0.8688889682539637,0.8904275201569494,0.8711486096369192
229
+ 3.3428165007112374,4700,0.826325,0.89855,0.92535,0.826325,0.826325,0.29951666666666665,0.89855,0.18507000000000004,0.92535,0.826325,0.8648566666666646,0.8692932539682527,0.8906216945646305,0.8716049278778036
230
+ 3.413940256045519,4800,0.8256,0.898475,0.92485,0.8256,0.8256,0.2994916666666666,0.898475,0.18497000000000002,0.92485,0.8256,0.8642804166666638,0.8687215277777753,0.890143837174269,0.8710299777162378
231
+ 3.485064011379801,4900,0.8265,0.89945,0.9264,0.8265,0.8265,0.2998166666666666,0.89945,0.18528000000000003,0.9264,0.8265,0.8654299999999965,0.8697809325396798,0.8911726290639512,0.8720622159211568
232
+ 3.5561877667140824,5000,0.8265,0.89915,0.926275,0.8265,0.8265,0.29971666666666663,0.89915,0.18525500000000003,0.926275,0.8265,0.8653016666666635,0.8696054067460289,0.8909265121877501,0.8719266555079046
233
+ 3.6273115220483643,5100,0.82565,0.899175,0.9258,0.82565,0.82565,0.29972499999999996,0.899175,0.18516000000000005,0.9258,0.82565,0.8647320833333304,0.8691491269841237,0.8906552891143423,0.8714300727214882
234
+ 3.6984352773826457,5200,0.8256,0.8991,0.9263,0.8256,0.8256,0.2996999999999999,0.8991,0.18526000000000004,0.9263,0.8256,0.8647333333333299,0.8690647321428537,0.8906112568228791,0.8713395023759158
235
+ 3.7695590327169275,5300,0.8267,0.90005,0.927175,0.8267,0.8267,0.30001666666666665,0.90005,0.18543500000000004,0.927175,0.8267,0.8658445833333303,0.870133650793649,0.8915589357540593,0.8724021579521188
236
+ 3.8406827880512093,5400,0.82635,0.89965,0.92665,0.82635,0.82635,0.2998833333333333,0.89965,0.18533000000000002,0.92665,0.82635,0.8654087499999968,0.8697750694444412,0.8913026933116265,0.8720366312771636
237
+ 3.9118065433854907,5500,0.826675,0.900275,0.927025,0.826675,0.826675,0.30009166666666665,0.900275,0.18540500000000001,0.927025,0.826675,0.8657945833333306,0.8701155952380921,0.891577510224056,0.8723879285250398
238
+ 3.9829302987197726,5600,0.82665,0.89945,0.9268,0.82665,0.82665,0.2998166666666666,0.89945,0.18536000000000002,0.9268,0.82665,0.8655083333333307,0.8698893154761871,0.8914393474383101,0.8721425772539715
239
+ 4.054054054054054,5700,0.826,0.898975,0.926125,0.826,0.826,0.29965833333333325,0.898975,0.185225,0.926125,0.826,0.8650079166666638,0.8693917162698382,0.8908775275994638,0.8717032813673281
240
+ 4.125177809388336,5800,0.82755,0.900225,0.92755,0.82755,0.82755,0.300075,0.900225,0.18551000000000004,0.92755,0.82755,0.8664683333333296,0.8708438392857104,0.8923375459234719,0.8730607380382754
241
+ 4.196301564722617,5900,0.82705,0.90065,0.9282,0.82705,0.82705,0.30021666666666663,0.90065,0.18564,0.9282,0.82705,0.8664145833333298,0.8706986011904715,0.8922585677258182,0.8729291403486316
242
+ 4.2674253200568995,6000,0.827125,0.899675,0.92705,0.827125,0.827125,0.2998916666666666,0.899675,0.18541000000000005,0.92705,0.827125,0.8659537499999962,0.8703908829365028,0.89196813784815,0.8726083427829648
243
+ 4.338549075391181,6100,0.82685,0.90015,0.927425,0.82685,0.82685,0.30004999999999993,0.90015,0.185485,0.927425,0.82685,0.8659616666666632,0.8703435515872973,0.891964842926896,0.8725465070798555
244
+ 4.409672830725462,6200,0.826875,0.900275,0.927075,0.826875,0.826875,0.3000916666666666,0.900275,0.185415,0.927075,0.826875,0.8658904166666626,0.8703097519841209,0.8918968662133886,0.8725315660787611
245
+ 4.480796586059744,6300,0.82725,0.9012,0.927925,0.82725,0.82725,0.30039999999999994,0.9012,0.185585,0.927925,0.82725,0.8665112499999957,0.8708849999999949,0.8924496641197642,0.8731048780889112
246
+ 4.551920341394026,6400,0.82685,0.900775,0.927275,0.82685,0.82685,0.3002583333333333,0.900775,0.185455,0.927275,0.82685,0.8660962499999955,0.8705262896825335,0.8921092548063507,0.8727407525006975
247
+ 4.623044096728307,6500,0.827575,0.901075,0.928275,0.827575,0.827575,0.3003583333333333,0.901075,0.18565500000000001,0.928275,0.827575,0.8668083333333294,0.8711607341269799,0.8926932674759781,0.8733821469708742
248
+ 4.694167852062589,6600,0.826725,0.90005,0.927225,0.826725,0.826725,0.3000166666666666,0.90005,0.18544500000000003,0.927225,0.826725,0.8657887499999959,0.8702033432539638,0.891834835753562,0.8724344252761722
249
+ 4.76529160739687,6700,0.827225,0.900125,0.927525,0.827225,0.827225,0.3000416666666666,0.900125,0.18550500000000003,0.927525,0.827225,0.8662174999999963,0.8706892460317409,0.8923377513084956,0.8729005329969197
250
+ 4.836415362731152,6800,0.827675,0.90055,0.9274,0.827675,0.827675,0.30018333333333325,0.90055,0.18548000000000003,0.9274,0.827675,0.8665483333333288,0.8710253075396777,0.8925563711068755,0.8732563983996522
251
+ 4.907539118065434,6900,0.827575,0.9004,0.92825,0.827575,0.827575,0.30013333333333325,0.9004,0.18565,0.92825,0.827575,0.8666908333333294,0.871095198412693,0.8927540200153565,0.873288654653562
252
+ 4.978662873399715,7000,0.82795,0.90095,0.9285,0.82795,0.82795,0.3003166666666666,0.90095,0.1857,0.9285,0.82795,0.8669474999999964,0.8713514285714237,0.8929598750593017,0.8735464299621104
253
+ 5.049786628733997,7100,0.82815,0.90145,0.928925,0.82815,0.82815,0.30048333333333327,0.90145,0.18578500000000003,0.928925,0.82815,0.8673516666666624,0.8717274305555508,0.893290068636731,0.8739296073428533
254
+ 5.120910384068279,7200,0.82795,0.90095,0.9286,0.82795,0.82795,0.3003166666666666,0.90095,0.18572000000000002,0.9286,0.82795,0.8670533333333291,0.8714805257936449,0.8931097492872178,0.8736657314635108
255
+ 5.19203413940256,7300,0.82755,0.9014,0.92895,0.82755,0.82755,0.3004666666666666,0.9014,0.18579,0.92895,0.82755,0.8670079166666628,0.8714806150793606,0.8932898007413307,0.8736056338472569
256
+ 5.2631578947368425,7400,0.827525,0.9012,0.9288,0.827525,0.827525,0.30039999999999994,0.9012,0.18576,0.9288,0.827525,0.8667895833333292,0.8712067261904715,0.8929619217995062,0.8733752014385081
257
+ 5.334281650071124,7500,0.8274,0.900825,0.928425,0.8274,0.8274,0.300275,0.900825,0.18568500000000002,0.928425,0.8274,0.8666358333333299,0.871154603174598,0.8930023691933456,0.8732816124281766
258
+ 5.405405405405405,7600,0.8278,0.90085,0.92875,0.8278,0.8278,0.3002833333333333,0.90085,0.18575000000000003,0.92875,0.8278,0.8669041666666626,0.8714118353174557,0.893248046324686,0.8735481954345135
259
+ 5.476529160739687,7700,0.8275,0.9006,0.929025,0.8275,0.8275,0.3001999999999999,0.9006,0.18580500000000003,0.929025,0.8275,0.8667345833333302,0.8711018650793607,0.8928943868112501,0.873268752899655
260
+ 5.547652916073969,7800,0.8282,0.90175,0.929125,0.8282,0.8282,0.3005833333333333,0.90175,0.185825,0.929125,0.8282,0.8674249999999957,0.8717897222222175,0.8934076888371761,0.8739759838230828
261
+ 5.61877667140825,7900,0.828125,0.90145,0.9286,0.828125,0.828125,0.3004833333333333,0.90145,0.18572000000000002,0.9286,0.828125,0.8671795833333292,0.8716491468253924,0.8933186665620889,0.8738263228050317
262
+ 5.689900426742532,8000,0.827775,0.901,0.928925,0.827775,0.827775,0.3003333333333333,0.901,0.18578500000000003,0.928925,0.827775,0.8669049999999959,0.8712585416666617,0.8929668333310461,0.8734331161309917
263
+ 5.761024182076814,8100,0.828325,0.901925,0.92915,0.828325,0.828325,0.3006416666666666,0.901925,0.18583000000000005,0.92915,0.828325,0.8674870833333292,0.8719067162698362,0.8935731518881023,0.874073460724267
264
+ 5.832147937411095,8200,0.82765,0.901475,0.9292,0.82765,0.82765,0.3004916666666666,0.901475,0.18584000000000006,0.9292,0.82765,0.867018749999996,0.8714473412698366,0.8932306586102523,0.8736176102556379
265
+ 5.903271692745377,8300,0.8282,0.9018,0.9294,0.8282,0.8282,0.3006,0.9018,0.18588000000000002,0.9294,0.8282,0.8674824999999957,0.8719401686507885,0.8937213383114889,0.8740581043681251
266
+ 5.974395448079658,8400,0.828325,0.90185,0.929925,0.828325,0.828325,0.30061666666666664,0.90185,0.185985,0.929925,0.828325,0.8677295833333308,0.8720828869047582,0.8937899946723228,0.8742234355394175
267
+ 6.0455192034139404,8500,0.8286,0.9023,0.929925,0.8286,0.8286,0.3007666666666666,0.9023,0.18598500000000004,0.929925,0.8286,0.8680758333333293,0.8724603968253921,0.894128127198057,0.8745885827849681
268
+ 6.116642958748222,8600,0.828175,0.90225,0.929325,0.828175,0.828175,0.30074999999999996,0.90225,0.185865,0.929325,0.828175,0.8675804166666634,0.8720532440476145,0.8938181845798283,0.8741853105958053
269
+ 6.187766714082503,8700,0.82805,0.90215,0.929725,0.82805,0.82805,0.30071666666666663,0.90215,0.18594500000000003,0.929725,0.82805,0.8674470833333299,0.8718606547619014,0.8936594724961032,0.8739825128657174
270
+ 6.2588904694167855,8800,0.82785,0.90205,0.9297,0.82785,0.82785,0.3006833333333333,0.90205,0.18594000000000002,0.9297,0.82785,0.8674191666666624,0.8718350892857094,0.893655148048901,0.873962454047865
271
+ 6.330014224751067,8900,0.828525,0.90235,0.930125,0.828525,0.828525,0.3007833333333333,0.90235,0.18602500000000002,0.930125,0.828525,0.8680445833333297,0.8724303174603135,0.8941609142278779,0.8745462546782486
272
+ 6.401137980085348,9000,0.828325,0.902,0.930025,0.828325,0.828325,0.3006666666666666,0.902,0.18600500000000003,0.930025,0.828325,0.8678520833333294,0.8722724900793611,0.8940775055650032,0.8743679763133748
273
+ 6.472261735419631,9100,0.828475,0.902075,0.92965,0.828475,0.828475,0.3006916666666666,0.902075,0.18593,0.92965,0.828475,0.8678004166666627,0.8722169146825343,0.8939301588915514,0.8743476154747785
274
+ 6.543385490753912,9200,0.8286,0.901525,0.929325,0.8286,0.8286,0.3005083333333333,0.901525,0.18586500000000003,0.929325,0.8286,0.867682083333329,0.8721559920634869,0.8938887495928067,0.8742804116459679
275
+ 6.614509246088193,9300,0.82855,0.902625,0.930175,0.82855,0.82855,0.300875,0.902625,0.18603500000000003,0.930175,0.82855,0.8681324999999959,0.8725687400793595,0.8943725725633536,0.8746570790204005
276
+ 6.685633001422475,9400,0.8288,0.90245,0.93045,0.8288,0.8288,0.3008166666666666,0.90245,0.18609,0.93045,0.8288,0.8681895833333286,0.8725463095238056,0.8942745973992838,0.8746616129494023
277
+ 6.756756756756757,9500,0.82835,0.902875,0.930225,0.82835,0.82835,0.3009583333333333,0.902875,0.18604500000000002,0.930225,0.82835,0.8679937499999963,0.8723895138888845,0.8941721987188964,0.8745020848542968
278
+ 6.827880512091038,9600,0.82875,0.9031,0.9302,0.82875,0.82875,0.3010333333333333,0.9031,0.18604,0.9302,0.82875,0.8681779166666623,0.8725444742063444,0.8942281345359215,0.8746847415751485
279
+ 6.89900426742532,9700,0.828775,0.903275,0.929975,0.828775,0.828775,0.3010916666666666,0.903275,0.18599500000000002,0.929975,0.828775,0.8680683333333283,0.8724758234126926,0.8941863586244911,0.8745995624422745
280
+ 6.970128022759602,9800,0.828725,0.90295,0.929725,0.828725,0.828725,0.3009833333333333,0.90295,0.18594500000000003,0.929725,0.828725,0.8680020833333291,0.8724609027777737,0.8941897885251688,0.8745898837951301
281
+ 7.0412517780938835,9900,0.82885,0.902975,0.9297,0.82885,0.82885,0.3009916666666666,0.902975,0.18594000000000002,0.9297,0.82885,0.8680699999999961,0.8725679563492021,0.8943256778966812,0.8746741755899257
282
+ 7.112375533428165,10000,0.828825,0.9032,0.930225,0.828825,0.828825,0.3010666666666666,0.9032,0.18604500000000002,0.930225,0.828825,0.8682116666666623,0.8726781349206311,0.8945003142115322,0.8747602011610026
283
+ 7.183499288762446,10100,0.82925,0.903175,0.9307,0.82925,0.82925,0.3010583333333333,0.903175,0.18614,0.9307,0.82925,0.8686112499999957,0.8730120634920597,0.8947492895278029,0.8750990256783056
284
+ 7.2546230440967285,10200,0.828975,0.9032,0.929975,0.828975,0.828975,0.30106666666666665,0.9032,0.185995,0.929975,0.828975,0.8683299999999957,0.8728390873015822,0.8946373730573125,0.8749157675276352
285
+ 7.32574679943101,10300,0.829325,0.902925,0.930125,0.829325,0.829325,0.3009749999999999,0.902925,0.18602500000000005,0.930125,0.829325,0.8684108333333296,0.8728565773809487,0.89456862638456,0.8749638243818381
286
+ 7.396870554765291,10400,0.82955,0.902825,0.930375,0.82955,0.82955,0.3009416666666666,0.902825,0.186075,0.930375,0.82955,0.8686049999999963,0.872999374999997,0.8946617306603606,0.8751100324975649
287
+ 7.467994310099574,10500,0.828575,0.90255,0.930075,0.828575,0.828575,0.30085,0.90255,0.186015,0.930075,0.828575,0.8679862499999965,0.8724021329365041,0.89419947680479,0.8745111244014055
288
+ 7.539118065433855,10600,0.82875,0.902775,0.93,0.82875,0.82875,0.300925,0.902775,0.186,0.93,0.82875,0.8681016666666623,0.8725284722222181,0.8942673037954326,0.874650465442734
289
+ 7.610241820768136,10700,0.829025,0.902775,0.93065,0.829025,0.829025,0.300925,0.902775,0.18613000000000005,0.93065,0.829025,0.8684041666666626,0.8728070238095202,0.8945965165100355,0.8749042361097843
290
+ 7.681365576102419,10800,0.828825,0.90305,0.930675,0.828825,0.828825,0.3010166666666666,0.90305,0.18613500000000002,0.930675,0.828825,0.8684024999999962,0.8727813988095199,0.8945637455567085,0.8748808984568961
291
+ 7.7524893314367,10900,0.8292,0.903125,0.93095,0.8292,0.8292,0.3010416666666666,0.903125,0.18619000000000002,0.93095,0.8292,0.8686620833333288,0.8730319146825355,0.8948043753706577,0.8751241215815942
292
+ 7.823613086770981,11000,0.829,0.9025,0.930925,0.829,0.829,0.3008333333333333,0.9025,0.18618500000000004,0.930925,0.829,0.8684387499999965,0.8728221924603132,0.8946611073435173,0.8748990293437295
293
+ 7.894736842105263,11100,0.828975,0.90305,0.930825,0.828975,0.828975,0.3010166666666666,0.90305,0.18616500000000005,0.930825,0.828975,0.8684504166666628,0.872878621031742,0.894732783487921,0.874949510397515
294
+ 7.965860597439545,11200,0.829025,0.90335,0.930975,0.829025,0.829025,0.3011166666666666,0.90335,0.18619500000000003,0.930975,0.829025,0.8685791666666625,0.8729389384920585,0.8947336103485135,0.8750321865169739
295
+ 8.036984352773827,11300,0.829175,0.903225,0.93075,0.829175,0.829175,0.301075,0.903225,0.18615,0.93075,0.829175,0.8685495833333294,0.8729351289682501,0.8947091937766589,0.8750362483818597
296
+ 8.108108108108109,11400,0.828775,0.902825,0.9307,0.828775,0.828775,0.30094166666666666,0.902825,0.18614000000000006,0.9307,0.828775,0.8683158333333291,0.8727185416666623,0.894565804979394,0.8748080562858987
297
+ 8.17923186344239,11500,0.82925,0.9035,0.931125,0.82925,0.82925,0.30116666666666664,0.9035,0.18622500000000006,0.931125,0.82925,0.8688029166666628,0.8731468551587266,0.8948947495660065,0.8752462046242943
298
+ 8.250355618776672,11600,0.82925,0.90315,0.930525,0.82925,0.82925,0.30105,0.90315,0.18610500000000002,0.930525,0.82925,0.8686154166666618,0.8730597619047572,0.8948103433261322,0.8751613059053036
299
+ 8.321479374110954,11700,0.8288,0.903,0.9303,0.8288,0.8288,0.30099999999999993,0.903,0.18606,0.9303,0.8288,0.8682366666666615,0.8727099900793599,0.8945588779528776,0.8748019389609702
300
+ 8.392603129445234,11800,0.82915,0.903,0.930975,0.82915,0.82915,0.30099999999999993,0.903,0.186195,0.930975,0.82915,0.8686054166666618,0.8730238888888842,0.8948571327655279,0.8750999528350607
301
+ 8.463726884779517,11900,0.8292,0.90305,0.930975,0.8292,0.8292,0.30101666666666665,0.90305,0.18619500000000003,0.930975,0.8292,0.8686795833333283,0.8731174404761854,0.8949856631148606,0.8751770568215895
302
+ 8.534850640113799,12000,0.82915,0.902825,0.930825,0.82915,0.82915,0.3009416666666666,0.902825,0.186165,0.930825,0.82915,0.8685429166666616,0.8729799206349158,0.8948425781107103,0.8750457194043624
303
+ 8.60597439544808,12100,0.829275,0.9026,0.9306,0.829275,0.829275,0.3008666666666666,0.9026,0.18612000000000004,0.9306,0.829275,0.8685587499999947,0.8730033134920586,0.8947984660739079,0.8750927454028319
304
+ 8.677098150782362,12200,0.829275,0.90285,0.930625,0.829275,0.829275,0.3009499999999999,0.90285,0.186125,0.930625,0.829275,0.8685366666666623,0.8729908829365044,0.894820570983912,0.8750702930735716
305
+ 8.748221906116642,12300,0.829275,0.902775,0.931025,0.829275,0.829275,0.300925,0.902775,0.186205,0.931025,0.829275,0.8686212499999958,0.8730137202380908,0.8948325362646716,0.8750985303378986
306
+ 8.819345661450924,12400,0.829275,0.90275,0.930925,0.829275,0.829275,0.3009166666666666,0.90275,0.18618500000000002,0.930925,0.829275,0.8686174999999962,0.8730114880952342,0.8948164169545031,0.8750965074481144
307
+ 8.890469416785207,12500,0.829325,0.902975,0.931,0.829325,0.829325,0.30099166666666666,0.902975,0.18620000000000003,0.931,0.829325,0.8687383333333292,0.8731361408730117,0.8949238533713902,0.8752293018007289
308
+ 8.961593172119487,12600,0.8295,0.902775,0.9313,0.8295,0.8295,0.30092499999999994,0.902775,0.18626000000000004,0.9313,0.8295,0.8689020833333292,0.8732708234126947,0.8950697708359244,0.8753445888295041
309
+ 9.03271692745377,12700,0.82955,0.903225,0.931225,0.82955,0.82955,0.3010749999999999,0.903225,0.18624500000000002,0.931225,0.82955,0.8689562499999957,0.8733353571428528,0.8951241731895416,0.8754143164227475
310
+ 9.103840682788052,12800,0.82935,0.903,0.931075,0.82935,0.82935,0.301,0.903,0.18621500000000002,0.931075,0.82935,0.8687095833333284,0.8731101884920585,0.8949432152346973,0.8751869179331543
311
+ 9.174964438122332,12900,0.829275,0.90295,0.930825,0.829275,0.829275,0.30098333333333327,0.90295,0.186165,0.930825,0.829275,0.868580416666662,0.8730058630952335,0.8948285709625057,0.8750920981147012
312
+ 9.246088193456615,13000,0.829225,0.902875,0.931025,0.829225,0.829225,0.3009583333333333,0.902875,0.186205,0.931025,0.829225,0.8686254166666617,0.8730646825396782,0.8949587209733254,0.8751212188013575
313
+ 9.317211948790897,13100,0.829375,0.9032,0.931225,0.829375,0.829375,0.30106666666666665,0.9032,0.18624500000000002,0.931225,0.829375,0.8687699999999946,0.8731812896825357,0.8950327784818084,0.8752468123212304
314
+ 9.388335704125177,13200,0.8296,0.90295,0.931375,0.8296,0.8296,0.3009833333333333,0.90295,0.186275,0.931375,0.8296,0.8689904166666618,0.8734019444444404,0.8952498050830631,0.8754484318300516
315
+ 9.45945945945946,13300,0.82955,0.902775,0.93115,0.82955,0.82955,0.30092499999999994,0.902775,0.18623,0.93115,0.82955,0.8688566666666613,0.8732831845238046,0.895126620421129,0.8753385862163494
316
+ 9.530583214793742,13400,0.82935,0.9029,0.931,0.82935,0.82935,0.3009666666666666,0.9029,0.1862,0.931,0.82935,0.8686979166666617,0.873155942460313,0.895046161583383,0.8751996221334275
317
+ 9.601706970128022,13500,0.829275,0.90305,0.9311,0.829275,0.829275,0.30101666666666665,0.90305,0.18622000000000002,0.9311,0.829275,0.8686891666666617,0.8731191964285671,0.895002811736966,0.875171085337729
318
+ 9.672830725462305,13600,0.82945,0.903025,0.931225,0.82945,0.82945,0.3010083333333333,0.903025,0.18624500000000002,0.931225,0.82945,0.8688266666666615,0.8732329166666623,0.8950792684099309,0.8752895149961647
319
+ 9.743954480796585,13700,0.82945,0.90305,0.931125,0.82945,0.82945,0.3010166666666666,0.90305,0.18622500000000003,0.931125,0.82945,0.8687870833333281,0.8732131448412651,0.895073130294718,0.8752667124340483
320
+ 9.815078236130867,13800,0.82945,0.9031,0.931025,0.82945,0.82945,0.3010333333333333,0.9031,0.186205,0.931025,0.82945,0.8687966666666614,0.8732481845238048,0.8951075415333131,0.8753000882548883
321
+ 9.88620199146515,13900,0.82935,0.903025,0.931025,0.82935,0.82935,0.3010083333333333,0.903025,0.18620500000000004,0.931025,0.82935,0.8687354166666617,0.8731747619047575,0.895030096393872,0.8752349923053447
322
+ 9.95732574679943,14000,0.82935,0.903025,0.9311,0.82935,0.82935,0.30100833333333327,0.903025,0.18622,0.9311,0.82935,0.8687558333333282,0.8731832242063449,0.8950372962037911,0.8752427301346968
final_metrics.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "val_cosine_accuracy@1": 0.9104,
3
- "val_cosine_accuracy@3": 0.9688,
4
- "val_cosine_accuracy@5": 0.9842,
5
- "val_cosine_precision@1": 0.9104,
6
- "val_cosine_precision@3": 0.32293333333333335,
7
- "val_cosine_precision@5": 0.19683999999999996,
8
- "val_cosine_recall@1": 0.9104,
9
- "val_cosine_recall@3": 0.9688,
10
- "val_cosine_recall@5": 0.9842,
11
- "val_cosine_ndcg@10": 0.9545809774353143,
12
- "val_cosine_mrr@1": 0.9104,
13
- "val_cosine_mrr@5": 0.9402433333333332,
14
- "val_cosine_mrr@10": 0.9416250793650793,
15
- "val_cosine_map@100": 0.9420576026548708
16
  }
 
1
  {
2
+ "val_cosine_accuracy@1": 0.8281,
3
+ "val_cosine_accuracy@3": 0.9026,
4
+ "val_cosine_accuracy@5": 0.93105,
5
+ "val_cosine_precision@1": 0.8281,
6
+ "val_cosine_precision@3": 0.3008666666666666,
7
+ "val_cosine_precision@5": 0.18621000000000004,
8
+ "val_cosine_recall@1": 0.8281,
9
+ "val_cosine_recall@3": 0.9026,
10
+ "val_cosine_recall@5": 0.93105,
11
+ "val_cosine_ndcg@10": 0.8942437004811851,
12
+ "val_cosine_mrr@1": 0.8281,
13
+ "val_cosine_mrr@5": 0.8677437499999962,
14
+ "val_cosine_mrr@10": 0.8721381249999942,
15
+ "val_cosine_map@100": 0.874246358340888
16
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a4f8953c7f93243ebe09fe7c0daa9cffac072577f9e8000399b77733108e58b
3
  size 114011616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07288e0fdf774af44164e803f88a32086d4b380ab0234eb9fc654ddb063c8150
3
  size 114011616