radoslavralev commited on
Commit
02b082c
·
verified ·
1 Parent(s): 1bfae1d

Training in progress, step 5000

Browse files
Information-Retrieval_evaluation_val_results.csv CHANGED
@@ -3,3 +3,4 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
3
  -1,-1,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451366666666663,0.9466038095238087,0.9586270476620361,0.946959374340519
4
  -1,-1,0.829575,0.9048,0.9324,0.829575,0.829575,0.3016,0.9048,0.18648000000000003,0.9324,0.829575,0.8693266666666628,0.873717658730154,0.8957411186558171,0.8757871539962314
5
  -1,-1,0.829275,0.9051,0.9329,0.829275,0.829275,0.30169999999999997,0.9051,0.18658000000000002,0.9329,0.829275,0.8692179166666618,0.8735753373015815,0.8956869608914538,0.8756452160249361
 
 
3
  -1,-1,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451366666666663,0.9466038095238087,0.9586270476620361,0.946959374340519
4
  -1,-1,0.829575,0.9048,0.9324,0.829575,0.829575,0.3016,0.9048,0.18648000000000003,0.9324,0.829575,0.8693266666666628,0.873717658730154,0.8957411186558171,0.8757871539962314
5
  -1,-1,0.829275,0.9051,0.9329,0.829275,0.829275,0.30169999999999997,0.9051,0.18658000000000002,0.9329,0.829275,0.8692179166666618,0.8735753373015815,0.8956869608914538,0.8756452160249361
6
+ -1,-1,0.76415,0.822175,0.84665,0.76415,0.76415,0.2740583333333333,0.822175,0.16933,0.84665,0.76415,0.7953370833333295,0.8000376587301573,0.8195085862842543,0.8034052122492592
README.md CHANGED
@@ -5,110 +5,38 @@ tags:
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
- - dataset_size:359997
9
  - loss:MultipleNegativesRankingLoss
10
  base_model: prajjwal1/bert-small
11
  widget:
12
- - source_sentence: When do you use Ms. or Mrs.? Is one for a married woman and one
13
- for one that's not married? Which one is for what?
14
  sentences:
15
- - When do you use Ms. or Mrs.? Is one for a married woman and one for one that's
16
- not married? Which one is for what?
17
- - Nations that do/does otherwise? Which one do I use?
18
- - Why don't bikes have a gear indicator?
19
- - source_sentence: Which ointment is applied to the face of UFC fighters at the commencement
20
- of a bout? What does it do?
21
  sentences:
22
- - How can I save a Snapchat video that others posted?
23
- - Which ointment is applied to the face of UFC fighters at the commencement of a
24
- bout? What does it do?
25
- - How do I get the body of a UFC Fighter?
26
- - source_sentence: Do you love the life you live?
27
  sentences:
28
- - How do I die and make it look like an accident?
29
- - Do you love the life you're living?
30
- - Why am I not motivated by money?
31
- - source_sentence: Ordered food on Swiggy 3 days ago.After accepting my money, said
32
- no more on Menu! When if ever will I atleast get refund in cr card a/c?
33
  sentences:
34
- - Is getting to the Tel Aviv airport to catch a 5:30 AM flight very expensive?
35
- - What is a cheap healthy diet I can keep the same and eat every day?
36
- - Ordered food on Swiggy 3 days ago.After accepting my money, said no more on Menu!
37
- When if ever will I atleast get refund in cr card a/c?
38
- - source_sentence: How do you earn money on Quora?
39
  sentences:
40
- - What should I do to make my life interesting?
41
- - What are the best blogs/sites to follow on fashion?
42
- - What is the best way to make money on Quora?
43
  pipeline_tag: sentence-similarity
44
  library_name: sentence-transformers
45
- metrics:
46
- - cosine_accuracy@1
47
- - cosine_accuracy@3
48
- - cosine_accuracy@5
49
- - cosine_precision@1
50
- - cosine_precision@3
51
- - cosine_precision@5
52
- - cosine_recall@1
53
- - cosine_recall@3
54
- - cosine_recall@5
55
- - cosine_ndcg@10
56
- - cosine_mrr@1
57
- - cosine_mrr@5
58
- - cosine_mrr@10
59
- - cosine_map@100
60
- model-index:
61
- - name: SentenceTransformer based on prajjwal1/bert-small
62
- results:
63
- - task:
64
- type: information-retrieval
65
- name: Information Retrieval
66
- dataset:
67
- name: val
68
- type: val
69
- metrics:
70
- - type: cosine_accuracy@1
71
- value: 0.764
72
- name: Cosine Accuracy@1
73
- - type: cosine_accuracy@3
74
- value: 0.8221
75
- name: Cosine Accuracy@3
76
- - type: cosine_accuracy@5
77
- value: 0.84635
78
- name: Cosine Accuracy@5
79
- - type: cosine_precision@1
80
- value: 0.764
81
- name: Cosine Precision@1
82
- - type: cosine_precision@3
83
- value: 0.27403333333333335
84
- name: Cosine Precision@3
85
- - type: cosine_precision@5
86
- value: 0.16927000000000003
87
- name: Cosine Precision@5
88
- - type: cosine_recall@1
89
- value: 0.764
90
- name: Cosine Recall@1
91
- - type: cosine_recall@3
92
- value: 0.8221
93
- name: Cosine Recall@3
94
- - type: cosine_recall@5
95
- value: 0.84635
96
- name: Cosine Recall@5
97
- - type: cosine_ndcg@10
98
- value: 0.8194586562525387
99
- name: Cosine Ndcg@10
100
- - type: cosine_mrr@1
101
- value: 0.764
102
- name: Cosine Mrr@1
103
- - type: cosine_mrr@5
104
- value: 0.7952029166666628
105
- name: Cosine Mrr@5
106
- - type: cosine_mrr@10
107
- value: 0.7999395039682529
108
- name: Cosine Mrr@10
109
- - type: cosine_map@100
110
- value: 0.8032950373722874
111
- name: Cosine Map@100
112
  ---
113
 
114
  # SentenceTransformer based on prajjwal1/bert-small
@@ -157,12 +85,12 @@ Then you can load this model and run inference.
157
  from sentence_transformers import SentenceTransformer
158
 
159
  # Download from the 🤗 Hub
160
- model = SentenceTransformer("redis/model-a-baseline")
161
  # Run inference
162
  sentences = [
163
- 'How do you earn money on Quora?',
164
- 'What is the best way to make money on Quora?',
165
- 'What are the best blogs/sites to follow on fashion?',
166
  ]
167
  embeddings = model.encode(sentences)
168
  print(embeddings.shape)
@@ -171,9 +99,9 @@ print(embeddings.shape)
171
  # Get the similarity scores for the embeddings
172
  similarities = model.similarity(embeddings, embeddings)
173
  print(similarities)
174
- # tensor([[ 0.9999, 0.9998, -0.0914],
175
- # [ 0.9998, 1.0001, -0.0884],
176
- # [-0.0914, -0.0884, 1.0000]])
177
  ```
178
 
179
  <!--
@@ -200,32 +128,6 @@ You can finetune this model on your own dataset.
200
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
201
  -->
202
 
203
- ## Evaluation
204
-
205
- ### Metrics
206
-
207
- #### Information Retrieval
208
-
209
- * Dataset: `val`
210
- * Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
211
-
212
- | Metric | Value |
213
- |:-------------------|:-----------|
214
- | cosine_accuracy@1 | 0.764 |
215
- | cosine_accuracy@3 | 0.8221 |
216
- | cosine_accuracy@5 | 0.8464 |
217
- | cosine_precision@1 | 0.764 |
218
- | cosine_precision@3 | 0.274 |
219
- | cosine_precision@5 | 0.1693 |
220
- | cosine_recall@1 | 0.764 |
221
- | cosine_recall@3 | 0.8221 |
222
- | cosine_recall@5 | 0.8464 |
223
- | **cosine_ndcg@10** | **0.8195** |
224
- | cosine_mrr@1 | 0.764 |
225
- | cosine_mrr@5 | 0.7952 |
226
- | cosine_mrr@10 | 0.7999 |
227
- | cosine_map@100 | 0.8033 |
228
-
229
  <!--
230
  ## Bias, Risks and Limitations
231
 
@@ -244,49 +146,23 @@ You can finetune this model on your own dataset.
244
 
245
  #### Unnamed Dataset
246
 
247
- * Size: 359,997 training samples
248
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
249
  * Approximate statistics based on the first 1000 samples:
250
- | | anchor | positive | negative |
251
- |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
252
- | type | string | string | string |
253
- | details | <ul><li>min: 4 tokens</li><li>mean: 15.46 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 15.52 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 5 tokens</li><li>mean: 17.1 tokens</li><li>max: 128 tokens</li></ul> |
254
  * Samples:
255
- | anchor | positive | negative |
256
- |:--------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
257
- | <code>Shall I upgrade my iPhone 5s to iOS 10 final version?</code> | <code>Should I upgrade an iPhone 5s to iOS 10?</code> | <code>What are the disadvantages and advantages of presidential democracy?</code> |
258
- | <code>Is Donald Trump really going to be the president of United States?</code> | <code>Do you think Donald Trump could conceivably be the next President of the United States?</code> | <code>What should we do when we are bored?</code> |
259
- | <code>What are real tips to improve work life balance?</code> | <code>What are the best ways to create a work life balance?</code> | <code>What are the best sites for college students to earn money online?</code> |
260
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
261
  ```json
262
  {
263
- "scale": 1.0,
264
- "similarity_fct": "cos_sim",
265
- "gather_across_devices": false
266
- }
267
- ```
268
-
269
- ### Evaluation Dataset
270
-
271
- #### Unnamed Dataset
272
-
273
- * Size: 40,000 evaluation samples
274
- * Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
275
- * Approximate statistics based on the first 1000 samples:
276
- | | anchor | positive | negative |
277
- |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
278
- | type | string | string | string |
279
- | details | <ul><li>min: 6 tokens</li><li>mean: 15.71 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.8 tokens</li><li>max: 78 tokens</li></ul> |
280
- * Samples:
281
- | anchor | positive | negative |
282
- |:------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------|
283
- | <code>Why were feathered dinosaur fossils only found in the last 20 years?</code> | <code>Why were feathered dinosaur fossils only found in the last 20 years?</code> | <code>Why are only few people aware that many dinosaurs had feathers?</code> |
284
- | <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>How much did Fox News and conservative leaning media networks stoke the anger that contributed to Donald Trump's popularity?</code> |
285
- | <code>How can guys last longer during sex?</code> | <code>How do I last longer in sex?</code> | <code>How do you get over the fear of death while fighting a war?</code> |
286
- * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
287
- ```json
288
- {
289
- "scale": 1.0,
290
  "similarity_fct": "cos_sim",
291
  "gather_across_devices": false
292
  }
@@ -295,49 +171,36 @@ You can finetune this model on your own dataset.
295
  ### Training Hyperparameters
296
  #### Non-Default Hyperparameters
297
 
298
- - `eval_strategy`: steps
299
- - `per_device_train_batch_size`: 256
300
- - `per_device_eval_batch_size`: 256
301
- - `learning_rate`: 2e-05
302
- - `weight_decay`: 0.001
303
- - `max_steps`: 14060
304
- - `warmup_ratio`: 0.1
305
  - `fp16`: True
306
- - `dataloader_drop_last`: True
307
- - `dataloader_num_workers`: 1
308
- - `dataloader_prefetch_factor`: 1
309
- - `load_best_model_at_end`: True
310
- - `optim`: adamw_torch
311
- - `ddp_find_unused_parameters`: False
312
- - `push_to_hub`: True
313
- - `hub_model_id`: redis/model-a-baseline
314
- - `eval_on_start`: True
315
 
316
  #### All Hyperparameters
317
  <details><summary>Click to expand</summary>
318
 
319
  - `overwrite_output_dir`: False
320
  - `do_predict`: False
321
- - `eval_strategy`: steps
322
  - `prediction_loss_only`: True
323
- - `per_device_train_batch_size`: 256
324
- - `per_device_eval_batch_size`: 256
325
  - `per_gpu_train_batch_size`: None
326
  - `per_gpu_eval_batch_size`: None
327
  - `gradient_accumulation_steps`: 1
328
  - `eval_accumulation_steps`: None
329
  - `torch_empty_cache_steps`: None
330
- - `learning_rate`: 2e-05
331
- - `weight_decay`: 0.001
332
  - `adam_beta1`: 0.9
333
  - `adam_beta2`: 0.999
334
  - `adam_epsilon`: 1e-08
335
- - `max_grad_norm`: 1.0
336
- - `num_train_epochs`: 3.0
337
- - `max_steps`: 14060
338
  - `lr_scheduler_type`: linear
339
  - `lr_scheduler_kwargs`: {}
340
- - `warmup_ratio`: 0.1
341
  - `warmup_steps`: 0
342
  - `log_level`: passive
343
  - `log_level_replica`: warning
@@ -365,14 +228,14 @@ You can finetune this model on your own dataset.
365
  - `tpu_num_cores`: None
366
  - `tpu_metrics_debug`: False
367
  - `debug`: []
368
- - `dataloader_drop_last`: True
369
- - `dataloader_num_workers`: 1
370
- - `dataloader_prefetch_factor`: 1
371
  - `past_index`: -1
372
  - `disable_tqdm`: False
373
  - `remove_unused_columns`: True
374
  - `label_names`: None
375
- - `load_best_model_at_end`: True
376
  - `ignore_data_skip`: False
377
  - `fsdp`: []
378
  - `fsdp_min_num_params`: 0
@@ -382,23 +245,23 @@ You can finetune this model on your own dataset.
382
  - `parallelism_config`: None
383
  - `deepspeed`: None
384
  - `label_smoothing_factor`: 0.0
385
- - `optim`: adamw_torch
386
  - `optim_args`: None
387
  - `adafactor`: False
388
  - `group_by_length`: False
389
  - `length_column_name`: length
390
  - `project`: huggingface
391
  - `trackio_space_id`: trackio
392
- - `ddp_find_unused_parameters`: False
393
  - `ddp_bucket_cap_mb`: None
394
  - `ddp_broadcast_buffers`: False
395
  - `dataloader_pin_memory`: True
396
  - `dataloader_persistent_workers`: False
397
  - `skip_memory_metrics`: True
398
  - `use_legacy_prediction_loop`: False
399
- - `push_to_hub`: True
400
  - `resume_from_checkpoint`: None
401
- - `hub_model_id`: redis/model-a-baseline
402
  - `hub_strategy`: every_save
403
  - `hub_private_repo`: None
404
  - `hub_always_push`: False
@@ -425,167 +288,32 @@ You can finetune this model on your own dataset.
425
  - `neftune_noise_alpha`: None
426
  - `optim_target_modules`: None
427
  - `batch_eval_metrics`: False
428
- - `eval_on_start`: True
429
  - `use_liger_kernel`: False
430
  - `liger_kernel_config`: None
431
  - `eval_use_gather_object`: False
432
  - `average_tokens_across_devices`: True
433
  - `prompts`: None
434
  - `batch_sampler`: batch_sampler
435
- - `multi_dataset_batch_sampler`: proportional
436
  - `router_mapping`: {}
437
  - `learning_rate_mapping`: {}
438
 
439
  </details>
440
 
441
  ### Training Logs
442
- <details><summary>Click to expand</summary>
 
 
 
 
 
 
 
 
 
 
443
 
444
- | Epoch | Step | Training Loss | Validation Loss | val_cosine_ndcg@10 |
445
- |:------:|:-----:|:-------------:|:---------------:|:------------------:|
446
- | 0 | 0 | - | 5.9519 | 0.8045 |
447
- | 0.0711 | 100 | 5.9667 | 5.7035 | 0.8414 |
448
- | 0.1422 | 200 | 5.7144 | 5.4232 | 0.8556 |
449
- | 0.2134 | 300 | 5.5593 | 5.3409 | 0.8496 |
450
- | 0.2845 | 400 | 5.4925 | 5.3198 | 0.8407 |
451
- | 0.3556 | 500 | 5.457 | 5.3155 | 0.8334 |
452
- | 0.4267 | 600 | 5.4353 | 5.3145 | 0.8287 |
453
- | 0.4979 | 700 | 5.4205 | 5.3114 | 0.8248 |
454
- | 0.5690 | 800 | 5.4107 | 5.3092 | 0.8222 |
455
- | 0.6401 | 900 | 5.4 | 5.3071 | 0.8192 |
456
- | 0.7112 | 1000 | 5.3927 | 5.3057 | 0.8176 |
457
- | 0.7824 | 1100 | 5.3867 | 5.3042 | 0.8173 |
458
- | 0.8535 | 1200 | 5.3815 | 5.3008 | 0.8171 |
459
- | 0.9246 | 1300 | 5.3757 | 5.2991 | 0.8163 |
460
- | 0.9957 | 1400 | 5.3711 | 5.2976 | 0.8156 |
461
- | 1.0669 | 1500 | 5.3671 | 5.2953 | 0.8160 |
462
- | 1.1380 | 1600 | 5.3641 | 5.2942 | 0.8162 |
463
- | 1.2091 | 1700 | 5.3611 | 5.2941 | 0.8166 |
464
- | 1.2802 | 1800 | 5.3588 | 5.2927 | 0.8167 |
465
- | 1.3514 | 1900 | 5.3549 | 5.2916 | 0.8162 |
466
- | 1.4225 | 2000 | 5.3542 | 5.2915 | 0.8162 |
467
- | 1.4936 | 2100 | 5.3512 | 5.2903 | 0.8164 |
468
- | 1.5647 | 2200 | 5.3502 | 5.2898 | 0.8166 |
469
- | 1.6358 | 2300 | 5.3479 | 5.2898 | 0.8162 |
470
- | 1.7070 | 2400 | 5.346 | 5.2889 | 0.8161 |
471
- | 1.7781 | 2500 | 5.3444 | 5.2881 | 0.8161 |
472
- | 1.8492 | 2600 | 5.3431 | 5.2879 | 0.8164 |
473
- | 1.9203 | 2700 | 5.3414 | 5.2876 | 0.8165 |
474
- | 1.9915 | 2800 | 5.34 | 5.2869 | 0.8166 |
475
- | 2.0626 | 2900 | 5.3384 | 5.2865 | 0.8166 |
476
- | 2.1337 | 3000 | 5.3365 | 5.2858 | 0.8163 |
477
- | 2.2048 | 3100 | 5.3356 | 5.2859 | 0.8168 |
478
- | 2.2760 | 3200 | 5.3346 | 5.2857 | 0.8163 |
479
- | 2.3471 | 3300 | 5.3338 | 5.2850 | 0.8166 |
480
- | 2.4182 | 3400 | 5.3324 | 5.2851 | 0.8171 |
481
- | 2.4893 | 3500 | 5.3323 | 5.2845 | 0.8170 |
482
- | 2.5605 | 3600 | 5.33 | 5.2837 | 0.8172 |
483
- | 2.6316 | 3700 | 5.3297 | 5.2841 | 0.8174 |
484
- | 2.7027 | 3800 | 5.3281 | 5.2838 | 0.8173 |
485
- | 2.7738 | 3900 | 5.3273 | 5.2835 | 0.8175 |
486
- | 2.8450 | 4000 | 5.3259 | 5.2829 | 0.8176 |
487
- | 2.9161 | 4100 | 5.3258 | 5.2834 | 0.8182 |
488
- | 2.9872 | 4200 | 5.3245 | 5.2827 | 0.8181 |
489
- | 3.0583 | 4300 | 5.3229 | 5.2821 | 0.8184 |
490
- | 3.1294 | 4400 | 5.3224 | 5.2820 | 0.8181 |
491
- | 3.2006 | 4500 | 5.3217 | 5.2819 | 0.8179 |
492
- | 3.2717 | 4600 | 5.3212 | 5.2820 | 0.8181 |
493
- | 3.3428 | 4700 | 5.3208 | 5.2814 | 0.8186 |
494
- | 3.4139 | 4800 | 5.3197 | 5.2819 | 0.8181 |
495
- | 3.4851 | 4900 | 5.3196 | 5.2815 | 0.8183 |
496
- | 3.5562 | 5000 | 5.3189 | 5.2811 | 0.8184 |
497
- | 3.6273 | 5100 | 5.3183 | 5.2807 | 0.8181 |
498
- | 3.6984 | 5200 | 5.3169 | 5.2808 | 0.8182 |
499
- | 3.7696 | 5300 | 5.3168 | 5.2805 | 0.8182 |
500
- | 3.8407 | 5400 | 5.3163 | 5.2806 | 0.8180 |
501
- | 3.9118 | 5500 | 5.3157 | 5.2806 | 0.8180 |
502
- | 3.9829 | 5600 | 5.3153 | 5.2803 | 0.8184 |
503
- | 4.0541 | 5700 | 5.3147 | 5.2801 | 0.8184 |
504
- | 4.1252 | 5800 | 5.3134 | 5.2800 | 0.8183 |
505
- | 4.1963 | 5900 | 5.313 | 5.2796 | 0.8185 |
506
- | 4.2674 | 6000 | 5.3131 | 5.2797 | 0.8186 |
507
- | 4.3385 | 6100 | 5.3118 | 5.2795 | 0.8184 |
508
- | 4.4097 | 6200 | 5.3115 | 5.2792 | 0.8183 |
509
- | 4.4808 | 6300 | 5.3111 | 5.2790 | 0.8186 |
510
- | 4.5519 | 6400 | 5.3112 | 5.2793 | 0.8188 |
511
- | 4.6230 | 6500 | 5.3103 | 5.2792 | 0.8190 |
512
- | 4.6942 | 6600 | 5.3104 | 5.2788 | 0.8192 |
513
- | 4.7653 | 6700 | 5.3093 | 5.2792 | 0.8192 |
514
- | 4.8364 | 6800 | 5.309 | 5.2784 | 0.8189 |
515
- | 4.9075 | 6900 | 5.3095 | 5.2788 | 0.8191 |
516
- | 4.9787 | 7000 | 5.3083 | 5.2781 | 0.8192 |
517
- | 5.0498 | 7100 | 5.3077 | 5.2784 | 0.8193 |
518
- | 5.1209 | 7200 | 5.3071 | 5.2783 | 0.8191 |
519
- | 5.1920 | 7300 | 5.307 | 5.2781 | 0.8189 |
520
- | 5.2632 | 7400 | 5.3069 | 5.2781 | 0.8192 |
521
- | 5.3343 | 7500 | 5.3064 | 5.2784 | 0.8193 |
522
- | 5.4054 | 7600 | 5.3061 | 5.2778 | 0.8193 |
523
- | 5.4765 | 7700 | 5.3054 | 5.2777 | 0.8193 |
524
- | 5.5477 | 7800 | 5.3056 | 5.2781 | 0.8193 |
525
- | 5.6188 | 7900 | 5.3054 | 5.2778 | 0.8192 |
526
- | 5.6899 | 8000 | 5.305 | 5.2774 | 0.8191 |
527
- | 5.7610 | 8100 | 5.3049 | 5.2773 | 0.8191 |
528
- | 5.8321 | 8200 | 5.3041 | 5.2774 | 0.8188 |
529
- | 5.9033 | 8300 | 5.3041 | 5.2771 | 0.8191 |
530
- | 5.9744 | 8400 | 5.3039 | 5.2775 | 0.8191 |
531
- | 6.0455 | 8500 | 5.3031 | 5.2770 | 0.8190 |
532
- | 6.1166 | 8600 | 5.3028 | 5.2771 | 0.8191 |
533
- | 6.1878 | 8700 | 5.3026 | 5.2770 | 0.8189 |
534
- | 6.2589 | 8800 | 5.302 | 5.2772 | 0.8189 |
535
- | 6.3300 | 8900 | 5.3025 | 5.2772 | 0.8190 |
536
- | 6.4011 | 9000 | 5.3025 | 5.2768 | 0.8189 |
537
- | 6.4723 | 9100 | 5.3022 | 5.2765 | 0.8191 |
538
- | 6.5434 | 9200 | 5.3021 | 5.2765 | 0.8193 |
539
- | 6.6145 | 9300 | 5.3014 | 5.2765 | 0.8196 |
540
- | 6.6856 | 9400 | 5.3012 | 5.2766 | 0.8194 |
541
- | 6.7568 | 9500 | 5.3008 | 5.2763 | 0.8194 |
542
- | 6.8279 | 9600 | 5.3009 | 5.2764 | 0.8193 |
543
- | 6.8990 | 9700 | 5.3009 | 5.2762 | 0.8195 |
544
- | 6.9701 | 9800 | 5.3008 | 5.2762 | 0.8192 |
545
- | 7.0413 | 9900 | 5.3001 | 5.2763 | 0.8193 |
546
- | 7.1124 | 10000 | 5.2998 | 5.2762 | 0.8194 |
547
- | 7.1835 | 10100 | 5.2996 | 5.2760 | 0.8195 |
548
- | 7.2546 | 10200 | 5.2996 | 5.2763 | 0.8192 |
549
- | 7.3257 | 10300 | 5.2992 | 5.2761 | 0.8194 |
550
- | 7.3969 | 10400 | 5.2995 | 5.2762 | 0.8193 |
551
- | 7.4680 | 10500 | 5.2994 | 5.2760 | 0.8191 |
552
- | 7.5391 | 10600 | 5.2986 | 5.2758 | 0.8192 |
553
- | 7.6102 | 10700 | 5.2985 | 5.2760 | 0.8190 |
554
- | 7.6814 | 10800 | 5.2988 | 5.2758 | 0.8193 |
555
- | 7.7525 | 10900 | 5.2993 | 5.2757 | 0.8190 |
556
- | 7.8236 | 11000 | 5.2988 | 5.2758 | 0.8194 |
557
- | 7.8947 | 11100 | 5.2989 | 5.2757 | 0.8192 |
558
- | 7.9659 | 11200 | 5.2987 | 5.2757 | 0.8195 |
559
- | 8.0370 | 11300 | 5.2983 | 5.2756 | 0.8194 |
560
- | 8.1081 | 11400 | 5.2981 | 5.2756 | 0.8196 |
561
- | 8.1792 | 11500 | 5.2981 | 5.2756 | 0.8192 |
562
- | 8.2504 | 11600 | 5.2977 | 5.2757 | 0.8194 |
563
- | 8.3215 | 11700 | 5.2979 | 5.2756 | 0.8194 |
564
- | 8.3926 | 11800 | 5.2975 | 5.2754 | 0.8195 |
565
- | 8.4637 | 11900 | 5.2976 | 5.2756 | 0.8195 |
566
- | 8.5349 | 12000 | 5.2973 | 5.2755 | 0.8194 |
567
- | 8.6060 | 12100 | 5.2969 | 5.2754 | 0.8195 |
568
- | 8.6771 | 12200 | 5.2979 | 5.2755 | 0.8193 |
569
- | 8.7482 | 12300 | 5.2976 | 5.2754 | 0.8194 |
570
- | 8.8193 | 12400 | 5.2977 | 5.2753 | 0.8194 |
571
- | 8.8905 | 12500 | 5.2973 | 5.2754 | 0.8193 |
572
- | 8.9616 | 12600 | 5.297 | 5.2753 | 0.8195 |
573
- | 9.0327 | 12700 | 5.2964 | 5.2754 | 0.8195 |
574
- | 9.1038 | 12800 | 5.2971 | 5.2754 | 0.8193 |
575
- | 9.1750 | 12900 | 5.2969 | 5.2753 | 0.8194 |
576
- | 9.2461 | 13000 | 5.2969 | 5.2752 | 0.8193 |
577
- | 9.3172 | 13100 | 5.2966 | 5.2753 | 0.8194 |
578
- | 9.3883 | 13200 | 5.2968 | 5.2752 | 0.8195 |
579
- | 9.4595 | 13300 | 5.2972 | 5.2753 | 0.8194 |
580
- | 9.5306 | 13400 | 5.2968 | 5.2752 | 0.8194 |
581
- | 9.6017 | 13500 | 5.2963 | 5.2751 | 0.8196 |
582
- | 9.6728 | 13600 | 5.2968 | 5.2751 | 0.8196 |
583
- | 9.7440 | 13700 | 5.2967 | 5.2752 | 0.8197 |
584
- | 9.8151 | 13800 | 5.2965 | 5.2752 | 0.8195 |
585
- | 9.8862 | 13900 | 5.2962 | 5.2751 | 0.8193 |
586
- | 9.9573 | 14000 | 5.2964 | 5.2751 | 0.8195 |
587
-
588
- </details>
589
 
590
  ### Framework Versions
591
  - Python: 3.10.18
 
5
  - feature-extraction
6
  - dense
7
  - generated_from_trainer
8
+ - dataset_size:100000
9
  - loss:MultipleNegativesRankingLoss
10
  base_model: prajjwal1/bert-small
11
  widget:
12
+ - source_sentence: How do I polish my English skills?
 
13
  sentences:
14
+ - How can we polish English skills?
15
+ - Why should I move to Israel as a Jew?
16
+ - What are vitamins responsible for?
17
+ - source_sentence: Can I use the Kozuka Gothic Pro font as a font-face on my web site?
 
 
18
  sentences:
19
+ - Can I use the Kozuka Gothic Pro font as a font-face on my web site?
20
+ - Why are Google, Facebook, YouTube and other social networking sites banned in
21
+ China?
22
+ - What font is used in Bloomberg Terminal?
23
+ - source_sentence: Is Quora the best Q&A site?
24
  sentences:
25
+ - What was the best Quora question ever?
26
+ - Is Quora the best inquiry site?
27
+ - Where do I buy Oway hair products online?
28
+ - source_sentence: How can I customize my walking speed on Google Maps?
 
29
  sentences:
30
+ - How do I bring back Google maps icon in my home screen?
31
+ - How many pages are there in all the Harry Potter books combined?
32
+ - How can I customize my walking speed on Google Maps?
33
+ - source_sentence: DId something exist before the Big Bang?
 
34
  sentences:
35
+ - How can I improve my memory problem?
36
+ - Where can I buy Fairy Tail Manga?
37
+ - Is there a scientific name for what existed before the Big Bang?
38
  pipeline_tag: sentence-similarity
39
  library_name: sentence-transformers
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  ---
41
 
42
  # SentenceTransformer based on prajjwal1/bert-small
 
85
  from sentence_transformers import SentenceTransformer
86
 
87
  # Download from the 🤗 Hub
88
+ model = SentenceTransformer("sentence_transformers_model_id")
89
  # Run inference
90
  sentences = [
91
+ 'DId something exist before the Big Bang?',
92
+ 'Is there a scientific name for what existed before the Big Bang?',
93
+ 'Where can I buy Fairy Tail Manga?',
94
  ]
95
  embeddings = model.encode(sentences)
96
  print(embeddings.shape)
 
99
  # Get the similarity scores for the embeddings
100
  similarities = model.similarity(embeddings, embeddings)
101
  print(similarities)
102
+ # tensor([[ 1.0000, 0.7596, -0.0398],
103
+ # [ 0.7596, 1.0000, -0.0308],
104
+ # [-0.0398, -0.0308, 1.0000]])
105
  ```
106
 
107
  <!--
 
128
  *List how the model may foreseeably be misused and address what users ought not to do with the model.*
129
  -->
130
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  <!--
132
  ## Bias, Risks and Limitations
133
 
 
146
 
147
  #### Unnamed Dataset
148
 
149
+ * Size: 100,000 training samples
150
+ * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
151
  * Approximate statistics based on the first 1000 samples:
152
+ | | sentence_0 | sentence_1 | sentence_2 |
153
+ |:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
154
+ | type | string | string | string |
155
+ | details | <ul><li>min: 3 tokens</li><li>mean: 15.53 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 15.5 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.87 tokens</li><li>max: 128 tokens</li></ul> |
156
  * Samples:
157
+ | sentence_0 | sentence_1 | sentence_2 |
158
+ |:----------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------|:-----------------------------------------------------------------------|
159
+ | <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>How much is the airport tax in bogota?</code> |
160
+ | <code>Which concept is more important: good planning or hard work?</code> | <code>Which concept is more important: good planning or hard work?</code> | <code>What is important in life: luck or hard work?</code> |
161
+ | <code>What is the most efficient way to make money?</code> | <code>How can I make my money make money?</code> | <code>What can one learn about Quantum Mechanics in 10 minutes?</code> |
162
  * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
163
  ```json
164
  {
165
+ "scale": 20.0,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
166
  "similarity_fct": "cos_sim",
167
  "gather_across_devices": false
168
  }
 
171
  ### Training Hyperparameters
172
  #### Non-Default Hyperparameters
173
 
174
+ - `per_device_train_batch_size`: 64
175
+ - `per_device_eval_batch_size`: 64
 
 
 
 
 
176
  - `fp16`: True
177
+ - `multi_dataset_batch_sampler`: round_robin
 
 
 
 
 
 
 
 
178
 
179
  #### All Hyperparameters
180
  <details><summary>Click to expand</summary>
181
 
182
  - `overwrite_output_dir`: False
183
  - `do_predict`: False
184
+ - `eval_strategy`: no
185
  - `prediction_loss_only`: True
186
+ - `per_device_train_batch_size`: 64
187
+ - `per_device_eval_batch_size`: 64
188
  - `per_gpu_train_batch_size`: None
189
  - `per_gpu_eval_batch_size`: None
190
  - `gradient_accumulation_steps`: 1
191
  - `eval_accumulation_steps`: None
192
  - `torch_empty_cache_steps`: None
193
+ - `learning_rate`: 5e-05
194
+ - `weight_decay`: 0.0
195
  - `adam_beta1`: 0.9
196
  - `adam_beta2`: 0.999
197
  - `adam_epsilon`: 1e-08
198
+ - `max_grad_norm`: 1
199
+ - `num_train_epochs`: 3
200
+ - `max_steps`: -1
201
  - `lr_scheduler_type`: linear
202
  - `lr_scheduler_kwargs`: {}
203
+ - `warmup_ratio`: 0.0
204
  - `warmup_steps`: 0
205
  - `log_level`: passive
206
  - `log_level_replica`: warning
 
228
  - `tpu_num_cores`: None
229
  - `tpu_metrics_debug`: False
230
  - `debug`: []
231
+ - `dataloader_drop_last`: False
232
+ - `dataloader_num_workers`: 0
233
+ - `dataloader_prefetch_factor`: None
234
  - `past_index`: -1
235
  - `disable_tqdm`: False
236
  - `remove_unused_columns`: True
237
  - `label_names`: None
238
+ - `load_best_model_at_end`: False
239
  - `ignore_data_skip`: False
240
  - `fsdp`: []
241
  - `fsdp_min_num_params`: 0
 
245
  - `parallelism_config`: None
246
  - `deepspeed`: None
247
  - `label_smoothing_factor`: 0.0
248
+ - `optim`: adamw_torch_fused
249
  - `optim_args`: None
250
  - `adafactor`: False
251
  - `group_by_length`: False
252
  - `length_column_name`: length
253
  - `project`: huggingface
254
  - `trackio_space_id`: trackio
255
+ - `ddp_find_unused_parameters`: None
256
  - `ddp_bucket_cap_mb`: None
257
  - `ddp_broadcast_buffers`: False
258
  - `dataloader_pin_memory`: True
259
  - `dataloader_persistent_workers`: False
260
  - `skip_memory_metrics`: True
261
  - `use_legacy_prediction_loop`: False
262
+ - `push_to_hub`: False
263
  - `resume_from_checkpoint`: None
264
+ - `hub_model_id`: None
265
  - `hub_strategy`: every_save
266
  - `hub_private_repo`: None
267
  - `hub_always_push`: False
 
288
  - `neftune_noise_alpha`: None
289
  - `optim_target_modules`: None
290
  - `batch_eval_metrics`: False
291
+ - `eval_on_start`: False
292
  - `use_liger_kernel`: False
293
  - `liger_kernel_config`: None
294
  - `eval_use_gather_object`: False
295
  - `average_tokens_across_devices`: True
296
  - `prompts`: None
297
  - `batch_sampler`: batch_sampler
298
+ - `multi_dataset_batch_sampler`: round_robin
299
  - `router_mapping`: {}
300
  - `learning_rate_mapping`: {}
301
 
302
  </details>
303
 
304
  ### Training Logs
305
+ | Epoch | Step | Training Loss |
306
+ |:------:|:----:|:-------------:|
307
+ | 0.3199 | 500 | 0.2284 |
308
+ | 0.6398 | 1000 | 0.0571 |
309
+ | 0.9597 | 1500 | 0.0486 |
310
+ | 1.2796 | 2000 | 0.0378 |
311
+ | 1.5995 | 2500 | 0.0367 |
312
+ | 1.9194 | 3000 | 0.0338 |
313
+ | 2.2393 | 3500 | 0.0327 |
314
+ | 2.5592 | 4000 | 0.0285 |
315
+ | 2.8791 | 4500 | 0.0285 |
316
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
317
 
318
  ### Framework Versions
319
  - Python: 3.10.18
eval/Information-Retrieval_evaluation_val_results.csv CHANGED
@@ -462,3 +462,147 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
462
  9.815078236130867,13800,0.763825,0.82225,0.847,0.763825,0.763825,0.2740833333333333,0.82225,0.16940000000000002,0.847,0.763825,0.7952858333333293,0.799954811507935,0.8195212182618785,0.803303660228457
463
  9.88620199146515,13900,0.763875,0.82185,0.84695,0.763875,0.763875,0.27395,0.82185,0.16939000000000004,0.84695,0.763875,0.7952058333333295,0.7998402182539672,0.8193424066611443,0.8032217930140984
464
  9.95732574679943,14000,0.764,0.8221,0.84635,0.764,0.764,0.27403333333333335,0.8221,0.16927000000000003,0.84635,0.764,0.7952029166666628,0.7999395039682529,0.8194586562525387,0.8032950373722874
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
462
  9.815078236130867,13800,0.763825,0.82225,0.847,0.763825,0.763825,0.2740833333333333,0.82225,0.16940000000000002,0.847,0.763825,0.7952858333333293,0.799954811507935,0.8195212182618785,0.803303660228457
463
  9.88620199146515,13900,0.763875,0.82185,0.84695,0.763875,0.763875,0.27395,0.82185,0.16939000000000004,0.84695,0.763875,0.7952058333333295,0.7998402182539672,0.8193424066611443,0.8032217930140984
464
  9.95732574679943,14000,0.764,0.8221,0.84635,0.764,0.764,0.27403333333333335,0.8221,0.16927000000000003,0.84635,0.764,0.7952029166666628,0.7999395039682529,0.8194586562525387,0.8032950373722874
465
+ 0,0,0.754525,0.807575,0.8301,0.754525,0.754525,0.26919166666666666,0.807575,0.16601999999999997,0.8301,0.754525,0.7830537499999958,0.7870734424603156,0.8044818711611815,0.790208838398216
466
+ 0.07112375533428165,100,0.7846,0.8459,0.8711,0.7846,0.7846,0.2819666666666666,0.8459,0.17422000000000004,0.8711,0.7846,0.8175595833333287,0.8218771726190455,0.8413714264178179,0.8249413998110878
467
+ 0.1422475106685633,200,0.79835,0.8623,0.885675,0.79835,0.79835,0.2874333333333333,0.8623,0.17713500000000004,0.885675,0.79835,0.8318512499999939,0.836128045634916,0.855655110536833,0.8388773888658537
468
+ 0.21337126600284495,300,0.793725,0.85595,0.879675,0.793725,0.793725,0.28531666666666666,0.85595,0.175935,0.879675,0.793725,0.826637499999994,0.8306767956349161,0.8496438038402215,0.833452231279345
469
+ 0.2844950213371266,400,0.786225,0.84655,0.8699,0.786225,0.786225,0.2821833333333333,0.84655,0.17398000000000002,0.8699,0.786225,0.8181699999999944,0.8221539781746012,0.8406979711380332,0.8248666969847387
470
+ 0.35561877667140823,500,0.779925,0.838625,0.862225,0.779925,0.779925,0.27954166666666663,0.838625,0.172445,0.862225,0.779925,0.8113874999999955,0.8152349007936488,0.8333842785333231,0.8179856709019278
471
+ 0.4267425320056899,600,0.77615,0.83305,0.85555,0.77615,0.77615,0.2776833333333333,0.83305,0.17111,0.85555,0.77615,0.8065837499999955,0.8106722420634908,0.8286924647605718,0.8134236724071078
472
+ 0,0,0.757,0.811375,0.832975,0.757,0.757,0.27045833333333325,0.811375,0.16659500000000002,0.832975,0.757,0.7860745833333305,0.7898006746031736,0.8068310819573903,0.7928863872268869
473
+ 0,0,0.7569,0.811325,0.832975,0.7569,0.7569,0.27044166666666664,0.811325,0.166595,0.832975,0.7569,0.7860187499999972,0.7897472123015863,0.8067966793307086,0.7928306424812177
474
+ 0.2849002849002849,100,0.79975,0.86215,0.8864,0.79975,0.79975,0.2873833333333333,0.86215,0.17728,0.8864,0.79975,0.8329958333333292,0.8371555059523791,0.8563566180802749,0.8398102921473892
475
+ 0.5698005698005698,200,0.7925,0.852,0.87525,0.7925,0.7925,0.284,0.852,0.17505,0.87525,0.7925,0.8242770833333285,0.8281965773809494,0.8465099519180141,0.830860034111173
476
+ 0.8547008547008547,300,0.781375,0.838775,0.86035,0.781375,0.781375,0.2795916666666666,0.838775,0.17207,0.86035,0.781375,0.8116999999999953,0.815356507936506,0.8327024022306619,0.8181508397271755
477
+ 1.1396011396011396,400,0.7759,0.8309,0.851525,0.7759,0.7759,0.27696666666666664,0.8309,0.170305,0.851525,0.7759,0.8048937499999954,0.8085814682539659,0.8254910406165096,0.8115408891716827
478
+ 1.4245014245014245,500,0.772425,0.82605,0.84745,0.772425,0.772425,0.27535,0.82605,0.16949,0.84745,0.772425,0.801122916666663,0.8049703571428553,0.8219946152512894,0.8079477908294356
479
+ 1.7094017094017095,600,0.7709,0.824975,0.8463,0.7709,0.7709,0.27499166666666663,0.824975,0.16926,0.8463,0.7709,0.7997583333333297,0.8036332837301585,0.8207335647024644,0.8066639744523467
480
+ 1.9943019943019942,700,0.769975,0.823675,0.846275,0.769975,0.769975,0.2745583333333333,0.823675,0.16925500000000002,0.846275,0.769975,0.7989949999999969,0.8027658630952357,0.8199615531109039,0.8058107407966907
481
+ 2.2792022792022792,800,0.7693,0.823875,0.846175,0.7693,0.7693,0.274625,0.823875,0.169235,0.846175,0.7693,0.7986245833333296,0.8024050099206331,0.8196545256913207,0.805510324876921
482
+ 2.564102564102564,900,0.769575,0.82335,0.84615,0.769575,0.769575,0.27444999999999997,0.82335,0.16923000000000002,0.84615,0.769575,0.7986983333333304,0.802586666666665,0.819957118736493,0.8056878206576987
483
+ 2.849002849002849,1000,0.7691,0.8234,0.846125,0.7691,0.7691,0.27446666666666664,0.8234,0.16922500000000001,0.846125,0.7691,0.7984204166666636,0.8023565674603175,0.8198217799919566,0.8055107499465582
484
+ 0,0,0.756925,0.81135,0.832975,0.756925,0.756925,0.27044999999999997,0.81135,0.166595,0.832975,0.756925,0.7860320833333306,0.7897576984126973,0.8067989560864608,0.7928436652081169
485
+ 0.2849002849002849,100,0.799775,0.862075,0.8864,0.799775,0.799775,0.2873583333333333,0.862075,0.17728,0.8864,0.799775,0.8329937499999959,0.8371539186507918,0.8563545850975712,0.8398090360749207
486
+ 0.5698005698005698,200,0.792525,0.852,0.8752,0.792525,0.792525,0.284,0.852,0.17504000000000003,0.8752,0.792525,0.824271249999995,0.8281995238095209,0.8465115466169034,0.8308628965209384
487
+ 0.8547008547008547,300,0.781425,0.83875,0.860325,0.781425,0.781425,0.27958333333333324,0.83875,0.172065,0.860325,0.781425,0.8117137499999953,0.8153769940476172,0.8327223716360451,0.8181691434152646
488
+ 1.1396011396011396,400,0.776,0.8309,0.851525,0.776,0.776,0.27696666666666664,0.8309,0.170305,0.851525,0.776,0.8049424999999955,0.808629940476188,0.8255265529181479,0.8115893097540797
489
+ 1.4245014245014245,500,0.77245,0.826075,0.847425,0.77245,0.77245,0.2753583333333333,0.826075,0.169485,0.847425,0.77245,0.8011295833333297,0.8049792857142839,0.8219959765286132,0.807958647632387
490
+ 1.7094017094017095,600,0.770975,0.825,0.84635,0.770975,0.770975,0.275,0.825,0.16926999999999998,0.84635,0.770975,0.7998049999999963,0.8036710218253964,0.8207617608415884,0.8067018917058776
491
+ 1.9943019943019942,700,0.770025,0.823675,0.846275,0.770025,0.770025,0.2745583333333333,0.823675,0.16925500000000002,0.846275,0.770025,0.7990199999999967,0.8027929166666641,0.8199867865324695,0.805835145066436
492
+ 2.2792022792022792,800,0.769375,0.823875,0.846175,0.769375,0.769375,0.274625,0.823875,0.169235,0.846175,0.769375,0.7986662499999962,0.8024467063492045,0.8196855673585592,0.8055517545231466
493
+ 2.564102564102564,900,0.7696,0.82335,0.84615,0.7696,0.7696,0.27444999999999997,0.82335,0.16923000000000002,0.84615,0.7696,0.7987095833333303,0.8025977678571411,0.8199651247639522,0.8056987038056344
494
+ 2.849002849002849,1000,0.769175,0.8234,0.846075,0.769175,0.769175,0.27446666666666664,0.8234,0.169215,0.846075,0.769175,0.7984474999999968,0.8023916369047619,0.819847580644878,0.8055455645928971
495
+ 0,0,0.756925,0.811325,0.832975,0.756925,0.756925,0.27044166666666664,0.811325,0.16659500000000002,0.832975,0.756925,0.7860366666666637,0.7897715277777766,0.8068261092449236,0.7928502735721635
496
+ 0.07112375533428165,100,0.7999,0.861975,0.886425,0.7999,0.7999,0.287325,0.861975,0.177285,0.886425,0.7999,0.8329808333333287,0.8371251984126964,0.8563155772584116,0.8398022217446666
497
+ 0.1422475106685633,200,0.794325,0.8553,0.878525,0.794325,0.794325,0.2851,0.8553,0.17570500000000003,0.878525,0.794325,0.8267374999999947,0.8305394841269801,0.8489094362604251,0.8332286266074361
498
+ 0.21337126600284495,300,0.78485,0.842075,0.863375,0.78485,0.78485,0.2806916666666667,0.842075,0.172675,0.863375,0.78485,0.8150720833333287,0.8188160119047592,0.83619245007385,0.8215492677449724
499
+ 0.2844950213371266,400,0.777775,0.833175,0.8542,0.777775,0.777775,0.27772499999999994,0.833175,0.17084000000000002,0.8542,0.777775,0.8070979166666624,0.8107996230158705,0.8278241316562024,0.8136652065977691
500
+ 0.35561877667140823,500,0.77395,0.8277,0.848925,0.77395,0.77395,0.2759,0.8277,0.169785,0.848925,0.77395,0.8026024999999963,0.8064195436507905,0.8234369127110808,0.8092864273342918
501
+ 0,0,0.75695,0.81135,0.833025,0.75695,0.75695,0.27044999999999997,0.81135,0.16660500000000003,0.833025,0.75695,0.7860599999999972,0.7897825992063481,0.8068288099196645,0.79286346457412
502
+ 0.07112375533428165,100,0.787525,0.84765,0.872825,0.787525,0.787525,0.28255,0.84765,0.174565,0.872825,0.787525,0.8200616666666627,0.8242667361111095,0.8434021080706371,0.8272603316013856
503
+ 0.1422475106685633,200,0.800875,0.862875,0.887625,0.800875,0.800875,0.287625,0.862875,0.177525,0.887625,0.800875,0.8341379166666631,0.8382101091269837,0.8573139451618763,0.840874494584925
504
+ 0.21337126600284495,300,0.796675,0.857525,0.88125,0.796675,0.796675,0.2858416666666666,0.857525,0.17625,0.88125,0.796675,0.8291537499999944,0.8330430654761858,0.8516132915919113,0.8357202570608268
505
+ 0.2844950213371266,400,0.7899,0.848,0.8712,0.7899,0.7899,0.2826666666666666,0.848,0.17424,0.8712,0.7899,0.8210562499999958,0.8248586706349186,0.8427570230120356,0.8275746227632077
506
+ 0.35561877667140823,500,0.78435,0.84085,0.863225,0.78435,0.78435,0.28028333333333333,0.84085,0.172645,0.863225,0.78435,0.8145704166666621,0.8182268749999977,0.8355804330527246,0.8209647560975787
507
+ 0.4267425320056899,600,0.7802,0.8361,0.8576,0.7802,0.7802,0.27869999999999995,0.8361,0.17152,0.8576,0.7802,0.8098908333333292,0.8135463492063476,0.830654625279805,0.8163927962763059
508
+ 0.49786628733997157,700,0.777175,0.832,0.85325,0.777175,0.777175,0.2773333333333334,0.832,0.17065,0.85325,0.777175,0.80638208333333,0.810140039682537,0.8272090556062501,0.8130014044089133
509
+ 0.5689900426742532,800,0.7758,0.829675,0.850625,0.7758,0.7758,0.27655833333333324,0.829675,0.17012500000000005,0.850625,0.7758,0.8044174999999962,0.8081950198412682,0.8250726167267949,0.8111216889217711
510
+ 0.6401137980085349,900,0.77365,0.827625,0.84905,0.77365,0.77365,0.275875,0.827625,0.16981000000000002,0.84905,0.77365,0.8023862499999959,0.8060900198412684,0.8229627553948096,0.8090680588419786
511
+ 0.7112375533428165,1000,0.772425,0.826775,0.84795,0.772425,0.772425,0.2755916666666667,0.826775,0.16959000000000002,0.84795,0.772425,0.8013654166666629,0.8051413888888874,0.8221576910797325,0.808084264541656
512
+ 0.7823613086770982,1100,0.771925,0.824975,0.8465,0.771925,0.771925,0.27499166666666663,0.824975,0.1693,0.8465,0.771925,0.800552916666664,0.8043763392857131,0.8212707180610094,0.8073654553559804
513
+ 0.8534850640113798,1200,0.771025,0.8245,0.8465,0.771025,0.771025,0.2748333333333333,0.8245,0.1693,0.8465,0.771025,0.7997141666666631,0.8034813194444425,0.8205412738249844,0.8064654982641106
514
+ 0.9246088193456614,1300,0.770125,0.823975,0.845875,0.770125,0.770125,0.27465833333333334,0.823975,0.16917500000000005,0.845875,0.770125,0.7990629166666633,0.8028743353174591,0.819989477253076,0.8059157176756445
515
+ 0.9957325746799431,1400,0.770025,0.823775,0.845675,0.770025,0.770025,0.2745916666666667,0.823775,0.169135,0.845675,0.770025,0.7988824999999963,0.8026089186507919,0.8195745782328137,0.8057224194613716
516
+ 1.0668563300142249,1500,0.76935,0.823125,0.845025,0.76935,0.76935,0.274375,0.823125,0.16900500000000002,0.845025,0.76935,0.7982599999999963,0.8021238492063483,0.8192664063697898,0.8052476907087441
517
+ 1.1379800853485065,1600,0.769675,0.824125,0.8459,0.769675,0.769675,0.27470833333333333,0.824125,0.16918000000000002,0.8459,0.769675,0.7988324999999966,0.802651726190475,0.8198127616171903,0.8057819627069863
518
+ 1.209103840682788,1700,0.769925,0.82335,0.845475,0.769925,0.769925,0.27444999999999997,0.82335,0.169095,0.845475,0.769925,0.7985574999999965,0.8024133928571431,0.8195806889185923,0.8055723826945884
519
+ 1.2802275960170697,1800,0.76945,0.823225,0.845675,0.76945,0.76945,0.2744083333333333,0.823225,0.16913499999999998,0.845675,0.76945,0.7984820833333308,0.8023268353174606,0.819532061273888,0.8055067107223939
520
+ 1.3513513513513513,1900,0.7691,0.82345,0.8454,0.7691,0.7691,0.27448333333333336,0.82345,0.16908,0.8454,0.7691,0.7982649999999966,0.8022191071428562,0.8196089924598134,0.805361535028667
521
+ 1.422475106685633,2000,0.76915,0.82355,0.84555,0.76915,0.76915,0.27451666666666663,0.82355,0.16911,0.84555,0.76915,0.7983374999999966,0.8022573412698407,0.8196429775778874,0.8054493205736073
522
+ 1.4935988620199145,2100,0.768975,0.82265,0.846275,0.768975,0.768975,0.27421666666666666,0.82265,0.16925500000000002,0.846275,0.768975,0.7982549999999972,0.8021490277777785,0.8196726605111274,0.8052885037093309
523
+ 1.5647226173541964,2200,0.76885,0.82275,0.845525,0.76885,0.76885,0.27425,0.82275,0.169105,0.845525,0.76885,0.7980941666666648,0.8021060019841282,0.8196099718058493,0.8052948246443948
524
+ 1.635846372688478,2300,0.769525,0.822775,0.84605,0.769525,0.769525,0.2742583333333333,0.822775,0.16921000000000003,0.84605,0.769525,0.7986262499999974,0.8027209226190485,0.8203517201642737,0.8058530753943357
525
+ 1.7069701280227596,2400,0.7692,0.82385,0.846825,0.7692,0.7692,0.2746166666666667,0.82385,0.16936500000000002,0.846825,0.7692,0.7988820833333301,0.8029087698412697,0.8205894446265792,0.8060693271840575
526
+ 1.7780938833570412,2500,0.76915,0.823725,0.846075,0.76915,0.76915,0.27457499999999996,0.823725,0.169215,0.846075,0.76915,0.7985683333333297,0.8026579761904753,0.8203458536798922,0.8058257617189151
527
+ 1.8492176386913228,2600,0.769625,0.8246,0.84655,0.769625,0.769625,0.2748666666666666,0.8246,0.16931000000000002,0.84655,0.769625,0.7989349999999956,0.8030517063492055,0.8208079574534372,0.8062010921954896
528
+ 1.9203413940256047,2700,0.769675,0.82405,0.8468,0.769675,0.769675,0.2746833333333333,0.82405,0.16936,0.8468,0.769675,0.7990249999999968,0.8030730158730155,0.8207917590552776,0.8062321544716866
529
+ 1.991465149359886,2800,0.769475,0.8239,0.8468,0.769475,0.769475,0.27463333333333334,0.8239,0.16936,0.8468,0.769475,0.7988862499999965,0.8029730853174593,0.8208163032289425,0.8061114050965597
530
+ 2.062588904694168,2900,0.769525,0.824125,0.846975,0.769525,0.769525,0.2747083333333333,0.824125,0.16939500000000002,0.846975,0.769525,0.7990287499999966,0.8031705853174578,0.821090801331755,0.8063063194045639
531
+ 2.1337126600284497,3000,0.769025,0.8238,0.846525,0.769025,0.769025,0.2746,0.8238,0.169305,0.846525,0.769025,0.7986137499999971,0.8027588888888897,0.8206359977622566,0.8059580410743369
532
+ 2.204836415362731,3100,0.7688,0.8244,0.846275,0.7688,0.7688,0.2748,0.8244,0.169255,0.846275,0.7688,0.798440833333329,0.8026224603174588,0.820558182322503,0.8058211991919103
533
+ 2.275960170697013,3200,0.769,0.824375,0.847525,0.769,0.769,0.27479166666666666,0.824375,0.169505,0.847525,0.769,0.7988995833333298,0.803090228174601,0.821212946340923,0.8062446723184936
534
+ 2.3470839260312943,3300,0.768775,0.8236,0.84665,0.768775,0.768775,0.2745333333333333,0.8236,0.16933,0.84665,0.768775,0.7983658333333298,0.8026019444444423,0.8207321092229084,0.8057709182568787
535
+ 2.418207681365576,3400,0.7693,0.824175,0.847725,0.7693,0.7693,0.27472499999999994,0.824175,0.16954500000000003,0.847725,0.7693,0.79904708333333,0.8031713095238084,0.8212219006875497,0.80638920392551
536
+ 2.4893314366998576,3500,0.769,0.823825,0.846725,0.769,0.769,0.2746083333333333,0.823825,0.169345,0.846725,0.769,0.7985754166666633,0.8027596726190457,0.8207592889256521,0.8059864347671793
537
+ 2.5604551920341394,3600,0.76885,0.82395,0.8472,0.76885,0.76885,0.27464999999999995,0.82395,0.16944,0.8472,0.76885,0.7986866666666633,0.8028899107142835,0.8210598982328613,0.8060860038886773
538
+ 2.6315789473684212,3700,0.768975,0.82375,0.8473,0.768975,0.768975,0.2745833333333333,0.82375,0.16946000000000003,0.8473,0.768975,0.7987062499999968,0.8029549503968239,0.8212211242661835,0.8061437513212889
539
+ 2.7027027027027026,3800,0.768475,0.823675,0.84715,0.768475,0.768475,0.27455833333333335,0.823675,0.16943000000000003,0.84715,0.768475,0.7984670833333309,0.8026703373015858,0.8208454989370393,0.805943199903395
540
+ 2.7738264580369845,3900,0.76845,0.824325,0.84735,0.76845,0.76845,0.27477499999999994,0.824325,0.16946999999999998,0.84735,0.76845,0.7985499999999965,0.8027500992063477,0.8209005123089202,0.8060343590369505
541
+ 0,0,0.754925,0.807525,0.8308,0.754925,0.754925,0.269175,0.807525,0.16616,0.8308,0.754925,0.7834312499999965,0.7873935019841247,0.8048432733807716,0.7904811958974799
542
+ 0.07112375533428165,100,0.784875,0.8462,0.8709,0.784875,0.784875,0.28206666666666663,0.8462,0.17418000000000003,0.8709,0.784875,0.8176641666666615,0.8220572023809489,0.8415809580549388,0.8251072204917347
543
+ 0.1422475106685633,200,0.798075,0.861975,0.885725,0.798075,0.798075,0.2873249999999999,0.861975,0.17714500000000002,0.885725,0.798075,0.8317829166666606,0.8360811706349159,0.8556872478966354,0.8388044601703539
544
+ 0.21337126600284495,300,0.793925,0.8557,0.88025,0.793925,0.793925,0.2852333333333333,0.8557,0.17605,0.88025,0.793925,0.8268337499999948,0.8308948511904715,0.8500208698553641,0.8336162172452645
545
+ 0.2844950213371266,400,0.786425,0.8466,0.870625,0.786425,0.786425,0.2822,0.8466,0.174125,0.870625,0.786425,0.8185116666666619,0.8224112103174577,0.8409779740110344,0.8251187590711554
546
+ 0.35561877667140823,500,0.780775,0.838675,0.8622,0.780775,0.780775,0.2795583333333333,0.838675,0.17244,0.8622,0.780775,0.811867083333329,0.815695327380951,0.8336883141924025,0.8184760373174987
547
+ 0.4267425320056899,600,0.77655,0.833175,0.856,0.77655,0.77655,0.277725,0.833175,0.1712,0.856,0.77655,0.807010833333329,0.8109420535714272,0.82873317460253,0.8137810123209752
548
+ 0,0,0.75535,0.8081,0.8314,0.75535,0.75535,0.26936666666666664,0.8081,0.16628,0.8314,0.75535,0.7840524999999964,0.7878348511904746,0.8050188990447376,0.7908941824856313
549
+ 0.07112375533428165,100,0.78485,0.845025,0.8707,0.78485,0.78485,0.28167499999999995,0.845025,0.17414000000000002,0.8707,0.78485,0.8174049999999953,0.8215923908730137,0.8408518209280792,0.824681932921971
550
+ 0.1422475106685633,200,0.797725,0.860375,0.885575,0.797725,0.797725,0.28679166666666667,0.860375,0.17711500000000002,0.885575,0.797725,0.831151666666662,0.8353726289682505,0.85499622100629,0.8380966938179938
551
+ 0.21337126600284495,300,0.794025,0.85545,0.88,0.794025,0.794025,0.28514999999999996,0.85545,0.17600000000000005,0.88,0.794025,0.826853749999995,0.830870674603171,0.8498411806165926,0.8336155787941057
552
+ 0.2844950213371266,400,0.7868,0.84585,0.87015,0.7868,0.7868,0.28195,0.84585,0.17403,0.87015,0.7868,0.8186045833333289,0.8224543551587292,0.840768411489289,0.8252239450411709
553
+ 0.35561877667140823,500,0.78115,0.838625,0.86195,0.78115,0.78115,0.27954166666666663,0.838625,0.17239000000000004,0.86195,0.78115,0.8119908333333283,0.8158102579365052,0.8337259437115861,0.8185528436776975
554
+ 0,0,0.75485,0.807325,0.8301,0.75485,0.75485,0.26910833333333334,0.807325,0.16602000000000003,0.8301,0.75485,0.7832499999999958,0.7872831249999973,0.8046637231378345,0.7903968386415016
555
+ 0.07112375533428165,100,0.7846,0.845875,0.870475,0.7846,0.7846,0.2819583333333333,0.845875,0.17409500000000003,0.870475,0.7846,0.8175220833333285,0.8218874404761879,0.8412730946280681,0.8249953488463487
556
+ 0.1422475106685633,200,0.7991,0.86235,0.88605,0.7991,0.7991,0.28745,0.86235,0.17721,0.88605,0.7991,0.8323445833333273,0.8366090079365022,0.856103705751086,0.8393309442404088
557
+ 0.21337126600284495,300,0.794425,0.85565,0.87995,0.794425,0.794425,0.2852166666666666,0.85565,0.17599,0.87995,0.794425,0.8270145833333279,0.8310784623015823,0.8500951169586342,0.8337921512636404
558
+ 0,0,0.75485,0.807325,0.8301,0.75485,0.75485,0.26910833333333334,0.807325,0.16602000000000003,0.8301,0.75485,0.783244583333329,0.7872755357142831,0.8046476992140257,0.7903936979751335
559
+ 0.2849002849002849,100,0.798025,0.8616,0.886625,0.798025,0.798025,0.2872,0.8616,0.177325,0.886625,0.798025,0.8318458333333274,0.8360672718253934,0.8557194444498546,0.8388601667645773
560
+ 0.5698005698005698,200,0.798625,0.860475,0.885875,0.798625,0.798625,0.28682499999999994,0.860475,0.177175,0.885875,0.798625,0.8319029166666622,0.8359366567460295,0.8551078160914656,0.8387170552876636
561
+ 0.8547008547008547,300,0.7974,0.8592,0.884475,0.7974,0.7974,0.28639999999999993,0.8592,0.176895,0.884475,0.7974,0.830667916666662,0.8349036011904735,0.8543353748755151,0.8377224912323109
562
+ 1.1396011396011396,400,0.79655,0.8598,0.8852,0.79655,0.79655,0.28659999999999997,0.8598,0.17704,0.8852,0.79655,0.8303362499999941,0.8345951587301531,0.8543686412006678,0.8374535585472126
563
+ 1.4245014245014245,500,0.797225,0.860075,0.8858,0.797225,0.797225,0.2866916666666667,0.860075,0.17716,0.8858,0.797225,0.8309320833333282,0.8352305456349177,0.8550016319959118,0.8381492588281301
564
+ 1.7094017094017095,600,0.797725,0.86015,0.88665,0.797725,0.797725,0.2867166666666666,0.86015,0.17733000000000002,0.88665,0.797725,0.8313991666666614,0.8356998115079319,0.8555999734107165,0.8385892139129612
565
+ 1.9943019943019942,700,0.797975,0.861325,0.887325,0.797975,0.797975,0.2871083333333333,0.861325,0.177465,0.887325,0.797975,0.8319016666666612,0.8362617460317411,0.8562739250749866,0.8391815253696681
566
+ 2.2792022792022792,800,0.79835,0.86145,0.888175,0.79835,0.79835,0.28715,0.86145,0.17763500000000002,0.888175,0.79835,0.8324391666666625,0.8366843948412661,0.8566247944874054,0.8396564176781772
567
+ 2.564102564102564,900,0.798375,0.862075,0.888225,0.798375,0.798375,0.2873583333333333,0.862075,0.17764500000000003,0.888225,0.798375,0.832563333333328,0.8369297420634868,0.8570087384867151,0.8398948290084045
568
+ 2.849002849002849,1000,0.7987,0.8623,0.88935,0.7987,0.7987,0.28743333333333326,0.8623,0.17787,0.88935,0.7987,0.8329933333333287,0.8371977083333292,0.8572161454797611,0.8402122967157111
569
+ 3.133903133903134,1100,0.7986,0.862325,0.88905,0.7986,0.7986,0.2874416666666666,0.862325,0.17781000000000002,0.88905,0.7986,0.8329541666666616,0.8373066468253925,0.8574478167950113,0.8403082974896099
570
+ 3.4188034188034186,1200,0.799125,0.8628,0.889525,0.799125,0.799125,0.2876,0.8628,0.177905,0.889525,0.799125,0.8334449999999944,0.8377979365079314,0.8579219826325345,0.8408453839836647
571
+ 3.7037037037037037,1300,0.7992,0.862875,0.890425,0.7992,0.7992,0.28762499999999996,0.862875,0.17808500000000005,0.890425,0.7992,0.8336524999999955,0.8379600992063455,0.8582124684478113,0.8409858197750711
572
+ 3.9886039886039883,1400,0.7994,0.86345,0.891,0.7994,0.7994,0.2878166666666666,0.86345,0.17820000000000003,0.891,0.7994,0.8340516666666619,0.8383523908730119,0.8586412473927562,0.8413785327956702
573
+ 4.273504273504273,1500,0.799825,0.863575,0.890875,0.799825,0.799825,0.2878583333333333,0.863575,0.178175,0.890875,0.799825,0.834187083333328,0.8385845238095195,0.8589450124547461,0.8415957957257291
574
+ 4.5584045584045585,1600,0.7999,0.864025,0.89135,0.7999,0.7999,0.2880083333333333,0.864025,0.17827000000000004,0.89135,0.7999,0.8344108333333281,0.8388025892857096,0.8592482613458784,0.841788283131001
575
+ 4.843304843304844,1700,0.799775,0.8643,0.891475,0.799775,0.799775,0.28809999999999997,0.8643,0.17829500000000004,0.891475,0.799775,0.8345333333333287,0.8389044543650747,0.8593056239763895,0.8419345783918155
576
+ 5.128205128205128,1800,0.800375,0.86435,0.8917,0.800375,0.800375,0.28811666666666663,0.86435,0.17834,0.8917,0.800375,0.8348483333333283,0.8392823412698379,0.8596923791808068,0.8423084831810096
577
+ 5.413105413105413,1900,0.800275,0.86425,0.892,0.800275,0.800275,0.2880833333333333,0.86425,0.17840000000000006,0.892,0.800275,0.8349387499999955,0.8393679861111072,0.8598629768476423,0.8423678921112896
578
+ 5.698005698005698,2000,0.80055,0.8648,0.892175,0.80055,0.80055,0.2882666666666666,0.8648,0.178435,0.892175,0.80055,0.8351624999999946,0.8395857440476139,0.8600660936618119,0.8426103480001628
579
+ 5.982905982905983,2100,0.80115,0.864925,0.89215,0.80115,0.80115,0.2883083333333333,0.864925,0.17843,0.89215,0.80115,0.8355258333333282,0.8399495039682501,0.8603280903031238,0.8429954225226167
580
+ 6.267806267806268,2200,0.801225,0.865025,0.892425,0.801225,0.801225,0.28834166666666666,0.865025,0.178485,0.892425,0.801225,0.835632499999995,0.8400607440476151,0.8605022088940468,0.8430806763267096
581
+ 6.552706552706553,2300,0.801275,0.865075,0.892575,0.801275,0.801275,0.2883583333333333,0.865075,0.178515,0.892575,0.801275,0.8357724999999949,0.8402953373015829,0.8608270909507948,0.8433080921002089
582
+ 6.837606837606837,2400,0.801575,0.865125,0.8924,0.801575,0.801575,0.288375,0.865125,0.17848000000000003,0.8924,0.801575,0.8358745833333283,0.840366349206346,0.8607774883891669,0.8434245904693333
583
+ 7.122507122507122,2500,0.80125,0.865,0.892525,0.80125,0.80125,0.2883333333333333,0.865,0.17850500000000002,0.892525,0.80125,0.8357508333333294,0.8403010317460295,0.8608727881432594,0.843316427466535
584
+ 7.407407407407407,2600,0.8017,0.86565,0.893325,0.8017,0.8017,0.28855,0.86565,0.17866500000000002,0.893325,0.8017,0.8362083333333287,0.8406969047619021,0.8612783749710475,0.843708366864761
585
+ 7.6923076923076925,2700,0.8017,0.86565,0.89265,0.8017,0.8017,0.28855,0.86565,0.17853000000000002,0.89265,0.8017,0.8360970833333286,0.8406737103174581,0.8612256926854687,0.8436828252043238
586
+ 7.977207977207978,2800,0.801675,0.865975,0.8929,0.801675,0.801675,0.2886583333333333,0.865975,0.17858000000000002,0.8929,0.801675,0.8361433333333281,0.8407370833333302,0.8613825554424442,0.8437267421268331
587
+ 8.262108262108262,2900,0.801675,0.866225,0.8937,0.801675,0.801675,0.2887416666666667,0.866225,0.17874,0.8937,0.801675,0.8364137499999955,0.84092972222222,0.8616041067360869,0.8439152706764873
588
+ 8.547008547008547,3000,0.801875,0.86645,0.893725,0.801875,0.801875,0.2888166666666666,0.86645,0.17874500000000001,0.893725,0.801875,0.8365145833333281,0.8410595634920605,0.8617667237853555,0.844030886025032
589
+ 8.831908831908832,3100,0.8023,0.86675,0.8937,0.8023,0.8023,0.2889166666666666,0.86675,0.17874,0.8937,0.8023,0.8368633333333286,0.8413887301587282,0.861958616771547,0.8443842822305276
590
+ 9.116809116809117,3200,0.80215,0.8668,0.89375,0.80215,0.80215,0.2889333333333333,0.8668,0.17874999999999996,0.89375,0.80215,0.8368420833333284,0.8413961408730133,0.8620346988055472,0.8443798484797
591
+ 9.401709401709402,3300,0.801675,0.8668,0.893875,0.801675,0.801675,0.2889333333333333,0.8668,0.17877500000000002,0.893875,0.801675,0.8366370833333286,0.8411731646825378,0.8618878022511639,0.8441485770363238
592
+ 9.686609686609687,3400,0.802025,0.867075,0.89435,0.802025,0.802025,0.2890249999999999,0.867075,0.17887000000000003,0.89435,0.802025,0.8369008333333284,0.841364017857141,0.8620247212678003,0.8443709033760651
593
+ 9.971509971509972,3500,0.802225,0.867275,0.89445,0.802225,0.802225,0.2890916666666666,0.867275,0.17889000000000002,0.89445,0.802225,0.8371345833333286,0.8416379861111096,0.8623100216286935,0.8446182487542423
594
+ 10.256410256410255,3600,0.80225,0.8672,0.8945,0.80225,0.80225,0.28906666666666664,0.8672,0.17890000000000003,0.8945,0.80225,0.837151249999995,0.8416526091269821,0.8623240909951686,0.8446431234027594
595
+ 10.54131054131054,3700,0.802125,0.86725,0.894475,0.802125,0.802125,0.2890833333333333,0.86725,0.17889500000000003,0.894475,0.802125,0.8371087499999949,0.8416043948412673,0.8622908336950419,0.844597619678181
596
+ 10.826210826210826,3800,0.8023,0.867175,0.8947,0.8023,0.8023,0.2890583333333333,0.867175,0.17894000000000002,0.8947,0.8023,0.8372020833333284,0.841680109126981,0.8623660731654591,0.8446758053360729
597
+ 11.11111111111111,3900,0.8024,0.866975,0.8947,0.8024,0.8024,0.2889916666666666,0.866975,0.17894000000000002,0.8947,0.8024,0.8372370833333284,0.8417220138888862,0.8624066595516806,0.8447202630510894
598
+ 11.396011396011396,4000,0.802475,0.867225,0.8946,0.802475,0.802475,0.289075,0.867225,0.17892,0.8946,0.802475,0.8373216666666616,0.8418482242063464,0.8625582027110184,0.84483463585437
599
+ 11.68091168091168,4100,0.802425,0.867075,0.894575,0.802425,0.802425,0.289025,0.867075,0.17891500000000002,0.894575,0.802425,0.837263749999995,0.8417595833333303,0.8624108110281115,0.844785572665587
600
+ 11.965811965811966,4200,0.802525,0.86745,0.894625,0.802525,0.802525,0.28914999999999996,0.86745,0.17892500000000003,0.894625,0.802525,0.8373899999999946,0.8418941269841238,0.8625485306060944,0.8449125742356473
601
+ 12.250712250712251,4300,0.802425,0.867425,0.8949,0.802425,0.802425,0.28914166666666663,0.867425,0.17898,0.8949,0.802425,0.8373679166666612,0.8418560416666631,0.862570562826382,0.8448589991636967
602
+ 12.535612535612536,4400,0.802475,0.8674,0.8947,0.802475,0.802475,0.2891333333333333,0.8674,0.17894000000000004,0.8947,0.802475,0.8373841666666616,0.8419126488095208,0.8626053598219531,0.8449282197722164
603
+ 12.820512820512821,4500,0.802325,0.86755,0.89495,0.802325,0.802325,0.2891833333333333,0.86755,0.17899,0.89495,0.802325,0.8373616666666612,0.8418324404761867,0.8625349588157627,0.8448507681783248
604
+ 13.105413105413106,4600,0.80245,0.867575,0.8949,0.80245,0.80245,0.2891916666666666,0.867575,0.17898000000000003,0.8949,0.80245,0.8374437499999949,0.8419402380952342,0.8626571238261973,0.8449448072816027
605
+ 13.39031339031339,4700,0.80245,0.867575,0.894825,0.80245,0.80245,0.2891916666666666,0.867575,0.178965,0.894825,0.80245,0.8374074999999948,0.841924206349203,0.8626414521760948,0.8449332980096382
606
+ 13.675213675213675,4800,0.8024,0.867775,0.8947,0.8024,0.8024,0.2892583333333333,0.867775,0.17894000000000004,0.8947,0.8024,0.837397083333328,0.8419506051587267,0.8627007599790489,0.844947246350743
607
+ 13.96011396011396,4900,0.802475,0.867625,0.8947,0.802475,0.802475,0.2892083333333333,0.867625,0.17894,0.8947,0.802475,0.8374054166666614,0.8419520734126953,0.8626792917708476,0.8449600866335961
608
+ 14.245014245014245,5000,0.802275,0.8676,0.89495,0.802275,0.802275,0.28919999999999996,0.8676,0.17899,0.89495,0.802275,0.8373504166666614,0.8418555952380918,0.8626142815266079,0.8448604581747029
final_metrics.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "val_cosine_accuracy@1": 0.829275,
3
- "val_cosine_accuracy@3": 0.9051,
4
- "val_cosine_accuracy@5": 0.9329,
5
- "val_cosine_precision@1": 0.829275,
6
- "val_cosine_precision@3": 0.30169999999999997,
7
- "val_cosine_precision@5": 0.18658000000000002,
8
- "val_cosine_recall@1": 0.829275,
9
- "val_cosine_recall@3": 0.9051,
10
- "val_cosine_recall@5": 0.9329,
11
- "val_cosine_ndcg@10": 0.8956869608914538,
12
- "val_cosine_mrr@1": 0.829275,
13
- "val_cosine_mrr@5": 0.8692179166666618,
14
- "val_cosine_mrr@10": 0.8735753373015815,
15
- "val_cosine_map@100": 0.8756452160249361
16
  }
 
1
  {
2
+ "val_cosine_accuracy@1": 0.76415,
3
+ "val_cosine_accuracy@3": 0.822175,
4
+ "val_cosine_accuracy@5": 0.84665,
5
+ "val_cosine_precision@1": 0.76415,
6
+ "val_cosine_precision@3": 0.2740583333333333,
7
+ "val_cosine_precision@5": 0.16933,
8
+ "val_cosine_recall@1": 0.76415,
9
+ "val_cosine_recall@3": 0.822175,
10
+ "val_cosine_recall@5": 0.84665,
11
+ "val_cosine_ndcg@10": 0.8195085862842543,
12
+ "val_cosine_mrr@1": 0.76415,
13
+ "val_cosine_mrr@5": 0.7953370833333295,
14
+ "val_cosine_mrr@10": 0.8000376587301573,
15
+ "val_cosine_map@100": 0.8034052122492592
16
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47776c0a310f3e5373785410310661fe1f5da06c5321098634df02df9cb53755
3
  size 114011616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45cdb91cc553556968b278c530594e05d926d8a2c99d01838656167e2114a507
3
  size 114011616
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:767f865ad778a3cf6f884deef6446e71dd8ddc58b4697f307d5df2876e65b529
3
  size 6161
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7258c121488b2e4966f22215c04593a6fcbb89f836828ccbf066defc2eda8b93
3
  size 6161