Training in progress, step 14060
Browse files- Information-Retrieval_evaluation_val_results.csv +1 -0
- README.md +71 -233
- eval/Information-Retrieval_evaluation_val_results.csv +141 -0
- final_metrics.json +14 -14
- model.safetensors +1 -1
- training_args.bin +1 -1
Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -1,2 +1,3 @@
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.908,0.9684,0.9834,0.908,0.908,0.3228,0.9684,0.19667999999999997,0.9834,0.908,0.9386633333333337,0.9400269841269848,0.9532296698470627,0.9404621256346036
|
|
|
|
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.908,0.9684,0.9834,0.908,0.908,0.3228,0.9684,0.19667999999999997,0.9834,0.908,0.9386633333333337,0.9400269841269848,0.9532296698470627,0.9404621256346036
|
| 3 |
+
-1,-1,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402433333333332,0.9416250793650793,0.9545809774353143,0.9420576026548708
|
README.md
CHANGED
|
@@ -5,108 +5,38 @@ tags:
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
-
- dataset_size:
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
-
- source_sentence: How do I
|
| 13 |
sentences:
|
| 14 |
-
-
|
| 15 |
-
- How do I
|
| 16 |
-
- What
|
| 17 |
-
- source_sentence:
|
| 18 |
sentences:
|
| 19 |
-
-
|
| 20 |
-
- What
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
- source_sentence:
|
| 24 |
sentences:
|
| 25 |
-
-
|
| 26 |
-
-
|
| 27 |
-
-
|
| 28 |
-
- source_sentence: What are
|
| 29 |
sentences:
|
| 30 |
-
- What are
|
| 31 |
-
-
|
| 32 |
-
- What are some
|
| 33 |
-
- source_sentence: What
|
| 34 |
-
at Opus Bank?
|
| 35 |
sentences:
|
| 36 |
-
-
|
| 37 |
-
|
| 38 |
-
-
|
| 39 |
-
Bank ?
|
| 40 |
-
- What are some tips on making it through the job interview process at Opus Bank?
|
| 41 |
pipeline_tag: sentence-similarity
|
| 42 |
library_name: sentence-transformers
|
| 43 |
-
metrics:
|
| 44 |
-
- cosine_accuracy@1
|
| 45 |
-
- cosine_accuracy@3
|
| 46 |
-
- cosine_accuracy@5
|
| 47 |
-
- cosine_precision@1
|
| 48 |
-
- cosine_precision@3
|
| 49 |
-
- cosine_precision@5
|
| 50 |
-
- cosine_recall@1
|
| 51 |
-
- cosine_recall@3
|
| 52 |
-
- cosine_recall@5
|
| 53 |
-
- cosine_ndcg@10
|
| 54 |
-
- cosine_mrr@1
|
| 55 |
-
- cosine_mrr@5
|
| 56 |
-
- cosine_mrr@10
|
| 57 |
-
- cosine_map@100
|
| 58 |
-
model-index:
|
| 59 |
-
- name: SentenceTransformer based on prajjwal1/bert-small
|
| 60 |
-
results:
|
| 61 |
-
- task:
|
| 62 |
-
type: information-retrieval
|
| 63 |
-
name: Information Retrieval
|
| 64 |
-
dataset:
|
| 65 |
-
name: val
|
| 66 |
-
type: val
|
| 67 |
-
metrics:
|
| 68 |
-
- type: cosine_accuracy@1
|
| 69 |
-
value: 0.9104
|
| 70 |
-
name: Cosine Accuracy@1
|
| 71 |
-
- type: cosine_accuracy@3
|
| 72 |
-
value: 0.9688
|
| 73 |
-
name: Cosine Accuracy@3
|
| 74 |
-
- type: cosine_accuracy@5
|
| 75 |
-
value: 0.9842
|
| 76 |
-
name: Cosine Accuracy@5
|
| 77 |
-
- type: cosine_precision@1
|
| 78 |
-
value: 0.9104
|
| 79 |
-
name: Cosine Precision@1
|
| 80 |
-
- type: cosine_precision@3
|
| 81 |
-
value: 0.32293333333333335
|
| 82 |
-
name: Cosine Precision@3
|
| 83 |
-
- type: cosine_precision@5
|
| 84 |
-
value: 0.19683999999999996
|
| 85 |
-
name: Cosine Precision@5
|
| 86 |
-
- type: cosine_recall@1
|
| 87 |
-
value: 0.9104
|
| 88 |
-
name: Cosine Recall@1
|
| 89 |
-
- type: cosine_recall@3
|
| 90 |
-
value: 0.9688
|
| 91 |
-
name: Cosine Recall@3
|
| 92 |
-
- type: cosine_recall@5
|
| 93 |
-
value: 0.9842
|
| 94 |
-
name: Cosine Recall@5
|
| 95 |
-
- type: cosine_ndcg@10
|
| 96 |
-
value: 0.954585167414727
|
| 97 |
-
name: Cosine Ndcg@10
|
| 98 |
-
- type: cosine_mrr@1
|
| 99 |
-
value: 0.9104
|
| 100 |
-
name: Cosine Mrr@1
|
| 101 |
-
- type: cosine_mrr@5
|
| 102 |
-
value: 0.9402533333333333
|
| 103 |
-
name: Cosine Mrr@5
|
| 104 |
-
- type: cosine_mrr@10
|
| 105 |
-
value: 0.9416303174603176
|
| 106 |
-
name: Cosine Mrr@10
|
| 107 |
-
- type: cosine_map@100
|
| 108 |
-
value: 0.9420641228013908
|
| 109 |
-
name: Cosine Map@100
|
| 110 |
---
|
| 111 |
|
| 112 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
@@ -155,12 +85,12 @@ Then you can load this model and run inference.
|
|
| 155 |
from sentence_transformers import SentenceTransformer
|
| 156 |
|
| 157 |
# Download from the 🤗 Hub
|
| 158 |
-
model = SentenceTransformer("
|
| 159 |
# Run inference
|
| 160 |
sentences = [
|
| 161 |
-
'What
|
| 162 |
-
'
|
| 163 |
-
'
|
| 164 |
]
|
| 165 |
embeddings = model.encode(sentences)
|
| 166 |
print(embeddings.shape)
|
|
@@ -169,9 +99,9 @@ print(embeddings.shape)
|
|
| 169 |
# Get the similarity scores for the embeddings
|
| 170 |
similarities = model.similarity(embeddings, embeddings)
|
| 171 |
print(similarities)
|
| 172 |
-
# tensor([[1.0000,
|
| 173 |
-
# [
|
| 174 |
-
# [0.
|
| 175 |
```
|
| 176 |
|
| 177 |
<!--
|
|
@@ -198,32 +128,6 @@ You can finetune this model on your own dataset.
|
|
| 198 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 199 |
-->
|
| 200 |
|
| 201 |
-
## Evaluation
|
| 202 |
-
|
| 203 |
-
### Metrics
|
| 204 |
-
|
| 205 |
-
#### Information Retrieval
|
| 206 |
-
|
| 207 |
-
* Dataset: `val`
|
| 208 |
-
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
| 209 |
-
|
| 210 |
-
| Metric | Value |
|
| 211 |
-
|:-------------------|:-----------|
|
| 212 |
-
| cosine_accuracy@1 | 0.9104 |
|
| 213 |
-
| cosine_accuracy@3 | 0.9688 |
|
| 214 |
-
| cosine_accuracy@5 | 0.9842 |
|
| 215 |
-
| cosine_precision@1 | 0.9104 |
|
| 216 |
-
| cosine_precision@3 | 0.3229 |
|
| 217 |
-
| cosine_precision@5 | 0.1968 |
|
| 218 |
-
| cosine_recall@1 | 0.9104 |
|
| 219 |
-
| cosine_recall@3 | 0.9688 |
|
| 220 |
-
| cosine_recall@5 | 0.9842 |
|
| 221 |
-
| **cosine_ndcg@10** | **0.9546** |
|
| 222 |
-
| cosine_mrr@1 | 0.9104 |
|
| 223 |
-
| cosine_mrr@5 | 0.9403 |
|
| 224 |
-
| cosine_mrr@10 | 0.9416 |
|
| 225 |
-
| cosine_map@100 | 0.9421 |
|
| 226 |
-
|
| 227 |
<!--
|
| 228 |
## Bias, Risks and Limitations
|
| 229 |
|
|
@@ -242,45 +146,19 @@ You can finetune this model on your own dataset.
|
|
| 242 |
|
| 243 |
#### Unnamed Dataset
|
| 244 |
|
| 245 |
-
* Size:
|
| 246 |
-
* Columns: <code>
|
| 247 |
-
* Approximate statistics based on the first 1000 samples:
|
| 248 |
-
| | anchor | positive | negative |
|
| 249 |
-
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
|
| 250 |
-
| type | string | string | string |
|
| 251 |
-
| details | <ul><li>min: 6 tokens</li><li>mean: 15.63 tokens</li><li>max: 75 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.77 tokens</li><li>max: 75 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.2 tokens</li><li>max: 75 tokens</li></ul> |
|
| 252 |
-
* Samples:
|
| 253 |
-
| anchor | positive | negative |
|
| 254 |
-
|:---------------------------------------------------------|:---------------------------------------------------------|:----------------------------------------------------------------------------|
|
| 255 |
-
| <code>How long did it take to develop Pokémon GO?</code> | <code>How long did it take to develop Pokémon GO?</code> | <code>Can I take more than one gym in Pokémon GO?</code> |
|
| 256 |
-
| <code>How bad is 6/18 eyesight?</code> | <code>How bad is 6/18 eyesight?</code> | <code>How was bad eyesight dealt with in ancient and medieval times?</code> |
|
| 257 |
-
| <code>How can I do learn speaking English easily?</code> | <code>How can I learn speaking English easily?</code> | <code>How can English do learn speaking Ieasily?</code> |
|
| 258 |
-
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 259 |
-
```json
|
| 260 |
-
{
|
| 261 |
-
"scale": 20.0,
|
| 262 |
-
"similarity_fct": "cos_sim",
|
| 263 |
-
"gather_across_devices": false
|
| 264 |
-
}
|
| 265 |
-
```
|
| 266 |
-
|
| 267 |
-
### Evaluation Dataset
|
| 268 |
-
|
| 269 |
-
#### Unnamed Dataset
|
| 270 |
-
|
| 271 |
-
* Size: 5,000 evaluation samples
|
| 272 |
-
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
| 273 |
* Approximate statistics based on the first 1000 samples:
|
| 274 |
-
| |
|
| 275 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
| 276 |
| type | string | string | string |
|
| 277 |
-
| details | <ul><li>min: 6 tokens</li><li>mean: 15.
|
| 278 |
* Samples:
|
| 279 |
-
|
|
| 280 |
-
|
| 281 |
-
| <code>
|
| 282 |
-
| <code>
|
| 283 |
-
| <code>
|
| 284 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 285 |
```json
|
| 286 |
{
|
|
@@ -293,49 +171,36 @@ You can finetune this model on your own dataset.
|
|
| 293 |
### Training Hyperparameters
|
| 294 |
#### Non-Default Hyperparameters
|
| 295 |
|
| 296 |
-
- `
|
| 297 |
-
- `
|
| 298 |
-
- `per_device_eval_batch_size`: 256
|
| 299 |
-
- `learning_rate`: 2e-05
|
| 300 |
-
- `weight_decay`: 0.001
|
| 301 |
-
- `max_steps`: 3510
|
| 302 |
-
- `warmup_ratio`: 0.1
|
| 303 |
- `fp16`: True
|
| 304 |
-
- `
|
| 305 |
-
- `dataloader_num_workers`: 1
|
| 306 |
-
- `dataloader_prefetch_factor`: 1
|
| 307 |
-
- `load_best_model_at_end`: True
|
| 308 |
-
- `optim`: adamw_torch
|
| 309 |
-
- `ddp_find_unused_parameters`: False
|
| 310 |
-
- `push_to_hub`: True
|
| 311 |
-
- `hub_model_id`: redis/model-b-structured
|
| 312 |
-
- `eval_on_start`: True
|
| 313 |
|
| 314 |
#### All Hyperparameters
|
| 315 |
<details><summary>Click to expand</summary>
|
| 316 |
|
| 317 |
- `overwrite_output_dir`: False
|
| 318 |
- `do_predict`: False
|
| 319 |
-
- `eval_strategy`:
|
| 320 |
- `prediction_loss_only`: True
|
| 321 |
-
- `per_device_train_batch_size`:
|
| 322 |
-
- `per_device_eval_batch_size`:
|
| 323 |
- `per_gpu_train_batch_size`: None
|
| 324 |
- `per_gpu_eval_batch_size`: None
|
| 325 |
- `gradient_accumulation_steps`: 1
|
| 326 |
- `eval_accumulation_steps`: None
|
| 327 |
- `torch_empty_cache_steps`: None
|
| 328 |
-
- `learning_rate`:
|
| 329 |
-
- `weight_decay`: 0.
|
| 330 |
- `adam_beta1`: 0.9
|
| 331 |
- `adam_beta2`: 0.999
|
| 332 |
- `adam_epsilon`: 1e-08
|
| 333 |
-
- `max_grad_norm`: 1
|
| 334 |
-
- `num_train_epochs`: 3
|
| 335 |
-
- `max_steps`:
|
| 336 |
- `lr_scheduler_type`: linear
|
| 337 |
- `lr_scheduler_kwargs`: {}
|
| 338 |
-
- `warmup_ratio`: 0.
|
| 339 |
- `warmup_steps`: 0
|
| 340 |
- `log_level`: passive
|
| 341 |
- `log_level_replica`: warning
|
|
@@ -363,14 +228,14 @@ You can finetune this model on your own dataset.
|
|
| 363 |
- `tpu_num_cores`: None
|
| 364 |
- `tpu_metrics_debug`: False
|
| 365 |
- `debug`: []
|
| 366 |
-
- `dataloader_drop_last`:
|
| 367 |
-
- `dataloader_num_workers`:
|
| 368 |
-
- `dataloader_prefetch_factor`:
|
| 369 |
- `past_index`: -1
|
| 370 |
- `disable_tqdm`: False
|
| 371 |
- `remove_unused_columns`: True
|
| 372 |
- `label_names`: None
|
| 373 |
-
- `load_best_model_at_end`:
|
| 374 |
- `ignore_data_skip`: False
|
| 375 |
- `fsdp`: []
|
| 376 |
- `fsdp_min_num_params`: 0
|
|
@@ -380,23 +245,23 @@ You can finetune this model on your own dataset.
|
|
| 380 |
- `parallelism_config`: None
|
| 381 |
- `deepspeed`: None
|
| 382 |
- `label_smoothing_factor`: 0.0
|
| 383 |
-
- `optim`:
|
| 384 |
- `optim_args`: None
|
| 385 |
- `adafactor`: False
|
| 386 |
- `group_by_length`: False
|
| 387 |
- `length_column_name`: length
|
| 388 |
- `project`: huggingface
|
| 389 |
- `trackio_space_id`: trackio
|
| 390 |
-
- `ddp_find_unused_parameters`:
|
| 391 |
- `ddp_bucket_cap_mb`: None
|
| 392 |
- `ddp_broadcast_buffers`: False
|
| 393 |
- `dataloader_pin_memory`: True
|
| 394 |
- `dataloader_persistent_workers`: False
|
| 395 |
- `skip_memory_metrics`: True
|
| 396 |
- `use_legacy_prediction_loop`: False
|
| 397 |
-
- `push_to_hub`:
|
| 398 |
- `resume_from_checkpoint`: None
|
| 399 |
-
- `hub_model_id`:
|
| 400 |
- `hub_strategy`: every_save
|
| 401 |
- `hub_private_repo`: None
|
| 402 |
- `hub_always_push`: False
|
|
@@ -423,58 +288,31 @@ You can finetune this model on your own dataset.
|
|
| 423 |
- `neftune_noise_alpha`: None
|
| 424 |
- `optim_target_modules`: None
|
| 425 |
- `batch_eval_metrics`: False
|
| 426 |
-
- `eval_on_start`:
|
| 427 |
- `use_liger_kernel`: False
|
| 428 |
- `liger_kernel_config`: None
|
| 429 |
- `eval_use_gather_object`: False
|
| 430 |
- `average_tokens_across_devices`: True
|
| 431 |
- `prompts`: None
|
| 432 |
- `batch_sampler`: batch_sampler
|
| 433 |
-
- `multi_dataset_batch_sampler`:
|
| 434 |
- `router_mapping`: {}
|
| 435 |
- `learning_rate_mapping`: {}
|
| 436 |
|
| 437 |
</details>
|
| 438 |
|
| 439 |
### Training Logs
|
| 440 |
-
| Epoch | Step | Training Loss |
|
| 441 |
-
|
| 442 |
-
| 0
|
| 443 |
-
| 0.
|
| 444 |
-
| 0.
|
| 445 |
-
|
|
| 446 |
-
| 1.
|
| 447 |
-
| 1.
|
| 448 |
-
|
|
| 449 |
-
|
|
| 450 |
-
| 2.
|
| 451 |
-
| 2.5641 | 900 | 0.1221 | 0.0863 | 0.9505 |
|
| 452 |
-
| 2.8490 | 1000 | 0.1124 | 0.0833 | 0.9512 |
|
| 453 |
-
| 3.1339 | 1100 | 0.1116 | 0.0816 | 0.9514 |
|
| 454 |
-
| 3.4188 | 1200 | 0.1019 | 0.0808 | 0.9522 |
|
| 455 |
-
| 3.7037 | 1300 | 0.1007 | 0.0784 | 0.9522 |
|
| 456 |
-
| 3.9886 | 1400 | 0.1009 | 0.0766 | 0.9525 |
|
| 457 |
-
| 4.2735 | 1500 | 0.0889 | 0.0759 | 0.9533 |
|
| 458 |
-
| 4.5584 | 1600 | 0.0891 | 0.0748 | 0.9536 |
|
| 459 |
-
| 4.8433 | 1700 | 0.0874 | 0.0734 | 0.9534 |
|
| 460 |
-
| 5.1282 | 1800 | 0.0856 | 0.0721 | 0.9539 |
|
| 461 |
-
| 5.4131 | 1900 | 0.082 | 0.0715 | 0.9544 |
|
| 462 |
-
| 5.6980 | 2000 | 0.0821 | 0.0704 | 0.9540 |
|
| 463 |
-
| 5.9829 | 2100 | 0.0804 | 0.0699 | 0.9540 |
|
| 464 |
-
| 6.2678 | 2200 | 0.076 | 0.0694 | 0.9544 |
|
| 465 |
-
| 6.5527 | 2300 | 0.0729 | 0.0693 | 0.9546 |
|
| 466 |
-
| 6.8376 | 2400 | 0.0758 | 0.0688 | 0.9547 |
|
| 467 |
-
| 7.1225 | 2500 | 0.0728 | 0.0673 | 0.9547 |
|
| 468 |
-
| 7.4074 | 2600 | 0.0707 | 0.0678 | 0.9546 |
|
| 469 |
-
| 7.6923 | 2700 | 0.0695 | 0.0678 | 0.9549 |
|
| 470 |
-
| 7.9772 | 2800 | 0.0716 | 0.0671 | 0.9546 |
|
| 471 |
-
| 8.2621 | 2900 | 0.0709 | 0.0668 | 0.9544 |
|
| 472 |
-
| 8.5470 | 3000 | 0.0692 | 0.0671 | 0.9551 |
|
| 473 |
-
| 8.8319 | 3100 | 0.0694 | 0.0666 | 0.9547 |
|
| 474 |
-
| 9.1168 | 3200 | 0.0713 | 0.0667 | 0.9547 |
|
| 475 |
-
| 9.4017 | 3300 | 0.0682 | 0.0665 | 0.9546 |
|
| 476 |
-
| 9.6866 | 3400 | 0.07 | 0.0664 | 0.9547 |
|
| 477 |
-
| 9.9715 | 3500 | 0.07 | 0.0664 | 0.9546 |
|
| 478 |
|
| 479 |
|
| 480 |
### Framework Versions
|
|
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
+
- dataset_size:100000
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
+
- source_sentence: How do I calculate IQ?
|
| 13 |
sentences:
|
| 14 |
+
- What is the easiest way to know my IQ?
|
| 15 |
+
- How do I calculate not IQ ?
|
| 16 |
+
- What are some creative and innovative business ideas with less investment in India?
|
| 17 |
+
- source_sentence: How can I learn martial arts in my home?
|
| 18 |
sentences:
|
| 19 |
+
- How can I learn martial arts by myself?
|
| 20 |
+
- What are the advantages and disadvantages of investing in gold?
|
| 21 |
+
- Can people see that I have looked at their pictures on instagram if I am not following
|
| 22 |
+
them?
|
| 23 |
+
- source_sentence: When Enterprise picks you up do you have to take them back?
|
| 24 |
sentences:
|
| 25 |
+
- Are there any software Training institute in Tuticorin?
|
| 26 |
+
- When Enterprise picks you up do you have to take them back?
|
| 27 |
+
- When Enterprise picks you up do them have to take youback?
|
| 28 |
+
- source_sentence: What are some non-capital goods?
|
| 29 |
sentences:
|
| 30 |
+
- What are capital goods?
|
| 31 |
+
- How is the value of [math]\pi[/math] calculated?
|
| 32 |
+
- What are some non-capital goods?
|
| 33 |
+
- source_sentence: What is the QuickBooks technical support phone number in New York?
|
|
|
|
| 34 |
sentences:
|
| 35 |
+
- What caused the Great Depression?
|
| 36 |
+
- Can I apply for PR in Canada?
|
| 37 |
+
- Which is the best QuickBooks Hosting Support Number in New York?
|
|
|
|
|
|
|
| 38 |
pipeline_tag: sentence-similarity
|
| 39 |
library_name: sentence-transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
---
|
| 41 |
|
| 42 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
|
|
| 85 |
from sentence_transformers import SentenceTransformer
|
| 86 |
|
| 87 |
# Download from the 🤗 Hub
|
| 88 |
+
model = SentenceTransformer("sentence_transformers_model_id")
|
| 89 |
# Run inference
|
| 90 |
sentences = [
|
| 91 |
+
'What is the QuickBooks technical support phone number in New York?',
|
| 92 |
+
'Which is the best QuickBooks Hosting Support Number in New York?',
|
| 93 |
+
'Can I apply for PR in Canada?',
|
| 94 |
]
|
| 95 |
embeddings = model.encode(sentences)
|
| 96 |
print(embeddings.shape)
|
|
|
|
| 99 |
# Get the similarity scores for the embeddings
|
| 100 |
similarities = model.similarity(embeddings, embeddings)
|
| 101 |
print(similarities)
|
| 102 |
+
# tensor([[1.0000, 0.8563, 0.0594],
|
| 103 |
+
# [0.8563, 1.0000, 0.1245],
|
| 104 |
+
# [0.0594, 0.1245, 1.0000]])
|
| 105 |
```
|
| 106 |
|
| 107 |
<!--
|
|
|
|
| 128 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 129 |
-->
|
| 130 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
<!--
|
| 132 |
## Bias, Risks and Limitations
|
| 133 |
|
|
|
|
| 146 |
|
| 147 |
#### Unnamed Dataset
|
| 148 |
|
| 149 |
+
* Size: 100,000 training samples
|
| 150 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 151 |
* Approximate statistics based on the first 1000 samples:
|
| 152 |
+
| | sentence_0 | sentence_1 | sentence_2 |
|
| 153 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
| 154 |
| type | string | string | string |
|
| 155 |
+
| details | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.68 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 16.37 tokens</li><li>max: 67 tokens</li></ul> |
|
| 156 |
* Samples:
|
| 157 |
+
| sentence_0 | sentence_1 | sentence_2 |
|
| 158 |
+
|:-----------------------------------------------------------------|:-----------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
| 159 |
+
| <code>Is masturbating bad for boys?</code> | <code>Is masturbating bad for boys?</code> | <code>How harmful or unhealthy is masturbation?</code> |
|
| 160 |
+
| <code>Does a train engine move in reverse?</code> | <code>Does a train engine move in reverse?</code> | <code>Time moves forward, not in reverse. Doesn't that make time a vector?</code> |
|
| 161 |
+
| <code>What is the most badass thing anyone has ever done?</code> | <code>What is the most badass thing anyone has ever done?</code> | <code>anyone is the most badass thing Whathas ever done?</code> |
|
| 162 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 163 |
```json
|
| 164 |
{
|
|
|
|
| 171 |
### Training Hyperparameters
|
| 172 |
#### Non-Default Hyperparameters
|
| 173 |
|
| 174 |
+
- `per_device_train_batch_size`: 64
|
| 175 |
+
- `per_device_eval_batch_size`: 64
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 176 |
- `fp16`: True
|
| 177 |
+
- `multi_dataset_batch_sampler`: round_robin
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 178 |
|
| 179 |
#### All Hyperparameters
|
| 180 |
<details><summary>Click to expand</summary>
|
| 181 |
|
| 182 |
- `overwrite_output_dir`: False
|
| 183 |
- `do_predict`: False
|
| 184 |
+
- `eval_strategy`: no
|
| 185 |
- `prediction_loss_only`: True
|
| 186 |
+
- `per_device_train_batch_size`: 64
|
| 187 |
+
- `per_device_eval_batch_size`: 64
|
| 188 |
- `per_gpu_train_batch_size`: None
|
| 189 |
- `per_gpu_eval_batch_size`: None
|
| 190 |
- `gradient_accumulation_steps`: 1
|
| 191 |
- `eval_accumulation_steps`: None
|
| 192 |
- `torch_empty_cache_steps`: None
|
| 193 |
+
- `learning_rate`: 5e-05
|
| 194 |
+
- `weight_decay`: 0.0
|
| 195 |
- `adam_beta1`: 0.9
|
| 196 |
- `adam_beta2`: 0.999
|
| 197 |
- `adam_epsilon`: 1e-08
|
| 198 |
+
- `max_grad_norm`: 1
|
| 199 |
+
- `num_train_epochs`: 3
|
| 200 |
+
- `max_steps`: -1
|
| 201 |
- `lr_scheduler_type`: linear
|
| 202 |
- `lr_scheduler_kwargs`: {}
|
| 203 |
+
- `warmup_ratio`: 0.0
|
| 204 |
- `warmup_steps`: 0
|
| 205 |
- `log_level`: passive
|
| 206 |
- `log_level_replica`: warning
|
|
|
|
| 228 |
- `tpu_num_cores`: None
|
| 229 |
- `tpu_metrics_debug`: False
|
| 230 |
- `debug`: []
|
| 231 |
+
- `dataloader_drop_last`: False
|
| 232 |
+
- `dataloader_num_workers`: 0
|
| 233 |
+
- `dataloader_prefetch_factor`: None
|
| 234 |
- `past_index`: -1
|
| 235 |
- `disable_tqdm`: False
|
| 236 |
- `remove_unused_columns`: True
|
| 237 |
- `label_names`: None
|
| 238 |
+
- `load_best_model_at_end`: False
|
| 239 |
- `ignore_data_skip`: False
|
| 240 |
- `fsdp`: []
|
| 241 |
- `fsdp_min_num_params`: 0
|
|
|
|
| 245 |
- `parallelism_config`: None
|
| 246 |
- `deepspeed`: None
|
| 247 |
- `label_smoothing_factor`: 0.0
|
| 248 |
+
- `optim`: adamw_torch_fused
|
| 249 |
- `optim_args`: None
|
| 250 |
- `adafactor`: False
|
| 251 |
- `group_by_length`: False
|
| 252 |
- `length_column_name`: length
|
| 253 |
- `project`: huggingface
|
| 254 |
- `trackio_space_id`: trackio
|
| 255 |
+
- `ddp_find_unused_parameters`: None
|
| 256 |
- `ddp_bucket_cap_mb`: None
|
| 257 |
- `ddp_broadcast_buffers`: False
|
| 258 |
- `dataloader_pin_memory`: True
|
| 259 |
- `dataloader_persistent_workers`: False
|
| 260 |
- `skip_memory_metrics`: True
|
| 261 |
- `use_legacy_prediction_loop`: False
|
| 262 |
+
- `push_to_hub`: False
|
| 263 |
- `resume_from_checkpoint`: None
|
| 264 |
+
- `hub_model_id`: None
|
| 265 |
- `hub_strategy`: every_save
|
| 266 |
- `hub_private_repo`: None
|
| 267 |
- `hub_always_push`: False
|
|
|
|
| 288 |
- `neftune_noise_alpha`: None
|
| 289 |
- `optim_target_modules`: None
|
| 290 |
- `batch_eval_metrics`: False
|
| 291 |
+
- `eval_on_start`: False
|
| 292 |
- `use_liger_kernel`: False
|
| 293 |
- `liger_kernel_config`: None
|
| 294 |
- `eval_use_gather_object`: False
|
| 295 |
- `average_tokens_across_devices`: True
|
| 296 |
- `prompts`: None
|
| 297 |
- `batch_sampler`: batch_sampler
|
| 298 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 299 |
- `router_mapping`: {}
|
| 300 |
- `learning_rate_mapping`: {}
|
| 301 |
|
| 302 |
</details>
|
| 303 |
|
| 304 |
### Training Logs
|
| 305 |
+
| Epoch | Step | Training Loss |
|
| 306 |
+
|:------:|:----:|:-------------:|
|
| 307 |
+
| 0.3199 | 500 | 0.4294 |
|
| 308 |
+
| 0.6398 | 1000 | 0.1268 |
|
| 309 |
+
| 0.9597 | 1500 | 0.1 |
|
| 310 |
+
| 1.2796 | 2000 | 0.0792 |
|
| 311 |
+
| 1.5995 | 2500 | 0.0706 |
|
| 312 |
+
| 1.9194 | 3000 | 0.0687 |
|
| 313 |
+
| 2.2393 | 3500 | 0.0584 |
|
| 314 |
+
| 2.5592 | 4000 | 0.057 |
|
| 315 |
+
| 2.8791 | 4500 | 0.0581 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 316 |
|
| 317 |
|
| 318 |
### Framework Versions
|
eval/Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -38,3 +38,144 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
|
|
| 38 |
9.401709401709402,3300,0.9106,0.969,0.984,0.9106,0.9106,0.323,0.969,0.19679999999999995,0.984,0.9106,0.9403533333333333,0.9417519841269842,0.9546344737257203,0.9422058505967718
|
| 39 |
9.686609686609687,3400,0.9106,0.969,0.9838,0.9106,0.9106,0.323,0.969,0.19675999999999996,0.9838,0.9106,0.94029,0.9417384126984129,0.9546651626751027,0.9421697455120135
|
| 40 |
9.971509971509972,3500,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402533333333333,0.9416303174603176,0.954585167414727,0.9420641228013908
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
9.401709401709402,3300,0.9106,0.969,0.984,0.9106,0.9106,0.323,0.969,0.19679999999999995,0.984,0.9106,0.9403533333333333,0.9417519841269842,0.9546344737257203,0.9422058505967718
|
| 39 |
9.686609686609687,3400,0.9106,0.969,0.9838,0.9106,0.9106,0.323,0.969,0.19675999999999996,0.9838,0.9106,0.94029,0.9417384126984129,0.9546651626751027,0.9421697455120135
|
| 40 |
9.971509971509972,3500,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402533333333333,0.9416303174603176,0.954585167414727,0.9420641228013908
|
| 41 |
+
0,0,0.708475,0.796825,0.822125,0.708475,0.708475,0.26560833333333334,0.796825,0.164425,0.822125,0.708475,0.7545720833333283,0.7589311706349164,0.7821232471271422,0.762190078448873
|
| 42 |
+
0.07112375533428165,100,0.722325,0.835475,0.864475,0.722325,0.722325,0.27849166666666664,0.835475,0.17289500000000002,0.864475,0.722325,0.7807887499999926,0.7853636309523765,0.8129700440800849,0.788695746119173
|
| 43 |
+
0.1422475106685633,200,0.775525,0.865825,0.893425,0.775525,0.775525,0.2886083333333333,0.865825,0.178685,0.893425,0.775525,0.8226208333333254,0.8270209226190415,0.8510344343464452,0.8297694974089365
|
| 44 |
+
0.21337126600284495,300,0.7953,0.8738,0.898525,0.7953,0.7953,0.2912666666666666,0.8738,0.17970500000000003,0.898525,0.7953,0.8362499999999936,0.8402300595238037,0.8614711546093078,0.8429135761695575
|
| 45 |
+
0.2844950213371266,400,0.80075,0.875975,0.8999,0.80075,0.80075,0.29199166666666665,0.875975,0.17998000000000003,0.8999,0.80075,0.8400066666666598,0.844234613095233,0.8652126035164182,0.8469108693511953
|
| 46 |
+
0.35561877667140823,500,0.804275,0.87805,0.90305,0.804275,0.804275,0.29268333333333324,0.87805,0.18061000000000002,0.90305,0.804275,0.8431766666666619,0.8472814087301547,0.8680716987185689,0.849959335480741
|
| 47 |
+
0.4267425320056899,600,0.807025,0.88005,0.9059,0.807025,0.807025,0.29334999999999994,0.88005,0.18118,0.9059,0.807025,0.8457383333333277,0.8498279365079326,0.8706979594584189,0.8524284487021079
|
| 48 |
+
0.49786628733997157,700,0.80925,0.88155,0.9079,0.80925,0.80925,0.29384999999999994,0.88155,0.18158000000000002,0.9079,0.80925,0.8476441666666613,0.8517574107142803,0.8726681323345749,0.8543576926337788
|
| 49 |
+
0.5689900426742532,800,0.8109,0.883625,0.9099,0.8109,0.8109,0.2945416666666666,0.883625,0.18198000000000003,0.9099,0.8109,0.8495879166666622,0.8536860714285669,0.8745685072106874,0.8563012372130501
|
| 50 |
+
0.6401137980085349,900,0.81195,0.885025,0.910475,0.81195,0.81195,0.29500833333333326,0.885025,0.18209500000000003,0.910475,0.81195,0.850385416666661,0.8545485218253918,0.8754790911316688,0.8571486995111011
|
| 51 |
+
0.7112375533428165,1000,0.8134,0.886325,0.911625,0.8134,0.8134,0.2954416666666666,0.886325,0.182325,0.911625,0.8134,0.851712916666662,0.8559180257936456,0.8769045253521254,0.858461313040403
|
| 52 |
+
0.7823613086770982,1100,0.814875,0.886975,0.9132,0.814875,0.814875,0.2956583333333333,0.886975,0.18264,0.9132,0.814875,0.8530883333333288,0.8571790972222177,0.8780328288384593,0.8597917042367506
|
| 53 |
+
0.8534850640113798,1200,0.815225,0.8876,0.914625,0.815225,0.815225,0.2958666666666666,0.8876,0.182925,0.914625,0.815225,0.8536179166666618,0.8577433928571372,0.8788102573542441,0.8603604705654514
|
| 54 |
+
0.9246088193456614,1300,0.8168,0.88965,0.916,0.8168,0.8168,0.29655,0.88965,0.18320000000000003,0.916,0.8168,0.8553495833333291,0.8594912698412667,0.8805103132423321,0.8620812371800759
|
| 55 |
+
0.9957325746799431,1400,0.81755,0.8904,0.91635,0.81755,0.81755,0.29679999999999995,0.8904,0.18327000000000002,0.91635,0.81755,0.8559383333333279,0.8601746230158671,0.8812821892817625,0.8627088755223402
|
| 56 |
+
1.0668563300142249,1500,0.818025,0.890825,0.91855,0.818025,0.818025,0.29694166666666666,0.890825,0.18371000000000004,0.91855,0.818025,0.8568879166666625,0.8609777380952331,0.8821959740808727,0.8634983104603688
|
| 57 |
+
1.1379800853485065,1600,0.818,0.8911,0.9188,0.818,0.818,0.29703333333333326,0.8911,0.18376,0.9188,0.818,0.8568829166666623,0.8609047619047576,0.8821272443292048,0.8634440919229286
|
| 58 |
+
1.209103840682788,1700,0.819175,0.89195,0.9181,0.819175,0.819175,0.2973166666666667,0.89195,0.18362000000000003,0.9181,0.819175,0.8575737499999945,0.8617549107142799,0.882844384353523,0.8642749205132398
|
| 59 |
+
1.2802275960170697,1800,0.8194,0.8924,0.919725,0.8194,0.8194,0.29746666666666666,0.8924,0.18394500000000003,0.919725,0.8194,0.8581379166666618,0.8622184722222166,0.8834552221414264,0.8647068630314819
|
| 60 |
+
1.3513513513513513,1900,0.8191,0.8923,0.9191,0.8191,0.8191,0.2974333333333333,0.8923,0.18382000000000004,0.9191,0.8191,0.8577912499999948,0.8620219146825335,0.8833214194983335,0.8645483869343061
|
| 61 |
+
1.422475106685633,2000,0.8196,0.892925,0.919325,0.8196,0.8196,0.2976416666666666,0.892925,0.183865,0.919325,0.8196,0.8582274999999945,0.8625463194444392,0.8839596520387931,0.8650054306218091
|
| 62 |
+
1.4935988620199145,2100,0.819925,0.893125,0.9203,0.819925,0.819925,0.2977083333333333,0.893125,0.18406,0.9203,0.819925,0.8587529166666618,0.8630120535714239,0.88446025808528,0.8654665677460687
|
| 63 |
+
1.5647226173541964,2200,0.820125,0.894225,0.920375,0.820125,0.820125,0.298075,0.894225,0.18407500000000002,0.920375,0.820125,0.8592241666666613,0.8636013293650736,0.8851266057760023,0.8659959512491583
|
| 64 |
+
1.635846372688478,2300,0.820075,0.8932,0.9209,0.820075,0.820075,0.29773333333333324,0.8932,0.18418,0.9209,0.820075,0.8588995833333279,0.8632591369047553,0.8849475053016227,0.8656661752985969
|
| 65 |
+
1.7069701280227596,2400,0.82185,0.8952,0.9231,0.82185,0.82185,0.29839999999999994,0.8952,0.18462000000000003,0.9231,0.82185,0.8608804166666614,0.8651157341269778,0.8866948372363526,0.8674770134243758
|
| 66 |
+
1.7780938833570412,2500,0.82185,0.8954,0.923,0.82185,0.82185,0.2984666666666666,0.8954,0.18460000000000001,0.923,0.82185,0.8608999999999958,0.8652058432539638,0.8868494073657219,0.8675460895736036
|
| 67 |
+
1.8492176386913228,2600,0.8215,0.895275,0.92175,0.8215,0.8215,0.298425,0.895275,0.18435000000000004,0.92175,0.8215,0.8603341666666618,0.8646135416666603,0.8860526209198091,0.8670679286069957
|
| 68 |
+
1.9203413940256047,2700,0.820625,0.8945,0.922,0.820625,0.820625,0.2981666666666666,0.8945,0.18440000000000004,0.922,0.820625,0.8598466666666612,0.8641913293650736,0.8859156872252449,0.8665781551131037
|
| 69 |
+
1.991465149359886,2800,0.822025,0.895975,0.922925,0.822025,0.822025,0.2986583333333333,0.895975,0.18458500000000003,0.922925,0.822025,0.8611374999999949,0.8654293253968208,0.8869782054645499,0.8678350492219843
|
| 70 |
+
2.062588904694168,2900,0.822375,0.8974,0.9234,0.822375,0.822375,0.2991333333333333,0.8974,0.18468000000000004,0.9234,0.822375,0.8616691666666622,0.8659512003968208,0.8874814593992038,0.868350202612051
|
| 71 |
+
2.1337126600284497,3000,0.82195,0.8957,0.92255,0.82195,0.82195,0.2985666666666666,0.8957,0.18451000000000004,0.92255,0.82195,0.8608883333333277,0.8652523214285655,0.8868455935332514,0.8676710098840409
|
| 72 |
+
2.204836415362731,3100,0.822625,0.8968,0.924125,0.822625,0.822625,0.29893333333333333,0.8968,0.184825,0.924125,0.822625,0.8619379166666613,0.8662695436507878,0.8879605042451206,0.868627403228108
|
| 73 |
+
2.275960170697013,3200,0.822275,0.896475,0.92305,0.822275,0.822275,0.29882499999999995,0.896475,0.18461000000000002,0.92305,0.822275,0.861370416666661,0.8657300198412631,0.8873368206534243,0.8681311777983121
|
| 74 |
+
2.3470839260312943,3300,0.822575,0.8966,0.923925,0.822575,0.822575,0.29886666666666667,0.8966,0.18478500000000003,0.923925,0.822575,0.8617070833333279,0.8659898313492007,0.8876205256797713,0.868391252660771
|
| 75 |
+
2.418207681365576,3400,0.823025,0.89705,0.924575,0.823025,0.823025,0.2990166666666666,0.89705,0.184915,0.924575,0.823025,0.8622604166666619,0.8665290773809465,0.8881827229716455,0.8689210814093723
|
| 76 |
+
2.4893314366998576,3500,0.823625,0.897725,0.925475,0.823625,0.823625,0.29924166666666663,0.897725,0.18509500000000004,0.925475,0.823625,0.8629520833333283,0.8672388888888832,0.8889497027095765,0.8695685936615529
|
| 77 |
+
2.5604551920341394,3600,0.8235,0.89685,0.924175,0.8235,0.8235,0.2989499999999999,0.89685,0.184835,0.924175,0.8235,0.8622354166666617,0.8666268154761853,0.888341415540098,0.8689953073597675
|
| 78 |
+
2.6315789473684212,3700,0.82355,0.897425,0.92415,0.82355,0.82355,0.2991416666666666,0.897425,0.18483000000000005,0.92415,0.82355,0.8624670833333292,0.8669097222222183,0.8886308654439339,0.8692619959122819
|
| 79 |
+
2.7027027027027026,3800,0.8238,0.897675,0.924725,0.8238,0.8238,0.29922499999999996,0.897675,0.18494500000000005,0.924725,0.8238,0.8628454166666616,0.8672871428571367,0.8891080403797096,0.8695675304567493
|
| 80 |
+
2.7738264580369845,3900,0.823575,0.897225,0.924975,0.823575,0.823575,0.2990749999999999,0.897225,0.18499500000000002,0.924975,0.823575,0.8627349999999951,0.8671197817460253,0.8889427948375415,0.8694255979358242
|
| 81 |
+
2.844950213371266,4000,0.8246,0.898625,0.9255,0.8246,0.8246,0.2995416666666666,0.898625,0.1851,0.9255,0.8246,0.8636449999999944,0.8680026785714217,0.8896766870518819,0.8703092429631565
|
| 82 |
+
2.9160739687055477,4100,0.82475,0.8984,0.925175,0.82475,0.82475,0.29946666666666666,0.8984,0.18503500000000003,0.925175,0.82475,0.863664166666661,0.8680787599206262,0.8897809878315085,0.8703838923006196
|
| 83 |
+
2.987197724039829,4200,0.824475,0.898775,0.925925,0.824475,0.824475,0.2995916666666666,0.898775,0.18518500000000002,0.925925,0.824475,0.8637179166666615,0.8681762301587244,0.8900659161989499,0.8704322254975827
|
| 84 |
+
3.058321479374111,4300,0.823775,0.898125,0.925725,0.823775,0.823775,0.2993749999999999,0.898125,0.18514500000000003,0.925725,0.823775,0.8631258333333284,0.8674574305555496,0.889263149093025,0.8697791518408887
|
| 85 |
+
3.1294452347083928,4400,0.82425,0.8994,0.92655,0.82425,0.82425,0.2997999999999999,0.8994,0.18531,0.92655,0.82425,0.8638191666666611,0.8681279563491986,0.8899985654190258,0.8704186236514423
|
| 86 |
+
3.200568990042674,4500,0.824375,0.897525,0.925025,0.824375,0.824375,0.299175,0.897525,0.18500500000000003,0.925025,0.824375,0.8632137499999953,0.867598730158724,0.8893111625023677,0.869929135071786
|
| 87 |
+
3.271692745376956,4600,0.824775,0.8987,0.926425,0.824775,0.824775,0.2995666666666666,0.8987,0.185285,0.926425,0.824775,0.863976666666661,0.868326567460311,0.8901824492164305,0.8705952707897266
|
| 88 |
+
3.3428165007112374,4700,0.82425,0.89855,0.92595,0.82425,0.82425,0.29951666666666665,0.89855,0.18519000000000002,0.92595,0.82425,0.8635716666666605,0.8679974404761835,0.889916074777261,0.870291257332588
|
| 89 |
+
3.413940256045519,4800,0.824425,0.8992,0.926925,0.824425,0.824425,0.2997333333333333,0.8992,0.18538500000000005,0.926925,0.824425,0.8640862499999946,0.8683873015872947,0.8902968657758403,0.8706472282678136
|
| 90 |
+
3.485064011379801,4900,0.8255,0.900225,0.9274,0.8255,0.8255,0.300075,0.900225,0.18548,0.9274,0.8255,0.864929583333328,0.8693004464285641,0.8911555528211864,0.8715296867059648
|
| 91 |
+
3.5561877667140824,5000,0.826175,0.9004,0.92745,0.826175,0.826175,0.30013333333333325,0.9004,0.18549000000000002,0.92745,0.826175,0.8653699999999938,0.8697670436507864,0.8915337858893878,0.8719912796724624
|
| 92 |
+
3.6273115220483643,5100,0.825575,0.89855,0.926725,0.825575,0.825575,0.2995166666666666,0.89855,0.18534500000000004,0.926725,0.825575,0.864612083333328,0.8690407142857088,0.8908701769082038,0.8712962428055211
|
| 93 |
+
3.6984352773826457,5200,0.8252,0.899325,0.92635,0.8252,0.8252,0.299775,0.899325,0.18527000000000002,0.92635,0.8252,0.8644154166666612,0.8688983134920578,0.89076275197231,0.8711567837909996
|
| 94 |
+
3.7695590327169275,5300,0.82605,0.900025,0.927225,0.82605,0.82605,0.3000083333333333,0.900025,0.18544500000000005,0.927225,0.82605,0.8653387499999955,0.8697499603174547,0.8915223711969326,0.871992943283452
|
| 95 |
+
3.8406827880512093,5400,0.826275,0.9004,0.92755,0.826275,0.826275,0.3001333333333333,0.9004,0.18551000000000004,0.92755,0.826275,0.865512916666662,0.8699264781745971,0.8917181202681815,0.8721455201676681
|
| 96 |
+
3.9118065433854907,5500,0.825775,0.899925,0.927325,0.825775,0.825775,0.29997499999999994,0.899925,0.18546500000000002,0.927325,0.825775,0.8650508333333288,0.8694735218253918,0.8913486775145258,0.8716941997380186
|
| 97 |
+
3.9829302987197726,5600,0.826125,0.89995,0.927225,0.826125,0.826125,0.2999833333333333,0.89995,0.18544500000000003,0.927225,0.826125,0.865341249999995,0.8697941666666599,0.89162530417659,0.8720109405246621
|
| 98 |
+
4.054054054054054,5700,0.826475,0.90025,0.9278,0.826475,0.826475,0.3000833333333333,0.90025,0.18556,0.9278,0.826475,0.8656508333333292,0.8699955456349148,0.8917350022910326,0.8722532738352884
|
| 99 |
+
4.125177809388336,5800,0.826225,0.899475,0.927,0.826225,0.826225,0.299825,0.899475,0.1854,0.927,0.826225,0.8651508333333295,0.869614156746027,0.89143847123497,0.8718547533641309
|
| 100 |
+
4.196301564722617,5900,0.825725,0.89975,0.9276,0.825725,0.825725,0.2999166666666666,0.89975,0.18552000000000002,0.9276,0.825725,0.8650741666666627,0.8695133035714238,0.8914381678978238,0.8717376488223952
|
| 101 |
+
4.2674253200568995,6000,0.825175,0.89985,0.927075,0.825175,0.825175,0.2999499999999999,0.89985,0.18541500000000002,0.927075,0.825175,0.8646229166666631,0.8690451587301552,0.8909441732114654,0.8713219430870955
|
| 102 |
+
4.338549075391181,6100,0.826325,0.9,0.928075,0.826325,0.826325,0.29999999999999993,0.9,0.18561500000000006,0.928075,0.826325,0.8655245833333285,0.8699142559523754,0.8918191631811841,0.8721377155219839
|
| 103 |
+
4.409672830725462,6200,0.825775,0.9,0.9274,0.825775,0.825775,0.3,0.9,0.18548,0.9274,0.825775,0.865164999999996,0.8696455753968206,0.8915460140151276,0.8718917068937276
|
| 104 |
+
4.480796586059744,6300,0.82585,0.900625,0.928175,0.82585,0.82585,0.3002083333333333,0.900625,0.18563500000000002,0.928175,0.82585,0.8654541666666612,0.8698656249999948,0.8918297577207097,0.8720836684391886
|
| 105 |
+
4.551920341394026,6400,0.8262,0.901,0.9284,0.8262,0.8262,0.3003333333333333,0.901,0.18568000000000004,0.9284,0.8262,0.8656649999999949,0.870096775793644,0.8921152393297598,0.8722803359988235
|
| 106 |
+
4.623044096728307,6500,0.825825,0.900225,0.92815,0.825825,0.825825,0.300075,0.900225,0.18563000000000002,0.92815,0.825825,0.8654254166666627,0.8698505357142811,0.8918221979339549,0.8720642323644188
|
| 107 |
+
4.694167852062589,6600,0.826825,0.9006,0.928325,0.826825,0.826825,0.30019999999999997,0.9006,0.18566500000000002,0.928325,0.826825,0.8660116666666614,0.8704902480158675,0.8924644328014245,0.8726653432105024
|
| 108 |
+
4.76529160739687,6700,0.82635,0.900575,0.928125,0.82635,0.82635,0.3001916666666666,0.900575,0.185625,0.928125,0.82635,0.8656424999999954,0.8701348214285656,0.8921800055683209,0.8723029664572237
|
| 109 |
+
4.836415362731152,6800,0.825825,0.90015,0.927925,0.825825,0.825825,0.30004999999999993,0.90015,0.18558500000000003,0.927925,0.825825,0.8652804166666618,0.8697154662698348,0.8916889085785515,0.8719573734033452
|
| 110 |
+
4.907539118065434,6900,0.825875,0.9007,0.9283,0.825875,0.825875,0.3002333333333333,0.9007,0.18566000000000005,0.9283,0.825875,0.8655537499999949,0.8700121825396756,0.8920441415559691,0.8722158863715503
|
| 111 |
+
4.978662873399715,7000,0.82585,0.90085,0.92835,0.82585,0.82585,0.30028333333333324,0.90085,0.18567000000000003,0.92835,0.82585,0.8656495833333278,0.8700978373015802,0.8921329634415721,0.8722989989753108
|
| 112 |
+
5.049786628733997,7100,0.82615,0.900125,0.928025,0.82615,0.82615,0.3000416666666666,0.900125,0.18560500000000002,0.928025,0.82615,0.8655370833333286,0.8699936607142797,0.8919609460457947,0.8722228385351324
|
| 113 |
+
5.120910384068279,7200,0.8266,0.90045,0.928625,0.8266,0.8266,0.30015,0.90045,0.18572500000000003,0.928625,0.8266,0.8659883333333286,0.8704057440476135,0.8923903227770039,0.8726006052318324
|
| 114 |
+
5.19203413940256,7300,0.826275,0.900525,0.928325,0.826275,0.826275,0.30017499999999997,0.900525,0.18566500000000005,0.928325,0.826275,0.865821666666663,0.8703211111111067,0.8923685679772724,0.8725058346472185
|
| 115 |
+
5.2631578947368425,7400,0.82715,0.900825,0.9287,0.82715,0.82715,0.30027499999999996,0.900825,0.18574000000000004,0.9287,0.82715,0.866372499999995,0.8708181845238036,0.8927230306601641,0.873016518689878
|
| 116 |
+
5.334281650071124,7500,0.8269,0.9007,0.92865,0.8269,0.8269,0.3002333333333333,0.9007,0.18573,0.92865,0.8269,0.8661345833333295,0.870578591269836,0.8925413608695965,0.8727783195270331
|
| 117 |
+
5.405405405405405,7600,0.8273,0.902125,0.92945,0.8273,0.8273,0.3007083333333333,0.902125,0.18589000000000003,0.92945,0.8273,0.866819999999996,0.8712253273809459,0.8931676895510297,0.8734009300232072
|
| 118 |
+
5.476529160739687,7700,0.8261,0.90025,0.9283,0.8261,0.8261,0.3000833333333333,0.90025,0.18566000000000005,0.9283,0.8261,0.865633333333329,0.8700428869047561,0.8919629226946234,0.8722891306212861
|
| 119 |
+
5.547652916073969,7800,0.826975,0.9014,0.929225,0.826975,0.826975,0.30046666666666666,0.9014,0.18584499999999998,0.929225,0.826975,0.8665362499999958,0.8709029761904703,0.8928165883433447,0.8731121920648562
|
| 120 |
+
5.61877667140825,7900,0.827225,0.9016,0.92965,0.827225,0.827225,0.3005333333333333,0.9016,0.18593,0.92965,0.827225,0.8667216666666622,0.8710348511904691,0.8929259824926039,0.8732431056526635
|
| 121 |
+
5.689900426742532,8000,0.827325,0.901575,0.9293,0.827325,0.827325,0.30052499999999993,0.901575,0.18586000000000003,0.9293,0.827325,0.8667749999999959,0.8711936111111052,0.8931338078257104,0.8733803433982779
|
| 122 |
+
5.761024182076814,8100,0.827325,0.901325,0.9294,0.827325,0.827325,0.30044166666666666,0.901325,0.18588000000000005,0.9294,0.827325,0.8668224999999962,0.8711923611111059,0.8930914853077098,0.8733914561201167
|
| 123 |
+
5.832147937411095,8200,0.8271,0.90135,0.929575,0.8271,0.8271,0.30045,0.90135,0.18591500000000002,0.929575,0.8271,0.8665966666666624,0.870971617063485,0.8929716232701749,0.8731473958399926
|
| 124 |
+
5.903271692745377,8300,0.827375,0.901275,0.929575,0.827375,0.827375,0.30042499999999994,0.901275,0.18591500000000005,0.929575,0.827375,0.8667629166666622,0.8711837003968187,0.8931756127653813,0.8733501042216866
|
| 125 |
+
5.974395448079658,8400,0.827075,0.9014,0.929225,0.827075,0.827075,0.3004666666666666,0.9014,0.18584500000000004,0.929225,0.827075,0.8664962499999962,0.8709454166666614,0.8929605309745368,0.8731388106861432
|
| 126 |
+
6.0455192034139404,8500,0.826925,0.901175,0.929275,0.826925,0.826925,0.30039166666666667,0.901175,0.18585500000000002,0.929275,0.826925,0.8663720833333296,0.8708128869047569,0.8928670137625475,0.8729984483900762
|
| 127 |
+
6.116642958748222,8600,0.827375,0.901275,0.929575,0.827375,0.827375,0.3004249999999999,0.901275,0.18591500000000005,0.929575,0.827375,0.8666758333333299,0.8710950198412644,0.8931014973905476,0.8732727102367857
|
| 128 |
+
6.187766714082503,8700,0.827325,0.901275,0.929825,0.827325,0.827325,0.300425,0.901275,0.18596500000000002,0.929825,0.827325,0.8667737499999961,0.8711758432539642,0.8932032489373747,0.8733509090687692
|
| 129 |
+
6.2588904694167855,8800,0.8273,0.901075,0.929525,0.8273,0.8273,0.3003583333333333,0.901075,0.18590500000000001,0.929525,0.8273,0.8666895833333296,0.8711485615079311,0.8931887967740513,0.8733185112174393
|
| 130 |
+
6.330014224751067,8900,0.8273,0.902075,0.929975,0.8273,0.8273,0.3006916666666666,0.902075,0.185995,0.929975,0.8273,0.866963333333329,0.8713748511904703,0.8934143433391218,0.8735371248005948
|
| 131 |
+
6.401137980085348,9000,0.82755,0.9019,0.929925,0.82755,0.82755,0.3006333333333333,0.9019,0.18598500000000004,0.929925,0.82755,0.8669908333333295,0.8713906249999939,0.8933946978262389,0.8735523391943721
|
| 132 |
+
6.472261735419631,9100,0.827075,0.90195,0.9295,0.827075,0.827075,0.30065,0.90195,0.1859,0.9295,0.827075,0.8666599999999954,0.8711001884920575,0.8931327166161569,0.873279831908762
|
| 133 |
+
6.543385490753912,9200,0.827275,0.901325,0.929325,0.827275,0.827275,0.3004416666666666,0.901325,0.185865,0.929325,0.827275,0.866724999999996,0.8711729761904704,0.893137421919378,0.8733594546877983
|
| 134 |
+
6.614509246088193,9300,0.82805,0.9021,0.93015,0.82805,0.82805,0.30069999999999997,0.9021,0.18603000000000003,0.93015,0.82805,0.8674033333333282,0.8718179761904695,0.8937801242746259,0.8739678233587836
|
| 135 |
+
6.685633001422475,9400,0.8274,0.90145,0.92975,0.8274,0.8274,0.3004833333333333,0.90145,0.18595000000000003,0.92975,0.8274,0.8669149999999956,0.8713803968253898,0.8934607392567395,0.8735101273721
|
| 136 |
+
6.756756756756757,9500,0.827425,0.90195,0.93005,0.827425,0.827425,0.30065,0.90195,0.18601000000000004,0.93005,0.827425,0.8670141666666629,0.8714513888888825,0.8935497802488053,0.8735801816011066
|
| 137 |
+
6.827880512091038,9600,0.827225,0.90175,0.9299,0.827225,0.827225,0.3005833333333333,0.90175,0.18598,0.9299,0.827225,0.8668416666666624,0.8713365873015803,0.8935022790918725,0.8734559739901633
|
| 138 |
+
6.89900426742532,9700,0.827375,0.90225,0.9302,0.827375,0.827375,0.30074999999999996,0.90225,0.18604000000000004,0.9302,0.827375,0.8670516666666624,0.8714862400793584,0.8935805344348627,0.8736238052627762
|
| 139 |
+
6.970128022759602,9800,0.82675,0.90165,0.930025,0.82675,0.82675,0.30055,0.90165,0.18600500000000003,0.930025,0.82675,0.8666212499999965,0.8710792857142798,0.893291732812247,0.8732089879292773
|
| 140 |
+
7.0412517780938835,9900,0.827025,0.90265,0.9307,0.827025,0.827025,0.3008833333333333,0.90265,0.18614,0.9307,0.827025,0.8670595833333293,0.871396656746025,0.893538239528841,0.8735295595199406
|
| 141 |
+
7.112375533428165,10000,0.827425,0.902575,0.930475,0.827425,0.827425,0.3008583333333333,0.902575,0.18609500000000004,0.930475,0.827425,0.8672554166666627,0.8716640079365022,0.8937487432845319,0.873803083516281
|
| 142 |
+
7.183499288762446,10100,0.8271,0.90185,0.93035,0.8271,0.8271,0.3006166666666666,0.90185,0.18607000000000004,0.93035,0.8271,0.866940833333329,0.8713414087301533,0.893460804923355,0.8734842714542749
|
| 143 |
+
7.2546230440967285,10200,0.8273,0.9024,0.930825,0.8273,0.8273,0.3008,0.9024,0.18616500000000002,0.930825,0.8273,0.8672354166666637,0.871576527777773,0.8936706222925621,0.8737201708523414
|
| 144 |
+
7.32574679943101,10300,0.827,0.90195,0.93035,0.827,0.827,0.30065,0.90195,0.18607000000000004,0.93035,0.827,0.8668745833333292,0.8712707142857095,0.8933687994725479,0.8734377113367171
|
| 145 |
+
7.396870554765291,10400,0.827425,0.90205,0.93045,0.827425,0.827425,0.3006833333333333,0.90205,0.18609000000000006,0.93045,0.827425,0.8671599999999958,0.8715786904761852,0.8936834412141746,0.8737168495708162
|
| 146 |
+
7.467994310099574,10500,0.827675,0.902275,0.930375,0.827675,0.827675,0.3007583333333333,0.902275,0.18607500000000002,0.930375,0.827675,0.8671958333333292,0.8715664880952326,0.8935959272616397,0.8737339779395145
|
| 147 |
+
7.539118065433855,10600,0.827525,0.902075,0.930825,0.827525,0.827525,0.3006916666666666,0.902075,0.186165,0.930825,0.827525,0.8673195833333289,0.8716814186507876,0.8937695384748613,0.8738217383831259
|
| 148 |
+
7.610241820768136,10700,0.826825,0.901825,0.930125,0.826825,0.826825,0.3006083333333333,0.901825,0.186025,0.930125,0.826825,0.8666412499999949,0.8710677281745962,0.8932242969670978,0.8732328510015733
|
| 149 |
+
7.681365576102419,10800,0.827125,0.9014,0.9303,0.827125,0.827125,0.3004666666666666,0.9014,0.18606000000000003,0.9303,0.827125,0.8668729166666626,0.8712897321428514,0.8934112707000093,0.8734470132059551
|
| 150 |
+
7.7524893314367,10900,0.827725,0.90215,0.930775,0.827725,0.827725,0.30071666666666663,0.90215,0.18615500000000001,0.930775,0.827725,0.8674787499999947,0.8718391567460251,0.8938481523794622,0.8740012494161281
|
| 151 |
+
7.823613086770981,11000,0.827725,0.902775,0.931,0.827725,0.827725,0.3009249999999999,0.902775,0.18620000000000003,0.931,0.827725,0.8675758333333291,0.871905257936502,0.8939239746475901,0.8740667873515862
|
| 152 |
+
7.894736842105263,11100,0.8274,0.9023,0.9305,0.8274,0.8274,0.3007666666666666,0.9023,0.18610000000000002,0.9305,0.8274,0.8671412499999951,0.8715991567460251,0.8937906785238047,0.8737041181091459
|
| 153 |
+
7.965860597439545,11200,0.82765,0.902725,0.93115,0.82765,0.82765,0.3009083333333333,0.902725,0.18623,0.93115,0.82765,0.867559166666662,0.8719196825396771,0.8940223198634536,0.874045803227272
|
| 154 |
+
8.036984352773827,11300,0.827575,0.9023,0.9309,0.827575,0.827575,0.3007666666666666,0.9023,0.18618,0.9309,0.827575,0.8673533333333288,0.8717183928571373,0.8937985230658438,0.8738695292087821
|
| 155 |
+
8.108108108108109,11400,0.8272,0.902325,0.9307,0.8272,0.8272,0.3007749999999999,0.902325,0.18614,0.9307,0.8272,0.8671591666666624,0.8715653869047552,0.8937392409414204,0.8736849934539683
|
| 156 |
+
8.17923186344239,11500,0.827425,0.902375,0.93065,0.827425,0.827425,0.3007916666666666,0.902375,0.18613000000000002,0.93065,0.827425,0.8672470833333283,0.8716846726190411,0.8938569816776198,0.8737965294795397
|
| 157 |
+
8.250355618776672,11600,0.826975,0.902,0.93055,0.826975,0.826975,0.3006666666666666,0.902,0.18611000000000003,0.93055,0.826975,0.8669816666666622,0.8713896130952322,0.8935380269606285,0.8735380533496878
|
| 158 |
+
8.321479374110954,11700,0.82735,0.902175,0.93045,0.82735,0.82735,0.30072499999999996,0.902175,0.18609000000000006,0.93045,0.82735,0.8672008333333284,0.8716245535714215,0.8937331752233314,0.8737676385625377
|
| 159 |
+
8.392603129445234,11800,0.827325,0.902075,0.930825,0.827325,0.827325,0.3006916666666666,0.902075,0.18616500000000002,0.930825,0.827325,0.8672229166666626,0.8716189682539626,0.8937900710030335,0.8737401825146364
|
| 160 |
+
8.463726884779517,11900,0.82825,0.902375,0.93095,0.82825,0.82825,0.3007916666666666,0.902375,0.18619000000000002,0.93095,0.82825,0.867789999999996,0.8722008035714228,0.8942588368608063,0.8743165391515858
|
| 161 |
+
8.534850640113799,12000,0.828025,0.902375,0.93065,0.828025,0.828025,0.3007916666666666,0.902375,0.18613000000000002,0.93065,0.828025,0.8675908333333291,0.8720352281745967,0.8941107124359962,0.8741616170195275
|
| 162 |
+
8.60597439544808,12100,0.8274,0.90205,0.93065,0.8274,0.8274,0.3006833333333333,0.90205,0.18613000000000005,0.93065,0.8274,0.8671479166666625,0.8715756051587236,0.8937531540939795,0.8737020154783539
|
| 163 |
+
8.677098150782362,12200,0.8279,0.902575,0.9311,0.8279,0.8279,0.3008583333333333,0.902575,0.18622000000000002,0.9311,0.8279,0.8676112499999951,0.8719609920634847,0.8940504954192067,0.8740880305749279
|
| 164 |
+
8.748221906116642,12300,0.828025,0.9026,0.93095,0.828025,0.828025,0.3008666666666666,0.9026,0.18619000000000002,0.93095,0.828025,0.8676454166666625,0.872042718253962,0.8941141499515919,0.8741736270226745
|
| 165 |
+
8.819345661450924,12400,0.827675,0.902525,0.930725,0.827675,0.827675,0.3008416666666666,0.902525,0.18614500000000003,0.930725,0.827675,0.8674170833333291,0.8718653571428513,0.8940080486660721,0.8739854046109397
|
| 166 |
+
8.890469416785207,12500,0.828175,0.90275,0.930875,0.828175,0.828175,0.3009166666666666,0.90275,0.18617500000000003,0.930875,0.828175,0.8677533333333289,0.872185367063486,0.8942613640979699,0.874300699597779
|
| 167 |
+
8.961593172119487,12600,0.828025,0.902575,0.930725,0.828025,0.828025,0.3008583333333333,0.902575,0.18614500000000003,0.930725,0.828025,0.8675987499999955,0.8720321726190418,0.8941092062004434,0.874160024782015
|
| 168 |
+
9.03271692745377,12700,0.8279,0.9022,0.930825,0.8279,0.8279,0.30073333333333324,0.9022,0.18616500000000005,0.930825,0.8279,0.8675487499999955,0.8719490773809458,0.8940326075238234,0.874081080000685
|
| 169 |
+
9.103840682788052,12800,0.8276,0.9022,0.930775,0.8276,0.8276,0.3007333333333333,0.9022,0.18615500000000001,0.930775,0.8276,0.867366249999996,0.8717785813492008,0.8939194902077963,0.8739011931174752
|
| 170 |
+
9.174964438122332,12900,0.82805,0.902425,0.931075,0.82805,0.82805,0.3008083333333333,0.902425,0.18621500000000005,0.931075,0.82805,0.8677170833333292,0.8720985119047558,0.8941916115423987,0.8742162023078077
|
| 171 |
+
9.246088193456615,13000,0.82815,0.902575,0.93085,0.82815,0.82815,0.3008583333333333,0.902575,0.18617000000000003,0.93085,0.82815,0.8676999999999966,0.8721441765872964,0.894261787826358,0.8742444867490919
|
| 172 |
+
9.317211948790897,13100,0.828125,0.902475,0.930975,0.828125,0.828125,0.30082499999999995,0.902475,0.18619500000000003,0.930975,0.828125,0.8677062499999961,0.8721303075396769,0.8942478025806649,0.8742292661264696
|
| 173 |
+
9.388335704125177,13200,0.8281,0.90265,0.931025,0.8281,0.8281,0.3008833333333333,0.90265,0.18620500000000004,0.931025,0.8281,0.8678104166666628,0.8722282936507881,0.8943284212825287,0.8743275076393491
|
| 174 |
+
9.45945945945946,13300,0.828,0.902525,0.930925,0.828,0.828,0.3008416666666667,0.902525,0.18618500000000004,0.930925,0.828,0.8676795833333294,0.8720963095238036,0.8942075462606434,0.8742051203797395
|
| 175 |
+
9.530583214793742,13400,0.828025,0.90255,0.9309,0.828025,0.828025,0.3008499999999999,0.90255,0.18618000000000004,0.9309,0.828025,0.8676949999999959,0.8721207837301521,0.89423621668204,0.8742209824708375
|
| 176 |
+
9.601706970128022,13500,0.828075,0.902525,0.9309,0.828075,0.828075,0.3008416666666666,0.902525,0.18618,0.9309,0.828075,0.8677216666666631,0.8721371230158671,0.894225847427998,0.8742471205251703
|
| 177 |
+
9.672830725462305,13600,0.82815,0.902675,0.93085,0.82815,0.82815,0.30089166666666667,0.902675,0.18617000000000003,0.93085,0.82815,0.8677495833333296,0.8721985813492004,0.8943215425397492,0.8742927261961481
|
| 178 |
+
9.743954480796585,13700,0.828125,0.902625,0.93105,0.828125,0.828125,0.3008749999999999,0.902625,0.18621000000000001,0.93105,0.828125,0.8677633333333299,0.8721709722222164,0.8942849736621314,0.8742720567226947
|
| 179 |
+
9.815078236130867,13800,0.827975,0.90265,0.931,0.827975,0.827975,0.3008833333333333,0.90265,0.18620000000000003,0.931,0.827975,0.8676762499999963,0.8720721230158671,0.894178246968981,0.8741864677767469
|
| 180 |
+
9.88620199146515,13900,0.828025,0.9026,0.9311,0.828025,0.828025,0.3008666666666666,0.9026,0.18622000000000002,0.9311,0.828025,0.867727916666663,0.8721113690476133,0.8942189714318465,0.8742215047241502
|
| 181 |
+
9.95732574679943,14000,0.828025,0.9027,0.931025,0.828025,0.828025,0.3008999999999999,0.9027,0.186205,0.931025,0.828025,0.8677179166666629,0.8721162896825339,0.8942284691055087,0.8742240723304836
|
final_metrics.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
{
|
| 2 |
-
"val_cosine_accuracy@1": 0.
|
| 3 |
-
"val_cosine_accuracy@3": 0.
|
| 4 |
-
"val_cosine_accuracy@5": 0.
|
| 5 |
-
"val_cosine_precision@1": 0.
|
| 6 |
-
"val_cosine_precision@3": 0.
|
| 7 |
-
"val_cosine_precision@5": 0.
|
| 8 |
-
"val_cosine_recall@1": 0.
|
| 9 |
-
"val_cosine_recall@3": 0.
|
| 10 |
-
"val_cosine_recall@5": 0.
|
| 11 |
-
"val_cosine_ndcg@10": 0.
|
| 12 |
-
"val_cosine_mrr@1": 0.
|
| 13 |
-
"val_cosine_mrr@5": 0.
|
| 14 |
-
"val_cosine_mrr@10": 0.
|
| 15 |
-
"val_cosine_map@100": 0.
|
| 16 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"val_cosine_accuracy@1": 0.9104,
|
| 3 |
+
"val_cosine_accuracy@3": 0.9688,
|
| 4 |
+
"val_cosine_accuracy@5": 0.9842,
|
| 5 |
+
"val_cosine_precision@1": 0.9104,
|
| 6 |
+
"val_cosine_precision@3": 0.32293333333333335,
|
| 7 |
+
"val_cosine_precision@5": 0.19683999999999996,
|
| 8 |
+
"val_cosine_recall@1": 0.9104,
|
| 9 |
+
"val_cosine_recall@3": 0.9688,
|
| 10 |
+
"val_cosine_recall@5": 0.9842,
|
| 11 |
+
"val_cosine_ndcg@10": 0.9545809774353143,
|
| 12 |
+
"val_cosine_mrr@1": 0.9104,
|
| 13 |
+
"val_cosine_mrr@5": 0.9402433333333332,
|
| 14 |
+
"val_cosine_mrr@10": 0.9416250793650793,
|
| 15 |
+
"val_cosine_map@100": 0.9420576026548708
|
| 16 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 114011616
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1a4f8953c7f93243ebe09fe7c0daa9cffac072577f9e8000399b77733108e58b
|
| 3 |
size 114011616
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 6161
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:97dfabb4d1b0f4e7472f63fe7d96632afc7ca09080738d20201c498fcf74eb45
|
| 3 |
size 6161
|