Training in progress, step 14060
Browse files- Information-Retrieval_evaluation_val_results.csv +1 -0
- README.md +73 -236
- eval/Information-Retrieval_evaluation_val_results.csv +142 -0
- final_metrics.json +14 -14
- model.safetensors +1 -1
- training_args.bin +1 -1
Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -1,2 +1,3 @@
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.9208,0.9698,0.9842,0.9208,0.9208,0.3232666666666667,0.9698,0.19684,0.9842,0.9208,0.9460899999999998,0.9476021428571432,0.9593212690041523,0.9479260307963899
|
|
|
|
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.9208,0.9698,0.9842,0.9208,0.9208,0.3232666666666667,0.9698,0.19684,0.9842,0.9208,0.9460899999999998,0.9476021428571432,0.9593212690041523,0.9479260307963899
|
| 3 |
+
-1,-1,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451366666666663,0.9466038095238087,0.9586270476620361,0.946959374340519
|
README.md
CHANGED
|
@@ -5,109 +5,38 @@ tags:
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
-
- dataset_size:
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
-
- source_sentence: How do I
|
| 13 |
sentences:
|
| 14 |
-
- How
|
| 15 |
-
-
|
| 16 |
-
- What
|
| 17 |
-
- source_sentence:
|
| 18 |
sentences:
|
| 19 |
-
-
|
| 20 |
-
-
|
| 21 |
-
|
| 22 |
-
-
|
| 23 |
-
- source_sentence:
|
| 24 |
sentences:
|
| 25 |
-
- What
|
| 26 |
-
-
|
| 27 |
-
-
|
| 28 |
-
|
| 29 |
-
- source_sentence: What are the differences between eccentric and concentric contraction?
|
| 30 |
-
What are some examples?
|
| 31 |
sentences:
|
| 32 |
-
-
|
| 33 |
-
- How
|
| 34 |
-
-
|
| 35 |
-
- source_sentence:
|
| 36 |
-
app. Is it necessary to use that coupon only when I order online?
|
| 37 |
sentences:
|
| 38 |
-
-
|
| 39 |
-
-
|
| 40 |
-
-
|
| 41 |
-
necessary to use that coupon only when I order online?
|
| 42 |
pipeline_tag: sentence-similarity
|
| 43 |
library_name: sentence-transformers
|
| 44 |
-
metrics:
|
| 45 |
-
- cosine_accuracy@1
|
| 46 |
-
- cosine_accuracy@3
|
| 47 |
-
- cosine_accuracy@5
|
| 48 |
-
- cosine_precision@1
|
| 49 |
-
- cosine_precision@3
|
| 50 |
-
- cosine_precision@5
|
| 51 |
-
- cosine_recall@1
|
| 52 |
-
- cosine_recall@3
|
| 53 |
-
- cosine_recall@5
|
| 54 |
-
- cosine_ndcg@10
|
| 55 |
-
- cosine_mrr@1
|
| 56 |
-
- cosine_mrr@5
|
| 57 |
-
- cosine_mrr@10
|
| 58 |
-
- cosine_map@100
|
| 59 |
-
model-index:
|
| 60 |
-
- name: SentenceTransformer based on prajjwal1/bert-small
|
| 61 |
-
results:
|
| 62 |
-
- task:
|
| 63 |
-
type: information-retrieval
|
| 64 |
-
name: Information Retrieval
|
| 65 |
-
dataset:
|
| 66 |
-
name: val
|
| 67 |
-
type: val
|
| 68 |
-
metrics:
|
| 69 |
-
- type: cosine_accuracy@1
|
| 70 |
-
value: 0.9184
|
| 71 |
-
name: Cosine Accuracy@1
|
| 72 |
-
- type: cosine_accuracy@3
|
| 73 |
-
value: 0.97
|
| 74 |
-
name: Cosine Accuracy@3
|
| 75 |
-
- type: cosine_accuracy@5
|
| 76 |
-
value: 0.9852
|
| 77 |
-
name: Cosine Accuracy@5
|
| 78 |
-
- type: cosine_precision@1
|
| 79 |
-
value: 0.9184
|
| 80 |
-
name: Cosine Precision@1
|
| 81 |
-
- type: cosine_precision@3
|
| 82 |
-
value: 0.3233333333333333
|
| 83 |
-
name: Cosine Precision@3
|
| 84 |
-
- type: cosine_precision@5
|
| 85 |
-
value: 0.19703999999999997
|
| 86 |
-
name: Cosine Precision@5
|
| 87 |
-
- type: cosine_recall@1
|
| 88 |
-
value: 0.9184
|
| 89 |
-
name: Cosine Recall@1
|
| 90 |
-
- type: cosine_recall@3
|
| 91 |
-
value: 0.97
|
| 92 |
-
name: Cosine Recall@3
|
| 93 |
-
- type: cosine_recall@5
|
| 94 |
-
value: 0.9852
|
| 95 |
-
name: Cosine Recall@5
|
| 96 |
-
- type: cosine_ndcg@10
|
| 97 |
-
value: 0.9585962869405669
|
| 98 |
-
name: Cosine Ndcg@10
|
| 99 |
-
- type: cosine_mrr@1
|
| 100 |
-
value: 0.9184
|
| 101 |
-
name: Cosine Mrr@1
|
| 102 |
-
- type: cosine_mrr@5
|
| 103 |
-
value: 0.9451033333333331
|
| 104 |
-
name: Cosine Mrr@5
|
| 105 |
-
- type: cosine_mrr@10
|
| 106 |
-
value: 0.9465657142857136
|
| 107 |
-
name: Cosine Mrr@10
|
| 108 |
-
- type: cosine_map@100
|
| 109 |
-
value: 0.9469212791024237
|
| 110 |
-
name: Cosine Map@100
|
| 111 |
---
|
| 112 |
|
| 113 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
@@ -156,12 +85,12 @@ Then you can load this model and run inference.
|
|
| 156 |
from sentence_transformers import SentenceTransformer
|
| 157 |
|
| 158 |
# Download from the 🤗 Hub
|
| 159 |
-
model = SentenceTransformer("
|
| 160 |
# Run inference
|
| 161 |
sentences = [
|
| 162 |
-
|
| 163 |
-
|
| 164 |
-
'
|
| 165 |
]
|
| 166 |
embeddings = model.encode(sentences)
|
| 167 |
print(embeddings.shape)
|
|
@@ -170,9 +99,9 @@ print(embeddings.shape)
|
|
| 170 |
# Get the similarity scores for the embeddings
|
| 171 |
similarities = model.similarity(embeddings, embeddings)
|
| 172 |
print(similarities)
|
| 173 |
-
# tensor([[1.0000,
|
| 174 |
-
# [
|
| 175 |
-
# [0.
|
| 176 |
```
|
| 177 |
|
| 178 |
<!--
|
|
@@ -199,32 +128,6 @@ You can finetune this model on your own dataset.
|
|
| 199 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 200 |
-->
|
| 201 |
|
| 202 |
-
## Evaluation
|
| 203 |
-
|
| 204 |
-
### Metrics
|
| 205 |
-
|
| 206 |
-
#### Information Retrieval
|
| 207 |
-
|
| 208 |
-
* Dataset: `val`
|
| 209 |
-
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
| 210 |
-
|
| 211 |
-
| Metric | Value |
|
| 212 |
-
|:-------------------|:-----------|
|
| 213 |
-
| cosine_accuracy@1 | 0.9184 |
|
| 214 |
-
| cosine_accuracy@3 | 0.97 |
|
| 215 |
-
| cosine_accuracy@5 | 0.9852 |
|
| 216 |
-
| cosine_precision@1 | 0.9184 |
|
| 217 |
-
| cosine_precision@3 | 0.3233 |
|
| 218 |
-
| cosine_precision@5 | 0.197 |
|
| 219 |
-
| cosine_recall@1 | 0.9184 |
|
| 220 |
-
| cosine_recall@3 | 0.97 |
|
| 221 |
-
| cosine_recall@5 | 0.9852 |
|
| 222 |
-
| **cosine_ndcg@10** | **0.9586** |
|
| 223 |
-
| cosine_mrr@1 | 0.9184 |
|
| 224 |
-
| cosine_mrr@5 | 0.9451 |
|
| 225 |
-
| cosine_mrr@10 | 0.9466 |
|
| 226 |
-
| cosine_map@100 | 0.9469 |
|
| 227 |
-
|
| 228 |
<!--
|
| 229 |
## Bias, Risks and Limitations
|
| 230 |
|
|
@@ -243,45 +146,19 @@ You can finetune this model on your own dataset.
|
|
| 243 |
|
| 244 |
#### Unnamed Dataset
|
| 245 |
|
| 246 |
-
* Size:
|
| 247 |
-
* Columns: <code>
|
| 248 |
-
* Approximate statistics based on the first 1000 samples:
|
| 249 |
-
| | anchor | positive | negative |
|
| 250 |
-
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
|
| 251 |
-
| type | string | string | string |
|
| 252 |
-
| details | <ul><li>min: 6 tokens</li><li>mean: 15.63 tokens</li><li>max: 75 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.77 tokens</li><li>max: 75 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.5 tokens</li><li>max: 67 tokens</li></ul> |
|
| 253 |
-
* Samples:
|
| 254 |
-
| anchor | positive | negative |
|
| 255 |
-
|:---------------------------------------------------------|:---------------------------------------------------------|:----------------------------------------------------------------------------|
|
| 256 |
-
| <code>How long did it take to develop Pokémon GO?</code> | <code>How long did it take to develop Pokémon GO?</code> | <code>Can I take more than one gym in Pokémon GO?</code> |
|
| 257 |
-
| <code>How bad is 6/18 eyesight?</code> | <code>How bad is 6/18 eyesight?</code> | <code>How was bad eyesight dealt with in ancient and medieval times?</code> |
|
| 258 |
-
| <code>How can I do learn speaking English easily?</code> | <code>How can I learn speaking English easily?</code> | <code>How do you hack an Instagram account?</code> |
|
| 259 |
-
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 260 |
-
```json
|
| 261 |
-
{
|
| 262 |
-
"scale": 20.0,
|
| 263 |
-
"similarity_fct": "cos_sim",
|
| 264 |
-
"gather_across_devices": false
|
| 265 |
-
}
|
| 266 |
-
```
|
| 267 |
-
|
| 268 |
-
### Evaluation Dataset
|
| 269 |
-
|
| 270 |
-
#### Unnamed Dataset
|
| 271 |
-
|
| 272 |
-
* Size: 5,000 evaluation samples
|
| 273 |
-
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
| 274 |
* Approximate statistics based on the first 1000 samples:
|
| 275 |
-
| |
|
| 276 |
-
|
| 277 |
-
| type | string | string
|
| 278 |
-
| details | <ul><li>min:
|
| 279 |
* Samples:
|
| 280 |
-
|
|
| 281 |
-
|
| 282 |
-
| <code>
|
| 283 |
-
| <code>
|
| 284 |
-
| <code>
|
| 285 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 286 |
```json
|
| 287 |
{
|
|
@@ -294,49 +171,36 @@ You can finetune this model on your own dataset.
|
|
| 294 |
### Training Hyperparameters
|
| 295 |
#### Non-Default Hyperparameters
|
| 296 |
|
| 297 |
-
- `
|
| 298 |
-
- `
|
| 299 |
-
- `per_device_eval_batch_size`: 256
|
| 300 |
-
- `learning_rate`: 2e-05
|
| 301 |
-
- `weight_decay`: 0.001
|
| 302 |
-
- `max_steps`: 3510
|
| 303 |
-
- `warmup_ratio`: 0.1
|
| 304 |
- `fp16`: True
|
| 305 |
-
- `
|
| 306 |
-
- `dataloader_num_workers`: 1
|
| 307 |
-
- `dataloader_prefetch_factor`: 1
|
| 308 |
-
- `load_best_model_at_end`: True
|
| 309 |
-
- `optim`: adamw_torch
|
| 310 |
-
- `ddp_find_unused_parameters`: False
|
| 311 |
-
- `push_to_hub`: True
|
| 312 |
-
- `hub_model_id`: redis/model-a-baseline
|
| 313 |
-
- `eval_on_start`: True
|
| 314 |
|
| 315 |
#### All Hyperparameters
|
| 316 |
<details><summary>Click to expand</summary>
|
| 317 |
|
| 318 |
- `overwrite_output_dir`: False
|
| 319 |
- `do_predict`: False
|
| 320 |
-
- `eval_strategy`:
|
| 321 |
- `prediction_loss_only`: True
|
| 322 |
-
- `per_device_train_batch_size`:
|
| 323 |
-
- `per_device_eval_batch_size`:
|
| 324 |
- `per_gpu_train_batch_size`: None
|
| 325 |
- `per_gpu_eval_batch_size`: None
|
| 326 |
- `gradient_accumulation_steps`: 1
|
| 327 |
- `eval_accumulation_steps`: None
|
| 328 |
- `torch_empty_cache_steps`: None
|
| 329 |
-
- `learning_rate`:
|
| 330 |
-
- `weight_decay`: 0.
|
| 331 |
- `adam_beta1`: 0.9
|
| 332 |
- `adam_beta2`: 0.999
|
| 333 |
- `adam_epsilon`: 1e-08
|
| 334 |
-
- `max_grad_norm`: 1
|
| 335 |
-
- `num_train_epochs`: 3
|
| 336 |
-
- `max_steps`:
|
| 337 |
- `lr_scheduler_type`: linear
|
| 338 |
- `lr_scheduler_kwargs`: {}
|
| 339 |
-
- `warmup_ratio`: 0.
|
| 340 |
- `warmup_steps`: 0
|
| 341 |
- `log_level`: passive
|
| 342 |
- `log_level_replica`: warning
|
|
@@ -364,14 +228,14 @@ You can finetune this model on your own dataset.
|
|
| 364 |
- `tpu_num_cores`: None
|
| 365 |
- `tpu_metrics_debug`: False
|
| 366 |
- `debug`: []
|
| 367 |
-
- `dataloader_drop_last`:
|
| 368 |
-
- `dataloader_num_workers`:
|
| 369 |
-
- `dataloader_prefetch_factor`:
|
| 370 |
- `past_index`: -1
|
| 371 |
- `disable_tqdm`: False
|
| 372 |
- `remove_unused_columns`: True
|
| 373 |
- `label_names`: None
|
| 374 |
-
- `load_best_model_at_end`:
|
| 375 |
- `ignore_data_skip`: False
|
| 376 |
- `fsdp`: []
|
| 377 |
- `fsdp_min_num_params`: 0
|
|
@@ -381,23 +245,23 @@ You can finetune this model on your own dataset.
|
|
| 381 |
- `parallelism_config`: None
|
| 382 |
- `deepspeed`: None
|
| 383 |
- `label_smoothing_factor`: 0.0
|
| 384 |
-
- `optim`:
|
| 385 |
- `optim_args`: None
|
| 386 |
- `adafactor`: False
|
| 387 |
- `group_by_length`: False
|
| 388 |
- `length_column_name`: length
|
| 389 |
- `project`: huggingface
|
| 390 |
- `trackio_space_id`: trackio
|
| 391 |
-
- `ddp_find_unused_parameters`:
|
| 392 |
- `ddp_bucket_cap_mb`: None
|
| 393 |
- `ddp_broadcast_buffers`: False
|
| 394 |
- `dataloader_pin_memory`: True
|
| 395 |
- `dataloader_persistent_workers`: False
|
| 396 |
- `skip_memory_metrics`: True
|
| 397 |
- `use_legacy_prediction_loop`: False
|
| 398 |
-
- `push_to_hub`:
|
| 399 |
- `resume_from_checkpoint`: None
|
| 400 |
-
- `hub_model_id`:
|
| 401 |
- `hub_strategy`: every_save
|
| 402 |
- `hub_private_repo`: None
|
| 403 |
- `hub_always_push`: False
|
|
@@ -424,58 +288,31 @@ You can finetune this model on your own dataset.
|
|
| 424 |
- `neftune_noise_alpha`: None
|
| 425 |
- `optim_target_modules`: None
|
| 426 |
- `batch_eval_metrics`: False
|
| 427 |
-
- `eval_on_start`:
|
| 428 |
- `use_liger_kernel`: False
|
| 429 |
- `liger_kernel_config`: None
|
| 430 |
- `eval_use_gather_object`: False
|
| 431 |
- `average_tokens_across_devices`: True
|
| 432 |
- `prompts`: None
|
| 433 |
- `batch_sampler`: batch_sampler
|
| 434 |
-
- `multi_dataset_batch_sampler`:
|
| 435 |
- `router_mapping`: {}
|
| 436 |
- `learning_rate_mapping`: {}
|
| 437 |
|
| 438 |
</details>
|
| 439 |
|
| 440 |
### Training Logs
|
| 441 |
-
| Epoch | Step | Training Loss |
|
| 442 |
-
|
| 443 |
-
| 0
|
| 444 |
-
| 0.
|
| 445 |
-
| 0.
|
| 446 |
-
|
|
| 447 |
-
| 1.
|
| 448 |
-
| 1.
|
| 449 |
-
|
|
| 450 |
-
|
|
| 451 |
-
| 2.
|
| 452 |
-
| 2.5641 | 900 | 0.0782 | 0.0439 | 0.9562 |
|
| 453 |
-
| 2.8490 | 1000 | 0.0745 | 0.0436 | 0.9572 |
|
| 454 |
-
| 3.1339 | 1100 | 0.0732 | 0.0421 | 0.9570 |
|
| 455 |
-
| 3.4188 | 1200 | 0.0688 | 0.0417 | 0.9577 |
|
| 456 |
-
| 3.7037 | 1300 | 0.0687 | 0.0411 | 0.9576 |
|
| 457 |
-
| 3.9886 | 1400 | 0.07 | 0.0412 | 0.9573 |
|
| 458 |
-
| 4.2735 | 1500 | 0.0635 | 0.0402 | 0.9578 |
|
| 459 |
-
| 4.5584 | 1600 | 0.0638 | 0.0397 | 0.9575 |
|
| 460 |
-
| 4.8433 | 1700 | 0.0613 | 0.0394 | 0.9579 |
|
| 461 |
-
| 5.1282 | 1800 | 0.0625 | 0.0388 | 0.9584 |
|
| 462 |
-
| 5.4131 | 1900 | 0.0585 | 0.0382 | 0.9586 |
|
| 463 |
-
| 5.6980 | 2000 | 0.0594 | 0.0379 | 0.9585 |
|
| 464 |
-
| 5.9829 | 2100 | 0.0566 | 0.0377 | 0.9584 |
|
| 465 |
-
| 6.2678 | 2200 | 0.0545 | 0.0376 | 0.9583 |
|
| 466 |
-
| 6.5527 | 2300 | 0.0535 | 0.0376 | 0.9580 |
|
| 467 |
-
| 6.8376 | 2400 | 0.0573 | 0.0373 | 0.9584 |
|
| 468 |
-
| 7.1225 | 2500 | 0.0528 | 0.0373 | 0.9583 |
|
| 469 |
-
| 7.4074 | 2600 | 0.053 | 0.0371 | 0.9587 |
|
| 470 |
-
| 7.6923 | 2700 | 0.0528 | 0.0368 | 0.9587 |
|
| 471 |
-
| 7.9772 | 2800 | 0.0531 | 0.0366 | 0.9585 |
|
| 472 |
-
| 8.2621 | 2900 | 0.0532 | 0.0365 | 0.9586 |
|
| 473 |
-
| 8.5470 | 3000 | 0.0516 | 0.0365 | 0.9584 |
|
| 474 |
-
| 8.8319 | 3100 | 0.0509 | 0.0364 | 0.9585 |
|
| 475 |
-
| 9.1168 | 3200 | 0.0544 | 0.0363 | 0.9587 |
|
| 476 |
-
| 9.4017 | 3300 | 0.0505 | 0.0364 | 0.9585 |
|
| 477 |
-
| 9.6866 | 3400 | 0.052 | 0.0363 | 0.9587 |
|
| 478 |
-
| 9.9715 | 3500 | 0.0536 | 0.0362 | 0.9586 |
|
| 479 |
|
| 480 |
|
| 481 |
### Framework Versions
|
|
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
+
- dataset_size:100000
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
+
- source_sentence: How do I polish my English skills?
|
| 13 |
sentences:
|
| 14 |
+
- How can we polish English skills?
|
| 15 |
+
- Why should I move to Israel as a Jew?
|
| 16 |
+
- What are vitamins responsible for?
|
| 17 |
+
- source_sentence: Can I use the Kozuka Gothic Pro font as a font-face on my web site?
|
| 18 |
sentences:
|
| 19 |
+
- Can I use the Kozuka Gothic Pro font as a font-face on my web site?
|
| 20 |
+
- Why are Google, Facebook, YouTube and other social networking sites banned in
|
| 21 |
+
China?
|
| 22 |
+
- What font is used in Bloomberg Terminal?
|
| 23 |
+
- source_sentence: Is Quora the best Q&A site?
|
| 24 |
sentences:
|
| 25 |
+
- What was the best Quora question ever?
|
| 26 |
+
- Is Quora the best inquiry site?
|
| 27 |
+
- Where do I buy Oway hair products online?
|
| 28 |
+
- source_sentence: How can I customize my walking speed on Google Maps?
|
|
|
|
|
|
|
| 29 |
sentences:
|
| 30 |
+
- How do I bring back Google maps icon in my home screen?
|
| 31 |
+
- How many pages are there in all the Harry Potter books combined?
|
| 32 |
+
- How can I customize my walking speed on Google Maps?
|
| 33 |
+
- source_sentence: DId something exist before the Big Bang?
|
|
|
|
| 34 |
sentences:
|
| 35 |
+
- How can I improve my memory problem?
|
| 36 |
+
- Where can I buy Fairy Tail Manga?
|
| 37 |
+
- Is there a scientific name for what existed before the Big Bang?
|
|
|
|
| 38 |
pipeline_tag: sentence-similarity
|
| 39 |
library_name: sentence-transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
---
|
| 41 |
|
| 42 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
|
|
| 85 |
from sentence_transformers import SentenceTransformer
|
| 86 |
|
| 87 |
# Download from the 🤗 Hub
|
| 88 |
+
model = SentenceTransformer("sentence_transformers_model_id")
|
| 89 |
# Run inference
|
| 90 |
sentences = [
|
| 91 |
+
'DId something exist before the Big Bang?',
|
| 92 |
+
'Is there a scientific name for what existed before the Big Bang?',
|
| 93 |
+
'Where can I buy Fairy Tail Manga?',
|
| 94 |
]
|
| 95 |
embeddings = model.encode(sentences)
|
| 96 |
print(embeddings.shape)
|
|
|
|
| 99 |
# Get the similarity scores for the embeddings
|
| 100 |
similarities = model.similarity(embeddings, embeddings)
|
| 101 |
print(similarities)
|
| 102 |
+
# tensor([[ 1.0000, 0.7596, -0.0398],
|
| 103 |
+
# [ 0.7596, 1.0000, -0.0308],
|
| 104 |
+
# [-0.0398, -0.0308, 1.0000]])
|
| 105 |
```
|
| 106 |
|
| 107 |
<!--
|
|
|
|
| 128 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 129 |
-->
|
| 130 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
<!--
|
| 132 |
## Bias, Risks and Limitations
|
| 133 |
|
|
|
|
| 146 |
|
| 147 |
#### Unnamed Dataset
|
| 148 |
|
| 149 |
+
* Size: 100,000 training samples
|
| 150 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 151 |
* Approximate statistics based on the first 1000 samples:
|
| 152 |
+
| | sentence_0 | sentence_1 | sentence_2 |
|
| 153 |
+
|:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
|
| 154 |
+
| type | string | string | string |
|
| 155 |
+
| details | <ul><li>min: 3 tokens</li><li>mean: 15.53 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 15.5 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.87 tokens</li><li>max: 128 tokens</li></ul> |
|
| 156 |
* Samples:
|
| 157 |
+
| sentence_0 | sentence_1 | sentence_2 |
|
| 158 |
+
|:----------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------|:-----------------------------------------------------------------------|
|
| 159 |
+
| <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>How much is the airport tax in bogota?</code> |
|
| 160 |
+
| <code>Which concept is more important: good planning or hard work?</code> | <code>Which concept is more important: good planning or hard work?</code> | <code>What is important in life: luck or hard work?</code> |
|
| 161 |
+
| <code>What is the most efficient way to make money?</code> | <code>How can I make my money make money?</code> | <code>What can one learn about Quantum Mechanics in 10 minutes?</code> |
|
| 162 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 163 |
```json
|
| 164 |
{
|
|
|
|
| 171 |
### Training Hyperparameters
|
| 172 |
#### Non-Default Hyperparameters
|
| 173 |
|
| 174 |
+
- `per_device_train_batch_size`: 64
|
| 175 |
+
- `per_device_eval_batch_size`: 64
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 176 |
- `fp16`: True
|
| 177 |
+
- `multi_dataset_batch_sampler`: round_robin
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 178 |
|
| 179 |
#### All Hyperparameters
|
| 180 |
<details><summary>Click to expand</summary>
|
| 181 |
|
| 182 |
- `overwrite_output_dir`: False
|
| 183 |
- `do_predict`: False
|
| 184 |
+
- `eval_strategy`: no
|
| 185 |
- `prediction_loss_only`: True
|
| 186 |
+
- `per_device_train_batch_size`: 64
|
| 187 |
+
- `per_device_eval_batch_size`: 64
|
| 188 |
- `per_gpu_train_batch_size`: None
|
| 189 |
- `per_gpu_eval_batch_size`: None
|
| 190 |
- `gradient_accumulation_steps`: 1
|
| 191 |
- `eval_accumulation_steps`: None
|
| 192 |
- `torch_empty_cache_steps`: None
|
| 193 |
+
- `learning_rate`: 5e-05
|
| 194 |
+
- `weight_decay`: 0.0
|
| 195 |
- `adam_beta1`: 0.9
|
| 196 |
- `adam_beta2`: 0.999
|
| 197 |
- `adam_epsilon`: 1e-08
|
| 198 |
+
- `max_grad_norm`: 1
|
| 199 |
+
- `num_train_epochs`: 3
|
| 200 |
+
- `max_steps`: -1
|
| 201 |
- `lr_scheduler_type`: linear
|
| 202 |
- `lr_scheduler_kwargs`: {}
|
| 203 |
+
- `warmup_ratio`: 0.0
|
| 204 |
- `warmup_steps`: 0
|
| 205 |
- `log_level`: passive
|
| 206 |
- `log_level_replica`: warning
|
|
|
|
| 228 |
- `tpu_num_cores`: None
|
| 229 |
- `tpu_metrics_debug`: False
|
| 230 |
- `debug`: []
|
| 231 |
+
- `dataloader_drop_last`: False
|
| 232 |
+
- `dataloader_num_workers`: 0
|
| 233 |
+
- `dataloader_prefetch_factor`: None
|
| 234 |
- `past_index`: -1
|
| 235 |
- `disable_tqdm`: False
|
| 236 |
- `remove_unused_columns`: True
|
| 237 |
- `label_names`: None
|
| 238 |
+
- `load_best_model_at_end`: False
|
| 239 |
- `ignore_data_skip`: False
|
| 240 |
- `fsdp`: []
|
| 241 |
- `fsdp_min_num_params`: 0
|
|
|
|
| 245 |
- `parallelism_config`: None
|
| 246 |
- `deepspeed`: None
|
| 247 |
- `label_smoothing_factor`: 0.0
|
| 248 |
+
- `optim`: adamw_torch_fused
|
| 249 |
- `optim_args`: None
|
| 250 |
- `adafactor`: False
|
| 251 |
- `group_by_length`: False
|
| 252 |
- `length_column_name`: length
|
| 253 |
- `project`: huggingface
|
| 254 |
- `trackio_space_id`: trackio
|
| 255 |
+
- `ddp_find_unused_parameters`: None
|
| 256 |
- `ddp_bucket_cap_mb`: None
|
| 257 |
- `ddp_broadcast_buffers`: False
|
| 258 |
- `dataloader_pin_memory`: True
|
| 259 |
- `dataloader_persistent_workers`: False
|
| 260 |
- `skip_memory_metrics`: True
|
| 261 |
- `use_legacy_prediction_loop`: False
|
| 262 |
+
- `push_to_hub`: False
|
| 263 |
- `resume_from_checkpoint`: None
|
| 264 |
+
- `hub_model_id`: None
|
| 265 |
- `hub_strategy`: every_save
|
| 266 |
- `hub_private_repo`: None
|
| 267 |
- `hub_always_push`: False
|
|
|
|
| 288 |
- `neftune_noise_alpha`: None
|
| 289 |
- `optim_target_modules`: None
|
| 290 |
- `batch_eval_metrics`: False
|
| 291 |
+
- `eval_on_start`: False
|
| 292 |
- `use_liger_kernel`: False
|
| 293 |
- `liger_kernel_config`: None
|
| 294 |
- `eval_use_gather_object`: False
|
| 295 |
- `average_tokens_across_devices`: True
|
| 296 |
- `prompts`: None
|
| 297 |
- `batch_sampler`: batch_sampler
|
| 298 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 299 |
- `router_mapping`: {}
|
| 300 |
- `learning_rate_mapping`: {}
|
| 301 |
|
| 302 |
</details>
|
| 303 |
|
| 304 |
### Training Logs
|
| 305 |
+
| Epoch | Step | Training Loss |
|
| 306 |
+
|:------:|:----:|:-------------:|
|
| 307 |
+
| 0.3199 | 500 | 0.2284 |
|
| 308 |
+
| 0.6398 | 1000 | 0.0571 |
|
| 309 |
+
| 0.9597 | 1500 | 0.0486 |
|
| 310 |
+
| 1.2796 | 2000 | 0.0378 |
|
| 311 |
+
| 1.5995 | 2500 | 0.0367 |
|
| 312 |
+
| 1.9194 | 3000 | 0.0338 |
|
| 313 |
+
| 2.2393 | 3500 | 0.0327 |
|
| 314 |
+
| 2.5592 | 4000 | 0.0285 |
|
| 315 |
+
| 2.8791 | 4500 | 0.0285 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 316 |
|
| 317 |
|
| 318 |
### Framework Versions
|
eval/Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -38,3 +38,145 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
|
|
| 38 |
9.401709401709402,3300,0.918,0.9702,0.9854,0.918,0.918,0.3234,0.9702,0.19707999999999998,0.9854,0.918,0.9449833333333331,0.9464248412698404,0.9585364250732368,0.9467604500159339
|
| 39 |
9.686609686609687,3400,0.9184,0.9702,0.9852,0.9184,0.9184,0.3234,0.9702,0.19703999999999997,0.9852,0.9184,0.9451433333333329,0.946628888888888,0.9586880746687382,0.9469662432012432
|
| 40 |
9.971509971509972,3500,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451033333333331,0.9465657142857136,0.9585962869405669,0.9469212791024237
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
9.401709401709402,3300,0.918,0.9702,0.9854,0.918,0.918,0.3234,0.9702,0.19707999999999998,0.9854,0.918,0.9449833333333331,0.9464248412698404,0.9585364250732368,0.9467604500159339
|
| 39 |
9.686609686609687,3400,0.9184,0.9702,0.9852,0.9184,0.9184,0.3234,0.9702,0.19703999999999997,0.9852,0.9184,0.9451433333333329,0.946628888888888,0.9586880746687382,0.9469662432012432
|
| 40 |
9.971509971509972,3500,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451033333333331,0.9465657142857136,0.9585962869405669,0.9469212791024237
|
| 41 |
+
0,0,0.755775,0.808875,0.831025,0.755775,0.755775,0.269625,0.808875,0.16620500000000002,0.831025,0.755775,0.7844620833333302,0.7884356150793639,0.8057171258182194,0.7915587339356275
|
| 42 |
+
0,0,0.7547,0.807275,0.83055,0.7547,0.7547,0.2690916666666667,0.807275,0.16611,0.83055,0.7547,0.7832520833333291,0.7872015575396802,0.8045860339061293,0.7903241559279329
|
| 43 |
+
0.07112375533428165,100,0.78785,0.851075,0.875425,0.78785,0.78785,0.2836916666666666,0.851075,0.17508500000000002,0.875425,0.78785,0.8214812499999945,0.8259418948412669,0.845761155370626,0.8289483897643156
|
| 44 |
+
0.1422475106685633,200,0.807425,0.8746,0.8983,0.807425,0.807425,0.2915333333333333,0.8746,0.17966000000000001,0.8983,0.807425,0.8427741666666617,0.8470684722222184,0.8670382366177412,0.8497105310760685
|
| 45 |
+
0.21337126600284495,300,0.81345,0.881175,0.906,0.81345,0.81345,0.293725,0.881175,0.1812,0.906,0.81345,0.8493633333333297,0.853414712301584,0.8733459236763736,0.856061621034904
|
| 46 |
+
0.2844950213371266,400,0.81635,0.88395,0.9094,0.81635,0.81635,0.2946499999999999,0.88395,0.18188000000000001,0.9094,0.81635,0.8522504166666626,0.8563771230158682,0.8765756531000587,0.8589065085962513
|
| 47 |
+
0.35561877667140823,500,0.818,0.886175,0.911175,0.818,0.818,0.2953916666666666,0.886175,0.18223500000000004,0.911175,0.818,0.8539554166666614,0.8580862499999943,0.8782710147523383,0.860644702658202
|
| 48 |
+
0.4267425320056899,600,0.819475,0.887525,0.91275,0.819475,0.819475,0.2958416666666666,0.887525,0.18255000000000002,0.91275,0.819475,0.8554229166666614,0.8595847023809465,0.8798219435803046,0.8621275028888417
|
| 49 |
+
0.49786628733997157,700,0.820625,0.888875,0.9142,0.820625,0.820625,0.2962916666666667,0.888875,0.18284,0.9142,0.820625,0.856661666666661,0.860812668650788,0.8810357441052341,0.8633834646157147
|
| 50 |
+
0.5689900426742532,800,0.820925,0.8898,0.916,0.820925,0.820925,0.2966,0.8898,0.1832,0.916,0.820925,0.8574945833333277,0.8615348710317401,0.881842076773151,0.8640919041835208
|
| 51 |
+
0.6401137980085349,900,0.82165,0.890475,0.91665,0.82165,0.82165,0.29682499999999995,0.890475,0.18333,0.91665,0.82165,0.8582687499999951,0.8624309920634873,0.8828656071112492,0.8649659283844867
|
| 52 |
+
0.7112375533428165,1000,0.822125,0.890975,0.9178,0.822125,0.822125,0.2969916666666666,0.890975,0.18356000000000003,0.9178,0.822125,0.8588987499999945,0.8629843452380895,0.8834907335792129,0.865520857294089
|
| 53 |
+
0.7823613086770982,1100,0.8221,0.892,0.918775,0.8221,0.8221,0.2973333333333333,0.892,0.18375500000000003,0.918775,0.8221,0.8593183333333276,0.8634522321428502,0.8841758741095582,0.8659255405241576
|
| 54 |
+
0.8534850640113798,1200,0.822275,0.8923,0.919725,0.822275,0.822275,0.2974333333333333,0.8923,0.18394500000000003,0.919725,0.822275,0.8597520833333285,0.8637605456349143,0.8844333496944642,0.8663024863202818
|
| 55 |
+
0.9246088193456614,1300,0.823225,0.893525,0.920575,0.823225,0.823225,0.29784166666666667,0.893525,0.18411500000000003,0.920575,0.823225,0.8606774999999949,0.8647051984126934,0.8854128554093778,0.867209868995025
|
| 56 |
+
0.9957325746799431,1400,0.823175,0.894025,0.921075,0.823175,0.823175,0.2980083333333333,0.894025,0.18421500000000005,0.921075,0.823175,0.860848333333329,0.8649403769841214,0.8858600106323372,0.867393318084569
|
| 57 |
+
1.0668563300142249,1500,0.82355,0.894475,0.921425,0.82355,0.82355,0.2981583333333333,0.894475,0.18428500000000003,0.921425,0.82355,0.8612766666666617,0.8654512003968201,0.8864450013473706,0.8678860308032963
|
| 58 |
+
1.1379800853485065,1600,0.824075,0.893975,0.92145,0.824075,0.824075,0.2979916666666666,0.893975,0.18429000000000006,0.92145,0.824075,0.8616245833333293,0.8658359027777728,0.8867900484928938,0.8682721259875404
|
| 59 |
+
1.209103840682788,1700,0.8243,0.89475,0.922,0.8243,0.8243,0.29824999999999996,0.89475,0.18440000000000004,0.922,0.8243,0.8618979166666617,0.8660758134920578,0.8870159340452088,0.8685273486995436
|
| 60 |
+
1.2802275960170697,1800,0.8239,0.894975,0.922625,0.8239,0.8239,0.2983249999999999,0.894975,0.184525,0.922625,0.8239,0.8619162499999955,0.8660865972222171,0.8871502465991052,0.868526031193211
|
| 61 |
+
1.3513513513513513,1900,0.8248,0.8953,0.92285,0.8248,0.8248,0.29843333333333333,0.8953,0.18457000000000004,0.92285,0.8248,0.8624554166666616,0.8667771626984062,0.887979642622687,0.8691551061429417
|
| 62 |
+
1.422475106685633,2000,0.825125,0.895675,0.92305,0.825125,0.825125,0.2985583333333333,0.895675,0.18461000000000002,0.92305,0.825125,0.8627966666666622,0.8671086408730106,0.8882894213555926,0.8694945359957533
|
| 63 |
+
1.4935988620199145,2100,0.82465,0.895475,0.9231,0.82465,0.82465,0.29849166666666666,0.895475,0.18462000000000003,0.9231,0.82465,0.8625087499999958,0.8667821329365039,0.8879802791439102,0.8691910176083405
|
| 64 |
+
1.5647226173541964,2200,0.824775,0.8964,0.92295,0.824775,0.824775,0.29879999999999995,0.8964,0.18459000000000006,0.92295,0.824775,0.8627320833333293,0.8670929067460275,0.8883622448503676,0.8694565192049544
|
| 65 |
+
1.635846372688478,2300,0.8248,0.896275,0.923925,0.8248,0.8248,0.2987583333333333,0.896275,0.18478500000000003,0.923925,0.8248,0.8630224999999968,0.8673088095238061,0.8886264876214104,0.8696932008768681
|
| 66 |
+
1.7069701280227596,2400,0.826025,0.89705,0.92455,0.826025,0.826025,0.29901666666666665,0.89705,0.18491,0.92455,0.826025,0.8638491666666629,0.8681291865079321,0.8894101179460802,0.87048095001459
|
| 67 |
+
1.7780938833570412,2500,0.82575,0.89745,0.924525,0.82575,0.82575,0.29914999999999997,0.89745,0.184905,0.924525,0.82575,0.863837499999996,0.8681003472222181,0.8893350803981342,0.8704980090657853
|
| 68 |
+
1.8492176386913228,2600,0.8262,0.89755,0.9245,0.8262,0.8262,0.29918333333333325,0.89755,0.18490000000000004,0.9245,0.8262,0.8639870833333294,0.8682846924603129,0.8895707677394977,0.8706525163734248
|
| 69 |
+
1.9203413940256047,2700,0.825625,0.897375,0.924725,0.825625,0.825625,0.299125,0.897375,0.18494500000000003,0.924725,0.825625,0.8637699999999955,0.8680662599206297,0.8894356770626558,0.8704346187020383
|
| 70 |
+
1.991465149359886,2800,0.8265,0.898175,0.9257,0.8265,0.8265,0.2993916666666666,0.898175,0.18514,0.9257,0.8265,0.8645654166666631,0.8688316964285668,0.8902285640417104,0.871149650917906
|
| 71 |
+
2.062588904694168,2900,0.82645,0.898625,0.92595,0.82645,0.82645,0.2995416666666666,0.898625,0.18519000000000002,0.92595,0.82645,0.8647274999999957,0.8689719345238045,0.890342026566598,0.871310932644214
|
| 72 |
+
2.1337126600284497,3000,0.826425,0.899225,0.9261,0.826425,0.826425,0.29974166666666663,0.899225,0.18522000000000002,0.9261,0.826425,0.8647395833333298,0.8689735416666621,0.890347823617768,0.8713261595062335
|
| 73 |
+
2.204836415362731,3100,0.826725,0.89905,0.92645,0.826725,0.826725,0.29968333333333325,0.89905,0.18529,0.92645,0.826725,0.8649899999999954,0.8692784920634858,0.8908060838763998,0.871566509775546
|
| 74 |
+
2.275960170697013,3200,0.8269,0.899,0.92655,0.8269,0.8269,0.2996666666666666,0.899,0.18531000000000003,0.92655,0.8269,0.8651883333333286,0.8694821428571373,0.8909713514566135,0.871765747281156
|
| 75 |
+
2.3470839260312943,3300,0.8271,0.8991,0.92655,0.8271,0.8271,0.2997,0.8991,0.18531,0.92655,0.8271,0.8653516666666632,0.8695972619047567,0.8909526338857612,0.8719357528477157
|
| 76 |
+
2.418207681365576,3400,0.826625,0.8992,0.926825,0.826625,0.826625,0.2997333333333333,0.8992,0.185365,0.926825,0.826625,0.8650733333333291,0.8692719940476129,0.8907103492984229,0.8716175978469811
|
| 77 |
+
2.4893314366998576,3500,0.826775,0.899325,0.926875,0.826775,0.826775,0.2997749999999999,0.899325,0.185375,0.926875,0.826775,0.8652879166666612,0.8695422321428495,0.8909951161321312,0.8718809967982727
|
| 78 |
+
2.5604551920341394,3600,0.827025,0.89965,0.926925,0.827025,0.827025,0.2998833333333333,0.89965,0.18538500000000005,0.926925,0.827025,0.8655404166666618,0.8698209027777717,0.8912443530415479,0.8721588161360633
|
| 79 |
+
2.6315789473684212,3700,0.8272,0.899875,0.92655,0.8272,0.8272,0.29995833333333327,0.899875,0.18531000000000003,0.92655,0.8272,0.8656041666666623,0.8700131448412636,0.8914837712801771,0.8723314630431057
|
| 80 |
+
2.7027027027027026,3800,0.8266,0.899825,0.927175,0.8266,0.8266,0.29994166666666666,0.899825,0.18543500000000004,0.927175,0.8266,0.8653658333333281,0.8696869841269774,0.8913161563114452,0.8719612942855474
|
| 81 |
+
2.7738264580369845,3900,0.8272,0.899825,0.92715,0.8272,0.8272,0.2999416666666666,0.899825,0.18543,0.92715,0.8272,0.8657754166666611,0.8701063789682477,0.8916378283367434,0.8723927691710214
|
| 82 |
+
2.844950213371266,4000,0.827075,0.89955,0.92675,0.827075,0.827075,0.29984999999999995,0.89955,0.18535000000000001,0.92675,0.827075,0.8655716666666616,0.8699127083333278,0.8913728457986007,0.8722460602722183
|
| 83 |
+
2.9160739687055477,4100,0.8273,0.900125,0.9276,0.8273,0.8273,0.3000416666666666,0.900125,0.18552000000000002,0.9276,0.8273,0.8659283333333285,0.870234494047613,0.8917878356256379,0.8725341820908176
|
| 84 |
+
2.987197724039829,4200,0.82765,0.899675,0.928275,0.82765,0.82765,0.29989166666666667,0.899675,0.18565500000000001,0.928275,0.82765,0.8661887499999953,0.8704055357142805,0.8919296060213343,0.8727003846925537
|
| 85 |
+
3.058321479374111,4300,0.827625,0.900125,0.92825,0.827625,0.827625,0.3000416666666666,0.900125,0.18565000000000004,0.92825,0.827625,0.8662208333333287,0.8705009821428515,0.8921252154319398,0.8727522181409114
|
| 86 |
+
3.1294452347083928,4400,0.828,0.900825,0.928325,0.828,0.828,0.300275,0.900825,0.18566500000000002,0.928325,0.828,0.8665966666666624,0.8708945734126934,0.8924466951830272,0.8731530635038529
|
| 87 |
+
3.200568990042674,4500,0.8277,0.9,0.92835,0.8277,0.8277,0.29999999999999993,0.9,0.18567000000000003,0.92835,0.8277,0.866319999999995,0.8706282341269782,0.8922483527994851,0.8728924711403877
|
| 88 |
+
3.271692745376956,4600,0.82805,0.90105,0.92865,0.82805,0.82805,0.30035,0.90105,0.18573,0.92865,0.82805,0.8666904166666612,0.8709508432539609,0.8925163619705987,0.8732096836080842
|
| 89 |
+
3.3428165007112374,4700,0.8278,0.900575,0.92825,0.8278,0.8278,0.3001916666666666,0.900575,0.18565,0.92825,0.8278,0.8665091666666603,0.8708543650793578,0.8924746282919447,0.8731299308768117
|
| 90 |
+
3.413940256045519,4800,0.827825,0.900525,0.92835,0.827825,0.827825,0.30017499999999997,0.900525,0.18567000000000003,0.92835,0.827825,0.8665587499999943,0.8708749603174522,0.8925184426900059,0.8731293966791309
|
| 91 |
+
3.485064011379801,4900,0.8282,0.901025,0.9294,0.8282,0.8282,0.3003416666666666,0.901025,0.18588000000000002,0.9294,0.8282,0.8669949999999941,0.8712464583333255,0.8929341649361835,0.873471354856535
|
| 92 |
+
3.5561877667140824,5000,0.82795,0.90165,0.928775,0.82795,0.82795,0.30055,0.90165,0.18575500000000003,0.928775,0.82795,0.8668574999999944,0.8712117658730092,0.8929485306096925,0.8734092206180317
|
| 93 |
+
3.6273115220483643,5100,0.82795,0.90155,0.9291,0.82795,0.82795,0.30051666666666665,0.90155,0.18582000000000004,0.9291,0.82795,0.8669424999999945,0.8712447321428517,0.8929399616791724,0.8734732719376523
|
| 94 |
+
3.6984352773826457,5200,0.82825,0.901425,0.929075,0.82825,0.82825,0.30047499999999994,0.901425,0.185815,0.929075,0.82825,0.8670362499999954,0.8712648313492005,0.8928164376378238,0.8735312781146788
|
| 95 |
+
3.7695590327169275,5300,0.827725,0.901425,0.92925,0.827725,0.827725,0.30047499999999994,0.901425,0.18585000000000002,0.92925,0.827725,0.8668066666666613,0.8710934523809452,0.8928289077266707,0.873324647003722
|
| 96 |
+
3.8406827880512093,5400,0.827825,0.90145,0.929,0.827825,0.827825,0.3004833333333333,0.90145,0.18580000000000002,0.929,0.827825,0.8668691666666618,0.871203353174598,0.8928865801326494,0.8734427586278187
|
| 97 |
+
3.9118065433854907,5500,0.828325,0.90155,0.929,0.828325,0.828325,0.30051666666666665,0.90155,0.18580000000000002,0.929,0.828325,0.8671308333333289,0.8714839384920581,0.8931380829352583,0.87370417401581
|
| 98 |
+
3.9829302987197726,5600,0.827875,0.901325,0.929625,0.827875,0.827875,0.3004416666666666,0.901325,0.18592500000000003,0.929625,0.827875,0.8670637499999959,0.8713995535714232,0.8932282155573206,0.8735739145143638
|
| 99 |
+
4.054054054054054,5700,0.828225,0.901825,0.92965,0.828225,0.828225,0.3006083333333333,0.901825,0.18592999999999998,0.92965,0.828225,0.8672995833333278,0.8715914484126921,0.8933318061811949,0.8737879900726819
|
| 100 |
+
4.125177809388336,5800,0.828125,0.902575,0.9298,0.828125,0.828125,0.3008583333333333,0.902575,0.18596000000000004,0.9298,0.828125,0.8674095833333273,0.871657361111104,0.8933427952172275,0.8738836312779211
|
| 101 |
+
4.196301564722617,5900,0.8281,0.901525,0.930075,0.8281,0.8281,0.30050833333333327,0.901525,0.18601500000000004,0.930075,0.8281,0.8673141666666621,0.8715502579365025,0.8932912071203153,0.8737613329234138
|
| 102 |
+
4.2674253200568995,6000,0.8285,0.90215,0.9301,0.8285,0.8285,0.30071666666666663,0.90215,0.18602000000000002,0.9301,0.8285,0.8677162499999951,0.8719605952380887,0.8935923803057578,0.8741910760623423
|
| 103 |
+
4.338549075391181,6100,0.828225,0.9024,0.929725,0.828225,0.828225,0.3008,0.9024,0.185945,0.929725,0.828225,0.8673716666666614,0.8717138293650726,0.8934810443914722,0.873907773748694
|
| 104 |
+
4.409672830725462,6200,0.828075,0.902125,0.929625,0.828075,0.828075,0.3007083333333333,0.902125,0.18592500000000003,0.929625,0.828075,0.8672837499999961,0.8716056646825336,0.8933245645232728,0.8738314386660578
|
| 105 |
+
4.480796586059744,6300,0.8285,0.902475,0.930375,0.8285,0.8285,0.300825,0.902475,0.186075,0.930375,0.8285,0.8677258333333291,0.8719892063492011,0.8937460890492909,0.8741866281708112
|
| 106 |
+
4.551920341394026,6400,0.829025,0.9031,0.930125,0.829025,0.829025,0.3010333333333333,0.9031,0.18602500000000005,0.930125,0.829025,0.868045833333328,0.8723535912698358,0.8939942369427725,0.8745777316869345
|
| 107 |
+
4.623044096728307,6500,0.828925,0.902675,0.93025,0.828925,0.828925,0.3008916666666666,0.902675,0.18605000000000002,0.93025,0.828925,0.8679429166666623,0.8722636011904705,0.894011390451915,0.8744507896495303
|
| 108 |
+
4.694167852062589,6600,0.828775,0.902675,0.930525,0.828775,0.828775,0.3008916666666666,0.902675,0.18610500000000002,0.930525,0.828775,0.8679779166666611,0.872248293650788,0.8939521912453529,0.874467781370097
|
| 109 |
+
4.76529160739687,6700,0.82895,0.90245,0.93035,0.82895,0.82895,0.3008166666666666,0.90245,0.18607000000000004,0.93035,0.82895,0.8679879166666622,0.8723038789682495,0.8940473756170835,0.8744915171684532
|
| 110 |
+
4.836415362731152,6800,0.8283,0.90305,0.9309,0.8283,0.8283,0.3010166666666666,0.90305,0.18618,0.9309,0.8283,0.867910416666662,0.8721540773809471,0.8939497437147061,0.8743487656718353
|
| 111 |
+
4.907539118065434,6900,0.828325,0.902925,0.93045,0.828325,0.828325,0.30097499999999994,0.902925,0.18609000000000003,0.93045,0.828325,0.8678854166666622,0.8721938888888839,0.8939514220767489,0.8744030273602739
|
| 112 |
+
4.978662873399715,7000,0.8283,0.903225,0.930375,0.8283,0.8283,0.301075,0.903225,0.18607500000000002,0.930375,0.8283,0.8678333333333287,0.8721642956349152,0.8939665996423323,0.8743696253316319
|
| 113 |
+
5.049786628733997,7100,0.827825,0.9032,0.930075,0.827825,0.827825,0.3010666666666666,0.9032,0.186015,0.930075,0.827825,0.8675479166666622,0.8719574007936451,0.8938974126211886,0.8741217627464118
|
| 114 |
+
5.120910384068279,7200,0.828225,0.903075,0.93035,0.828225,0.828225,0.30102499999999993,0.903075,0.18607000000000004,0.93035,0.828225,0.8677999999999951,0.8721783134920578,0.894072625572504,0.8743447352455296
|
| 115 |
+
5.19203413940256,7300,0.827875,0.903275,0.93085,0.827875,0.827875,0.3010916666666666,0.903275,0.18617000000000003,0.93085,0.827875,0.8677724999999952,0.8720725595238032,0.893982576896927,0.8742551897684425
|
| 116 |
+
5.2631578947368425,7400,0.82855,0.903125,0.930825,0.82855,0.82855,0.30104166666666665,0.903125,0.186165,0.930825,0.82855,0.8680941666666616,0.8724458630952325,0.8943261872169521,0.8746091010205986
|
| 117 |
+
5.334281650071124,7500,0.8282,0.9027,0.9307,0.8282,0.8282,0.3009,0.9027,0.18614000000000003,0.9307,0.8282,0.8677720833333288,0.8721882242063433,0.8942269763046193,0.8743096835645798
|
| 118 |
+
5.405405405405405,7600,0.828575,0.90245,0.930475,0.828575,0.828575,0.3008166666666667,0.90245,0.18609500000000004,0.930475,0.828575,0.8679016666666634,0.8722942460317409,0.894169304471156,0.8744688088000298
|
| 119 |
+
5.476529160739687,7700,0.82835,0.90285,0.9306,0.82835,0.82835,0.3009499999999999,0.90285,0.18612000000000004,0.9306,0.82835,0.8678508333333291,0.8722199404761842,0.8941142458749253,0.8744018627520107
|
| 120 |
+
5.547652916073969,7800,0.828625,0.903475,0.930225,0.828625,0.828625,0.3011583333333333,0.903475,0.18604500000000002,0.930225,0.828625,0.8680270833333281,0.8724584722222167,0.8942713349975829,0.8746604618279742
|
| 121 |
+
5.61877667140825,7900,0.82875,0.902825,0.930875,0.82875,0.82875,0.3009416666666666,0.902825,0.186175,0.930875,0.82875,0.8681183333333286,0.8724924107142793,0.8944002954247615,0.8746483877607266
|
| 122 |
+
5.689900426742532,8000,0.82915,0.903525,0.93085,0.82915,0.82915,0.301175,0.903525,0.18617000000000003,0.93085,0.82915,0.8684824999999949,0.8728611805555482,0.8947027956822773,0.8750076119907729
|
| 123 |
+
5.761024182076814,8100,0.829025,0.902775,0.9308,0.829025,0.829025,0.3009249999999999,0.902775,0.18616000000000005,0.9308,0.829025,0.8682724999999946,0.8727015178571357,0.8946280631253088,0.8748339744524721
|
| 124 |
+
5.832147937411095,8200,0.8286,0.903225,0.9308,0.8286,0.8286,0.3010749999999999,0.903225,0.18616000000000002,0.9308,0.8286,0.8680666666666624,0.8724622420634867,0.8944261476913996,0.8745954232847529
|
| 125 |
+
5.903271692745377,8300,0.82895,0.903725,0.930775,0.82895,0.82895,0.30124166666666663,0.903725,0.18615500000000001,0.930775,0.82895,0.8683333333333292,0.8727454464285668,0.8946178536849202,0.8748963303338884
|
| 126 |
+
5.974395448079658,8400,0.8288,0.9036,0.9312,0.8288,0.8288,0.3011999999999999,0.9036,0.18624000000000004,0.9312,0.8288,0.8683545833333292,0.8727151686507881,0.8946702577402386,0.8748394145685222
|
| 127 |
+
6.0455192034139404,8500,0.829175,0.90385,0.93095,0.829175,0.829175,0.30128333333333324,0.90385,0.18619000000000002,0.93095,0.829175,0.8685420833333285,0.872971696428565,0.8948846695179693,0.8750987853504063
|
| 128 |
+
6.116642958748222,8600,0.829,0.903525,0.9313,0.829,0.829,0.301175,0.903525,0.18626000000000004,0.9313,0.829,0.8685441666666623,0.8729358035714229,0.8949065301167233,0.87504230073798
|
| 129 |
+
6.187766714082503,8700,0.829075,0.903825,0.931125,0.829075,0.829075,0.301275,0.903825,0.18622500000000003,0.931125,0.829075,0.8685433333333281,0.8729486210317393,0.894908715557474,0.8750547367584172
|
| 130 |
+
6.2588904694167855,8800,0.8294,0.9034,0.931025,0.8294,0.8294,0.3011333333333333,0.9034,0.186205,0.931025,0.8294,0.8687245833333284,0.8731362599206285,0.8950249083831672,0.8752570268200927
|
| 131 |
+
6.330014224751067,8900,0.8291,0.904175,0.93095,0.8291,0.8291,0.3013916666666666,0.904175,0.18619,0.93095,0.8291,0.8686062499999953,0.873074950396819,0.8950197884709341,0.875183958791522
|
| 132 |
+
6.401137980085348,9000,0.829075,0.903525,0.9311,0.829075,0.829075,0.30117499999999997,0.903525,0.18622000000000002,0.9311,0.829075,0.8684637499999962,0.8728336607142791,0.8947201341464351,0.8749830024617687
|
| 133 |
+
6.472261735419631,9100,0.82895,0.903925,0.931125,0.82895,0.82895,0.3013083333333333,0.903925,0.18622500000000006,0.931125,0.82895,0.868523749999995,0.8729523313491995,0.8949264125232629,0.8750632591156137
|
| 134 |
+
6.543385490753912,9200,0.8287,0.903875,0.931525,0.8287,0.8287,0.3012916666666666,0.903875,0.18630500000000003,0.931525,0.8287,0.8685312499999951,0.8728960416666594,0.8948964473390973,0.8750060398957655
|
| 135 |
+
6.614509246088193,9300,0.828575,0.9037,0.930975,0.828575,0.828575,0.30123333333333324,0.9037,0.18619500000000003,0.930975,0.828575,0.8682566666666615,0.8727102876984051,0.8947419499833495,0.8748258780472974
|
| 136 |
+
6.685633001422475,9400,0.828725,0.903625,0.9312,0.828725,0.828725,0.3012083333333333,0.903625,0.18624000000000004,0.9312,0.828725,0.8684295833333284,0.8728350694444372,0.8948174140374873,0.8749621158488419
|
| 137 |
+
6.756756756756757,9500,0.828925,0.903825,0.9316,0.828925,0.828925,0.30127499999999996,0.903825,0.18632000000000004,0.9316,0.828925,0.8686979166666619,0.8730864484126918,0.8950858390267745,0.875175061860045
|
| 138 |
+
6.827880512091038,9600,0.828925,0.904025,0.931325,0.828925,0.828925,0.3013416666666666,0.904025,0.18626500000000004,0.931325,0.828925,0.8686649999999941,0.8730839285714214,0.8950756231504231,0.8751769766045896
|
| 139 |
+
6.89900426742532,9700,0.829325,0.90415,0.931525,0.829325,0.829325,0.3013833333333333,0.90415,0.18630500000000003,0.931525,0.829325,0.8689170833333283,0.8733062301587239,0.8952108076269908,0.8754240592488881
|
| 140 |
+
6.970128022759602,9800,0.829225,0.9039,0.9316,0.829225,0.829225,0.3013,0.9039,0.18632000000000004,0.9316,0.829225,0.8688504166666607,0.8732341468253897,0.8951715488614492,0.8753505377411699
|
| 141 |
+
7.0412517780938835,9900,0.828825,0.9036,0.931525,0.828825,0.828825,0.30119999999999997,0.9036,0.18630500000000003,0.931525,0.828825,0.8685466666666609,0.8729646825396757,0.8950075880619315,0.8750548303490677
|
| 142 |
+
7.112375533428165,10000,0.829125,0.903825,0.9317,0.829125,0.829125,0.301275,0.903825,0.18634,0.9317,0.829125,0.8687299999999943,0.8731202976190414,0.8951116703189326,0.8752221802611312
|
| 143 |
+
7.183499288762446,10100,0.8292,0.90395,0.932,0.8292,0.8292,0.3013166666666666,0.90395,0.18640000000000004,0.932,0.8292,0.8688920833333279,0.8732386111111037,0.8952233766688119,0.8753352290567349
|
| 144 |
+
7.2546230440967285,10200,0.829225,0.90445,0.93195,0.829225,0.829225,0.3014833333333333,0.90445,0.18639,0.93195,0.829225,0.8689337499999943,0.87331557539682,0.8953072675203337,0.8754143962131168
|
| 145 |
+
7.32574679943101,10300,0.828975,0.9043,0.9317,0.828975,0.828975,0.3014333333333333,0.9043,0.18634000000000003,0.9317,0.828975,0.8687420833333279,0.8731662003968191,0.8952068599802827,0.875254118597502
|
| 146 |
+
7.396870554765291,10400,0.829125,0.903825,0.932025,0.829125,0.829125,0.30127499999999996,0.903825,0.18640500000000002,0.932025,0.829125,0.8687904166666616,0.8731562202380894,0.8951799564271696,0.8752576132029881
|
| 147 |
+
7.467994310099574,10500,0.829125,0.90375,0.9321,0.829125,0.829125,0.30124999999999996,0.90375,0.18642000000000003,0.9321,0.829125,0.8687862499999952,0.8731379265872954,0.8951780327010416,0.8752325887196015
|
| 148 |
+
7.539118065433855,10600,0.82935,0.904225,0.932225,0.82935,0.82935,0.3014083333333333,0.904225,0.18644500000000003,0.932225,0.82935,0.8689916666666612,0.8733285813491991,0.8953223283998646,0.8754304197917856
|
| 149 |
+
7.610241820768136,10700,0.828875,0.904125,0.9319,0.828875,0.828875,0.30137499999999995,0.904125,0.18638000000000002,0.9319,0.828875,0.8686462499999948,0.8730549603174542,0.8951443508540531,0.8751426179544144
|
| 150 |
+
7.681365576102419,10800,0.8294,0.90425,0.932125,0.8294,0.8294,0.3014166666666666,0.90425,0.18642500000000004,0.932125,0.8294,0.8689254166666617,0.8733101587301526,0.8953412645129609,0.8753982744981171
|
| 151 |
+
7.7524893314367,10900,0.829525,0.904075,0.931875,0.829525,0.829525,0.3013583333333333,0.904075,0.18637500000000004,0.931875,0.829525,0.869002916666662,0.8734263988095179,0.8954372160730756,0.87550507311346
|
| 152 |
+
7.823613086770981,11000,0.829225,0.904325,0.93195,0.829225,0.829225,0.30144166666666666,0.904325,0.18639000000000003,0.93195,0.829225,0.8689704166666619,0.8734005753968189,0.8954592783354139,0.875473740734129
|
| 153 |
+
7.894736842105263,11100,0.829375,0.904325,0.931975,0.829375,0.829375,0.3014416666666666,0.904325,0.18639500000000003,0.931975,0.829375,0.8690224999999943,0.8734514186507863,0.8954979567496186,0.8755274020925966
|
| 154 |
+
7.965860597439545,11200,0.82955,0.9045,0.93215,0.82955,0.82955,0.30149999999999993,0.9045,0.18643,0.93215,0.82955,0.8692179166666613,0.8736397519841202,0.8956820121323833,0.8757002350962841
|
| 155 |
+
8.036984352773827,11300,0.82955,0.904175,0.9324,0.82955,0.82955,0.30139166666666667,0.904175,0.18648,0.9324,0.82955,0.8691774999999948,0.8735149603174545,0.8955172674364549,0.8756027777748313
|
| 156 |
+
8.108108108108109,11400,0.829325,0.90435,0.932425,0.829325,0.829325,0.30144999999999994,0.90435,0.18648500000000004,0.932425,0.829325,0.8691216666666615,0.8734770238095176,0.8955123296939637,0.8755540663940554
|
| 157 |
+
8.17923186344239,11500,0.82905,0.9043,0.9326,0.82905,0.82905,0.3014333333333333,0.9043,0.18652000000000002,0.9326,0.82905,0.8689616666666612,0.8733028571428504,0.8954049605787512,0.8753753151688954
|
| 158 |
+
8.250355618776672,11600,0.829575,0.904075,0.93235,0.829575,0.829575,0.3013583333333333,0.904075,0.18647000000000002,0.93235,0.829575,0.8692079166666619,0.8735856150793588,0.8955883966928999,0.8756651485608432
|
| 159 |
+
8.321479374110954,11700,0.829,0.90435,0.932375,0.829,0.829,0.30144999999999994,0.90435,0.186475,0.932375,0.829,0.8689162499999953,0.8733047916666604,0.8954331340853403,0.8753585929214951
|
| 160 |
+
8.392603129445234,11800,0.82955,0.9045,0.932425,0.82955,0.82955,0.3015,0.9045,0.18648500000000004,0.932425,0.82955,0.8692320833333294,0.8736002976190421,0.8956159725223579,0.8756756360478772
|
| 161 |
+
8.463726884779517,11900,0.829575,0.90465,0.932525,0.829575,0.829575,0.30154999999999993,0.90465,0.18650500000000003,0.932525,0.829575,0.8693274999999956,0.8736795238095177,0.8957038979386316,0.8757370242050787
|
| 162 |
+
8.534850640113799,12000,0.8295,0.90455,0.932525,0.8295,0.8295,0.30151666666666666,0.90455,0.18650500000000003,0.932525,0.8295,0.8693037499999957,0.8736758134920576,0.8957443910788215,0.8757191218855241
|
| 163 |
+
8.60597439544808,12100,0.829375,0.9046,0.932475,0.829375,0.829375,0.3015333333333333,0.9046,0.18649500000000002,0.932475,0.829375,0.869249999999996,0.8736443948412647,0.8957296640580314,0.8756884601745238
|
| 164 |
+
8.677098150782362,12200,0.829675,0.904425,0.9326,0.829675,0.829675,0.301475,0.904425,0.18652000000000002,0.9326,0.829675,0.8693591666666626,0.8736971329365028,0.8957079678384477,0.8757613361872285
|
| 165 |
+
8.748221906116642,12300,0.829675,0.90445,0.93235,0.829675,0.829675,0.3014833333333333,0.90445,0.18647,0.93235,0.829675,0.8693087499999952,0.8737007142857086,0.895718161327572,0.8757683518013507
|
| 166 |
+
8.819345661450924,12400,0.8295,0.904375,0.932575,0.8295,0.8295,0.30145833333333333,0.904375,0.18651500000000001,0.932575,0.8295,0.8692854166666626,0.8736491765872961,0.8957043908845526,0.875705749877939
|
| 167 |
+
8.890469416785207,12500,0.829525,0.904375,0.932475,0.829525,0.829525,0.3014583333333333,0.904375,0.18649500000000002,0.932475,0.829525,0.8692191666666625,0.8735937202380893,0.895662154246233,0.8756445157255779
|
| 168 |
+
8.961593172119487,12600,0.829475,0.904575,0.9323,0.829475,0.829475,0.30152499999999993,0.904575,0.18646000000000004,0.9323,0.829475,0.8691791666666621,0.8736097519841208,0.8957145356259026,0.8756490977339323
|
| 169 |
+
9.03271692745377,12700,0.82945,0.90425,0.932675,0.82945,0.82945,0.30141666666666667,0.90425,0.18653499999999998,0.932675,0.82945,0.869237916666663,0.8735875793650737,0.8956664069313527,0.8756409919548603
|
| 170 |
+
9.103840682788052,12800,0.8294,0.90425,0.932675,0.8294,0.8294,0.30141666666666667,0.90425,0.18653500000000003,0.932675,0.8294,0.8691570833333291,0.8734929662698362,0.8955523610447522,0.8755622615987919
|
| 171 |
+
9.174964438122332,12900,0.82915,0.904475,0.932675,0.82915,0.82915,0.3014916666666666,0.904475,0.186535,0.932675,0.82915,0.8690449999999957,0.8733805357142803,0.8954706666940537,0.8754538277866332
|
| 172 |
+
9.246088193456615,13000,0.82935,0.904525,0.9326,0.82935,0.82935,0.3015083333333333,0.904525,0.18652000000000002,0.9326,0.82935,0.8691433333333289,0.873507013888883,0.8955913527336096,0.8755684032749883
|
| 173 |
+
9.317211948790897,13100,0.8295,0.904475,0.932675,0.8295,0.8295,0.3014916666666666,0.904475,0.186535,0.932675,0.8295,0.869243749999996,0.8735823710317403,0.8956395313586778,0.8756462995333785
|
| 174 |
+
9.388335704125177,13200,0.829575,0.90465,0.9329,0.829575,0.829575,0.30155,0.90465,0.18658000000000002,0.9329,0.829575,0.869395416666663,0.8736798313492016,0.8956789415552683,0.8757611231012578
|
| 175 |
+
9.45945945945946,13300,0.8294,0.904625,0.93245,0.8294,0.8294,0.3015416666666666,0.904625,0.18649000000000004,0.93245,0.8294,0.8691949999999963,0.8735664980158677,0.8956132086027158,0.8756355010782446
|
| 176 |
+
9.530583214793742,13400,0.8295,0.904725,0.932475,0.8295,0.8295,0.3015749999999999,0.904725,0.18649500000000002,0.932475,0.8295,0.8692941666666627,0.873653789682535,0.8956503675255127,0.8757333507075223
|
| 177 |
+
9.601706970128022,13500,0.829525,0.904725,0.932375,0.829525,0.829525,0.301575,0.904725,0.18647500000000003,0.932375,0.829525,0.8692958333333296,0.8736637499999952,0.8956465103594379,0.8757491390118639
|
| 178 |
+
9.672830725462305,13600,0.82965,0.90465,0.9325,0.82965,0.82965,0.30154999999999993,0.90465,0.1865,0.9325,0.82965,0.8693666666666627,0.8737252380952334,0.8957081100848372,0.8758065761725513
|
| 179 |
+
9.743954480796585,13700,0.8295,0.90465,0.9325,0.8295,0.8295,0.30155,0.90465,0.18650000000000003,0.9325,0.8295,0.8693074999999961,0.8736817162698364,0.895700486958084,0.8757526814508423
|
| 180 |
+
9.815078236130867,13800,0.8296,0.904725,0.932475,0.8296,0.8296,0.301575,0.904725,0.18649500000000002,0.932475,0.8296,0.8693120833333294,0.8736897916666616,0.8957134946126672,0.8757585053341062
|
| 181 |
+
9.88620199146515,13900,0.829475,0.904825,0.932425,0.829475,0.829475,0.3016083333333333,0.904825,0.186485,0.932425,0.829475,0.8692816666666627,0.8736694444444394,0.8957054405004242,0.875738183071688
|
| 182 |
+
9.95732574679943,14000,0.829675,0.9048,0.93245,0.829675,0.829675,0.3016,0.9048,0.18649000000000004,0.93245,0.829675,0.8693824999999958,0.8737714285714238,0.8957919450437679,0.8758361833602419
|
final_metrics.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
{
|
| 2 |
-
"val_cosine_accuracy@1": 0.
|
| 3 |
-
"val_cosine_accuracy@3": 0.
|
| 4 |
-
"val_cosine_accuracy@5": 0.
|
| 5 |
-
"val_cosine_precision@1": 0.
|
| 6 |
-
"val_cosine_precision@3": 0.
|
| 7 |
-
"val_cosine_precision@5": 0.
|
| 8 |
-
"val_cosine_recall@1": 0.
|
| 9 |
-
"val_cosine_recall@3": 0.
|
| 10 |
-
"val_cosine_recall@5": 0.
|
| 11 |
-
"val_cosine_ndcg@10": 0.
|
| 12 |
-
"val_cosine_mrr@1": 0.
|
| 13 |
-
"val_cosine_mrr@5": 0.
|
| 14 |
-
"val_cosine_mrr@10": 0.
|
| 15 |
-
"val_cosine_map@100": 0.
|
| 16 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"val_cosine_accuracy@1": 0.9184,
|
| 3 |
+
"val_cosine_accuracy@3": 0.97,
|
| 4 |
+
"val_cosine_accuracy@5": 0.9852,
|
| 5 |
+
"val_cosine_precision@1": 0.9184,
|
| 6 |
+
"val_cosine_precision@3": 0.3233333333333333,
|
| 7 |
+
"val_cosine_precision@5": 0.19703999999999997,
|
| 8 |
+
"val_cosine_recall@1": 0.9184,
|
| 9 |
+
"val_cosine_recall@3": 0.97,
|
| 10 |
+
"val_cosine_recall@5": 0.9852,
|
| 11 |
+
"val_cosine_ndcg@10": 0.9586270476620361,
|
| 12 |
+
"val_cosine_mrr@1": 0.9184,
|
| 13 |
+
"val_cosine_mrr@5": 0.9451366666666663,
|
| 14 |
+
"val_cosine_mrr@10": 0.9466038095238087,
|
| 15 |
+
"val_cosine_map@100": 0.946959374340519
|
| 16 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 114011616
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4ae394348a491d3e3613014bc35b25e1e57987b52b334312c28991a2034cb6a2
|
| 3 |
size 114011616
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 6161
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:767f865ad778a3cf6f884deef6446e71dd8ddc58b4697f307d5df2876e65b529
|
| 3 |
size 6161
|