Training in progress, step 14060
Browse files- Information-Retrieval_evaluation_val_results.csv +1 -0
- README.md +71 -347
- eval/Information-Retrieval_evaluation_val_results.csv +141 -0
- final_metrics.json +14 -14
- model.safetensors +1 -1
Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -1,3 +1,4 @@
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.908,0.9684,0.9834,0.908,0.908,0.3228,0.9684,0.19667999999999997,0.9834,0.908,0.9386633333333337,0.9400269841269848,0.9532296698470627,0.9404621256346036
|
| 3 |
-1,-1,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402433333333332,0.9416250793650793,0.9545809774353143,0.9420576026548708
|
|
|
|
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.908,0.9684,0.9834,0.908,0.908,0.3228,0.9684,0.19667999999999997,0.9834,0.908,0.9386633333333337,0.9400269841269848,0.9532296698470627,0.9404621256346036
|
| 3 |
-1,-1,0.9104,0.9688,0.9842,0.9104,0.9104,0.32293333333333335,0.9688,0.19683999999999996,0.9842,0.9104,0.9402433333333332,0.9416250793650793,0.9545809774353143,0.9420576026548708
|
| 4 |
+
-1,-1,0.8281,0.9026,0.93105,0.8281,0.8281,0.3008666666666666,0.9026,0.18621000000000004,0.93105,0.8281,0.8677437499999962,0.8721381249999942,0.8942437004811851,0.874246358340888
|
README.md
CHANGED
|
@@ -5,114 +5,38 @@ tags:
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
-
- dataset_size:
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
-
- source_sentence:
|
| 13 |
-
for one that's not married? Which one is for what?
|
| 14 |
sentences:
|
| 15 |
-
-
|
| 16 |
-
|
| 17 |
-
-
|
| 18 |
-
|
| 19 |
-
- source_sentence: Which ointment is applied to the face of UFC fighters at the commencement
|
| 20 |
-
of a bout? What does it do?
|
| 21 |
sentences:
|
| 22 |
-
-
|
| 23 |
-
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
- source_sentence:
|
| 27 |
sentences:
|
| 28 |
-
-
|
| 29 |
-
-
|
| 30 |
-
-
|
| 31 |
-
|
| 32 |
-
- source_sentence: Can I do shoulder and triceps workout on same day? What other combinations
|
| 33 |
-
like this can I do?
|
| 34 |
sentences:
|
| 35 |
-
-
|
| 36 |
-
|
| 37 |
-
-
|
| 38 |
-
|
| 39 |
-
this can I do?
|
| 40 |
-
- source_sentence: I am a married woman and I'm in love with married man. what should
|
| 41 |
-
I do?
|
| 42 |
sentences:
|
| 43 |
-
-
|
| 44 |
-
- I
|
| 45 |
-
|
| 46 |
-
- I am a married woman and I'm in love with married man. what should I do?
|
| 47 |
pipeline_tag: sentence-similarity
|
| 48 |
library_name: sentence-transformers
|
| 49 |
-
metrics:
|
| 50 |
-
- cosine_accuracy@1
|
| 51 |
-
- cosine_accuracy@3
|
| 52 |
-
- cosine_accuracy@5
|
| 53 |
-
- cosine_precision@1
|
| 54 |
-
- cosine_precision@3
|
| 55 |
-
- cosine_precision@5
|
| 56 |
-
- cosine_recall@1
|
| 57 |
-
- cosine_recall@3
|
| 58 |
-
- cosine_recall@5
|
| 59 |
-
- cosine_ndcg@10
|
| 60 |
-
- cosine_mrr@1
|
| 61 |
-
- cosine_mrr@5
|
| 62 |
-
- cosine_mrr@10
|
| 63 |
-
- cosine_map@100
|
| 64 |
-
model-index:
|
| 65 |
-
- name: SentenceTransformer based on prajjwal1/bert-small
|
| 66 |
-
results:
|
| 67 |
-
- task:
|
| 68 |
-
type: information-retrieval
|
| 69 |
-
name: Information Retrieval
|
| 70 |
-
dataset:
|
| 71 |
-
name: val
|
| 72 |
-
type: val
|
| 73 |
-
metrics:
|
| 74 |
-
- type: cosine_accuracy@1
|
| 75 |
-
value: 0.828025
|
| 76 |
-
name: Cosine Accuracy@1
|
| 77 |
-
- type: cosine_accuracy@3
|
| 78 |
-
value: 0.9027
|
| 79 |
-
name: Cosine Accuracy@3
|
| 80 |
-
- type: cosine_accuracy@5
|
| 81 |
-
value: 0.931025
|
| 82 |
-
name: Cosine Accuracy@5
|
| 83 |
-
- type: cosine_precision@1
|
| 84 |
-
value: 0.828025
|
| 85 |
-
name: Cosine Precision@1
|
| 86 |
-
- type: cosine_precision@3
|
| 87 |
-
value: 0.3008999999999999
|
| 88 |
-
name: Cosine Precision@3
|
| 89 |
-
- type: cosine_precision@5
|
| 90 |
-
value: 0.186205
|
| 91 |
-
name: Cosine Precision@5
|
| 92 |
-
- type: cosine_recall@1
|
| 93 |
-
value: 0.828025
|
| 94 |
-
name: Cosine Recall@1
|
| 95 |
-
- type: cosine_recall@3
|
| 96 |
-
value: 0.9027
|
| 97 |
-
name: Cosine Recall@3
|
| 98 |
-
- type: cosine_recall@5
|
| 99 |
-
value: 0.931025
|
| 100 |
-
name: Cosine Recall@5
|
| 101 |
-
- type: cosine_ndcg@10
|
| 102 |
-
value: 0.8942284691055087
|
| 103 |
-
name: Cosine Ndcg@10
|
| 104 |
-
- type: cosine_mrr@1
|
| 105 |
-
value: 0.828025
|
| 106 |
-
name: Cosine Mrr@1
|
| 107 |
-
- type: cosine_mrr@5
|
| 108 |
-
value: 0.8677179166666629
|
| 109 |
-
name: Cosine Mrr@5
|
| 110 |
-
- type: cosine_mrr@10
|
| 111 |
-
value: 0.8721162896825339
|
| 112 |
-
name: Cosine Mrr@10
|
| 113 |
-
- type: cosine_map@100
|
| 114 |
-
value: 0.8742240723304836
|
| 115 |
-
name: Cosine Map@100
|
| 116 |
---
|
| 117 |
|
| 118 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
@@ -161,12 +85,12 @@ Then you can load this model and run inference.
|
|
| 161 |
from sentence_transformers import SentenceTransformer
|
| 162 |
|
| 163 |
# Download from the 🤗 Hub
|
| 164 |
-
model = SentenceTransformer("
|
| 165 |
# Run inference
|
| 166 |
sentences = [
|
| 167 |
-
|
| 168 |
-
|
| 169 |
-
|
| 170 |
]
|
| 171 |
embeddings = model.encode(sentences)
|
| 172 |
print(embeddings.shape)
|
|
@@ -175,9 +99,9 @@ print(embeddings.shape)
|
|
| 175 |
# Get the similarity scores for the embeddings
|
| 176 |
similarities = model.similarity(embeddings, embeddings)
|
| 177 |
print(similarities)
|
| 178 |
-
# tensor([[1.0000,
|
| 179 |
-
# [
|
| 180 |
-
# [0.
|
| 181 |
```
|
| 182 |
|
| 183 |
<!--
|
|
@@ -204,32 +128,6 @@ You can finetune this model on your own dataset.
|
|
| 204 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 205 |
-->
|
| 206 |
|
| 207 |
-
## Evaluation
|
| 208 |
-
|
| 209 |
-
### Metrics
|
| 210 |
-
|
| 211 |
-
#### Information Retrieval
|
| 212 |
-
|
| 213 |
-
* Dataset: `val`
|
| 214 |
-
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
| 215 |
-
|
| 216 |
-
| Metric | Value |
|
| 217 |
-
|:-------------------|:-----------|
|
| 218 |
-
| cosine_accuracy@1 | 0.828 |
|
| 219 |
-
| cosine_accuracy@3 | 0.9027 |
|
| 220 |
-
| cosine_accuracy@5 | 0.931 |
|
| 221 |
-
| cosine_precision@1 | 0.828 |
|
| 222 |
-
| cosine_precision@3 | 0.3009 |
|
| 223 |
-
| cosine_precision@5 | 0.1862 |
|
| 224 |
-
| cosine_recall@1 | 0.828 |
|
| 225 |
-
| cosine_recall@3 | 0.9027 |
|
| 226 |
-
| cosine_recall@5 | 0.931 |
|
| 227 |
-
| **cosine_ndcg@10** | **0.8942** |
|
| 228 |
-
| cosine_mrr@1 | 0.828 |
|
| 229 |
-
| cosine_mrr@5 | 0.8677 |
|
| 230 |
-
| cosine_mrr@10 | 0.8721 |
|
| 231 |
-
| cosine_map@100 | 0.8742 |
|
| 232 |
-
|
| 233 |
<!--
|
| 234 |
## Bias, Risks and Limitations
|
| 235 |
|
|
@@ -248,45 +146,19 @@ You can finetune this model on your own dataset.
|
|
| 248 |
|
| 249 |
#### Unnamed Dataset
|
| 250 |
|
| 251 |
-
* Size:
|
| 252 |
-
* Columns: <code>
|
| 253 |
-
* Approximate statistics based on the first 1000 samples:
|
| 254 |
-
| | anchor | positive | negative |
|
| 255 |
-
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
| 256 |
-
| type | string | string | string |
|
| 257 |
-
| details | <ul><li>min: 4 tokens</li><li>mean: 15.46 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 15.52 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 16.63 tokens</li><li>max: 59 tokens</li></ul> |
|
| 258 |
-
* Samples:
|
| 259 |
-
| anchor | positive | negative |
|
| 260 |
-
|:--------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|
|
| 261 |
-
| <code>Shall I upgrade my iPhone 5s to iOS 10 final version?</code> | <code>Should I upgrade an iPhone 5s to iOS 10?</code> | <code>Shall my iPhone 5s upgrade Ito iOS 10 final version?</code> |
|
| 262 |
-
| <code>Is Donald Trump really going to be the president of United States?</code> | <code>Do you think Donald Trump could conceivably be the next President of the United States?</code> | <code>Is Donald Trump really going not to be the president of United States ?</code> |
|
| 263 |
-
| <code>What are real tips to improve work life balance?</code> | <code>What are the best ways to create a work life balance?</code> | <code>How far is Miami from Fort Lauderdale?</code> |
|
| 264 |
-
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 265 |
-
```json
|
| 266 |
-
{
|
| 267 |
-
"scale": 20.0,
|
| 268 |
-
"similarity_fct": "cos_sim",
|
| 269 |
-
"gather_across_devices": false
|
| 270 |
-
}
|
| 271 |
-
```
|
| 272 |
-
|
| 273 |
-
### Evaluation Dataset
|
| 274 |
-
|
| 275 |
-
#### Unnamed Dataset
|
| 276 |
-
|
| 277 |
-
* Size: 40,000 evaluation samples
|
| 278 |
-
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
| 279 |
* Approximate statistics based on the first 1000 samples:
|
| 280 |
-
| |
|
| 281 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
| 282 |
| type | string | string | string |
|
| 283 |
-
| details | <ul><li>min: 6 tokens</li><li>mean: 15.
|
| 284 |
* Samples:
|
| 285 |
-
|
|
| 286 |
-
|
| 287 |
-
| <code>
|
| 288 |
-
| <code>
|
| 289 |
-
| <code>
|
| 290 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 291 |
```json
|
| 292 |
{
|
|
@@ -299,49 +171,36 @@ You can finetune this model on your own dataset.
|
|
| 299 |
### Training Hyperparameters
|
| 300 |
#### Non-Default Hyperparameters
|
| 301 |
|
| 302 |
-
- `
|
| 303 |
-
- `
|
| 304 |
-
- `per_device_eval_batch_size`: 256
|
| 305 |
-
- `learning_rate`: 2e-05
|
| 306 |
-
- `weight_decay`: 0.001
|
| 307 |
-
- `max_steps`: 14060
|
| 308 |
-
- `warmup_ratio`: 0.1
|
| 309 |
- `fp16`: True
|
| 310 |
-
- `
|
| 311 |
-
- `dataloader_num_workers`: 1
|
| 312 |
-
- `dataloader_prefetch_factor`: 1
|
| 313 |
-
- `load_best_model_at_end`: True
|
| 314 |
-
- `optim`: adamw_torch
|
| 315 |
-
- `ddp_find_unused_parameters`: False
|
| 316 |
-
- `push_to_hub`: True
|
| 317 |
-
- `hub_model_id`: redis/model-b-structured
|
| 318 |
-
- `eval_on_start`: True
|
| 319 |
|
| 320 |
#### All Hyperparameters
|
| 321 |
<details><summary>Click to expand</summary>
|
| 322 |
|
| 323 |
- `overwrite_output_dir`: False
|
| 324 |
- `do_predict`: False
|
| 325 |
-
- `eval_strategy`:
|
| 326 |
- `prediction_loss_only`: True
|
| 327 |
-
- `per_device_train_batch_size`:
|
| 328 |
-
- `per_device_eval_batch_size`:
|
| 329 |
- `per_gpu_train_batch_size`: None
|
| 330 |
- `per_gpu_eval_batch_size`: None
|
| 331 |
- `gradient_accumulation_steps`: 1
|
| 332 |
- `eval_accumulation_steps`: None
|
| 333 |
- `torch_empty_cache_steps`: None
|
| 334 |
-
- `learning_rate`:
|
| 335 |
-
- `weight_decay`: 0.
|
| 336 |
- `adam_beta1`: 0.9
|
| 337 |
- `adam_beta2`: 0.999
|
| 338 |
- `adam_epsilon`: 1e-08
|
| 339 |
-
- `max_grad_norm`: 1
|
| 340 |
-
- `num_train_epochs`: 3
|
| 341 |
-
- `max_steps`:
|
| 342 |
- `lr_scheduler_type`: linear
|
| 343 |
- `lr_scheduler_kwargs`: {}
|
| 344 |
-
- `warmup_ratio`: 0.
|
| 345 |
- `warmup_steps`: 0
|
| 346 |
- `log_level`: passive
|
| 347 |
- `log_level_replica`: warning
|
|
@@ -369,14 +228,14 @@ You can finetune this model on your own dataset.
|
|
| 369 |
- `tpu_num_cores`: None
|
| 370 |
- `tpu_metrics_debug`: False
|
| 371 |
- `debug`: []
|
| 372 |
-
- `dataloader_drop_last`:
|
| 373 |
-
- `dataloader_num_workers`:
|
| 374 |
-
- `dataloader_prefetch_factor`:
|
| 375 |
- `past_index`: -1
|
| 376 |
- `disable_tqdm`: False
|
| 377 |
- `remove_unused_columns`: True
|
| 378 |
- `label_names`: None
|
| 379 |
-
- `load_best_model_at_end`:
|
| 380 |
- `ignore_data_skip`: False
|
| 381 |
- `fsdp`: []
|
| 382 |
- `fsdp_min_num_params`: 0
|
|
@@ -386,23 +245,23 @@ You can finetune this model on your own dataset.
|
|
| 386 |
- `parallelism_config`: None
|
| 387 |
- `deepspeed`: None
|
| 388 |
- `label_smoothing_factor`: 0.0
|
| 389 |
-
- `optim`:
|
| 390 |
- `optim_args`: None
|
| 391 |
- `adafactor`: False
|
| 392 |
- `group_by_length`: False
|
| 393 |
- `length_column_name`: length
|
| 394 |
- `project`: huggingface
|
| 395 |
- `trackio_space_id`: trackio
|
| 396 |
-
- `ddp_find_unused_parameters`:
|
| 397 |
- `ddp_bucket_cap_mb`: None
|
| 398 |
- `ddp_broadcast_buffers`: False
|
| 399 |
- `dataloader_pin_memory`: True
|
| 400 |
- `dataloader_persistent_workers`: False
|
| 401 |
- `skip_memory_metrics`: True
|
| 402 |
- `use_legacy_prediction_loop`: False
|
| 403 |
-
- `push_to_hub`:
|
| 404 |
- `resume_from_checkpoint`: None
|
| 405 |
-
- `hub_model_id`:
|
| 406 |
- `hub_strategy`: every_save
|
| 407 |
- `hub_private_repo`: None
|
| 408 |
- `hub_always_push`: False
|
|
@@ -429,167 +288,32 @@ You can finetune this model on your own dataset.
|
|
| 429 |
- `neftune_noise_alpha`: None
|
| 430 |
- `optim_target_modules`: None
|
| 431 |
- `batch_eval_metrics`: False
|
| 432 |
-
- `eval_on_start`:
|
| 433 |
- `use_liger_kernel`: False
|
| 434 |
- `liger_kernel_config`: None
|
| 435 |
- `eval_use_gather_object`: False
|
| 436 |
- `average_tokens_across_devices`: True
|
| 437 |
- `prompts`: None
|
| 438 |
- `batch_sampler`: batch_sampler
|
| 439 |
-
- `multi_dataset_batch_sampler`:
|
| 440 |
- `router_mapping`: {}
|
| 441 |
- `learning_rate_mapping`: {}
|
| 442 |
|
| 443 |
</details>
|
| 444 |
|
| 445 |
### Training Logs
|
| 446 |
-
|
| 447 |
-
|
| 448 |
-
|
|
| 449 |
-
|
| 450 |
-
| 0
|
| 451 |
-
|
|
| 452 |
-
|
|
| 453 |
-
|
|
| 454 |
-
|
|
| 455 |
-
|
|
| 456 |
-
|
|
| 457 |
-
| 0.4979 | 700 | 0.2017 | 0.1443 | 0.8727 |
|
| 458 |
-
| 0.5690 | 800 | 0.1933 | 0.1322 | 0.8746 |
|
| 459 |
-
| 0.6401 | 900 | 0.1818 | 0.1217 | 0.8755 |
|
| 460 |
-
| 0.7112 | 1000 | 0.1714 | 0.1141 | 0.8769 |
|
| 461 |
-
| 0.7824 | 1100 | 0.157 | 0.1060 | 0.8780 |
|
| 462 |
-
| 0.8535 | 1200 | 0.1467 | 0.0998 | 0.8788 |
|
| 463 |
-
| 0.9246 | 1300 | 0.1394 | 0.0937 | 0.8805 |
|
| 464 |
-
| 0.9957 | 1400 | 0.1343 | 0.0910 | 0.8813 |
|
| 465 |
-
| 1.0669 | 1500 | 0.1222 | 0.0853 | 0.8822 |
|
| 466 |
-
| 1.1380 | 1600 | 0.1173 | 0.0820 | 0.8821 |
|
| 467 |
-
| 1.2091 | 1700 | 0.1082 | 0.0797 | 0.8828 |
|
| 468 |
-
| 1.2802 | 1800 | 0.1105 | 0.0777 | 0.8835 |
|
| 469 |
-
| 1.3514 | 1900 | 0.1093 | 0.0734 | 0.8833 |
|
| 470 |
-
| 1.4225 | 2000 | 0.1034 | 0.0744 | 0.8840 |
|
| 471 |
-
| 1.4936 | 2100 | 0.1016 | 0.0713 | 0.8845 |
|
| 472 |
-
| 1.5647 | 2200 | 0.0995 | 0.0699 | 0.8851 |
|
| 473 |
-
| 1.6358 | 2300 | 0.0994 | 0.0679 | 0.8849 |
|
| 474 |
-
| 1.7070 | 2400 | 0.1024 | 0.0667 | 0.8867 |
|
| 475 |
-
| 1.7781 | 2500 | 0.0911 | 0.0658 | 0.8868 |
|
| 476 |
-
| 1.8492 | 2600 | 0.0907 | 0.0640 | 0.8861 |
|
| 477 |
-
| 1.9203 | 2700 | 0.0941 | 0.0632 | 0.8859 |
|
| 478 |
-
| 1.9915 | 2800 | 0.093 | 0.0625 | 0.8870 |
|
| 479 |
-
| 2.0626 | 2900 | 0.0814 | 0.0618 | 0.8875 |
|
| 480 |
-
| 2.1337 | 3000 | 0.0811 | 0.0609 | 0.8868 |
|
| 481 |
-
| 2.2048 | 3100 | 0.0773 | 0.0602 | 0.8880 |
|
| 482 |
-
| 2.2760 | 3200 | 0.0813 | 0.0590 | 0.8873 |
|
| 483 |
-
| 2.3471 | 3300 | 0.0806 | 0.0584 | 0.8876 |
|
| 484 |
-
| 2.4182 | 3400 | 0.0765 | 0.0575 | 0.8882 |
|
| 485 |
-
| 2.4893 | 3500 | 0.0774 | 0.0581 | 0.8889 |
|
| 486 |
-
| 2.5605 | 3600 | 0.0761 | 0.0560 | 0.8883 |
|
| 487 |
-
| 2.6316 | 3700 | 0.0735 | 0.0560 | 0.8886 |
|
| 488 |
-
| 2.7027 | 3800 | 0.0711 | 0.0555 | 0.8891 |
|
| 489 |
-
| 2.7738 | 3900 | 0.0747 | 0.0551 | 0.8889 |
|
| 490 |
-
| 2.8450 | 4000 | 0.0731 | 0.0552 | 0.8897 |
|
| 491 |
-
| 2.9161 | 4100 | 0.0708 | 0.0543 | 0.8898 |
|
| 492 |
-
| 2.9872 | 4200 | 0.0778 | 0.0536 | 0.8901 |
|
| 493 |
-
| 3.0583 | 4300 | 0.0697 | 0.0540 | 0.8893 |
|
| 494 |
-
| 3.1294 | 4400 | 0.0668 | 0.0533 | 0.8900 |
|
| 495 |
-
| 3.2006 | 4500 | 0.0679 | 0.0526 | 0.8893 |
|
| 496 |
-
| 3.2717 | 4600 | 0.0652 | 0.0532 | 0.8902 |
|
| 497 |
-
| 3.3428 | 4700 | 0.0673 | 0.0520 | 0.8899 |
|
| 498 |
-
| 3.4139 | 4800 | 0.0625 | 0.0514 | 0.8903 |
|
| 499 |
-
| 3.4851 | 4900 | 0.0669 | 0.0515 | 0.8912 |
|
| 500 |
-
| 3.5562 | 5000 | 0.0641 | 0.0515 | 0.8915 |
|
| 501 |
-
| 3.6273 | 5100 | 0.0637 | 0.0509 | 0.8909 |
|
| 502 |
-
| 3.6984 | 5200 | 0.0635 | 0.0506 | 0.8908 |
|
| 503 |
-
| 3.7696 | 5300 | 0.0606 | 0.0499 | 0.8915 |
|
| 504 |
-
| 3.8407 | 5400 | 0.0633 | 0.0503 | 0.8917 |
|
| 505 |
-
| 3.9118 | 5500 | 0.0656 | 0.0498 | 0.8913 |
|
| 506 |
-
| 3.9829 | 5600 | 0.0658 | 0.0492 | 0.8916 |
|
| 507 |
-
| 4.0541 | 5700 | 0.0606 | 0.0489 | 0.8917 |
|
| 508 |
-
| 4.1252 | 5800 | 0.0585 | 0.0485 | 0.8914 |
|
| 509 |
-
| 4.1963 | 5900 | 0.0613 | 0.0490 | 0.8914 |
|
| 510 |
-
| 4.2674 | 6000 | 0.0568 | 0.0487 | 0.8909 |
|
| 511 |
-
| 4.3385 | 6100 | 0.0576 | 0.0481 | 0.8918 |
|
| 512 |
-
| 4.4097 | 6200 | 0.0603 | 0.0481 | 0.8915 |
|
| 513 |
-
| 4.4808 | 6300 | 0.0569 | 0.0480 | 0.8918 |
|
| 514 |
-
| 4.5519 | 6400 | 0.0553 | 0.0477 | 0.8921 |
|
| 515 |
-
| 4.6230 | 6500 | 0.057 | 0.0472 | 0.8918 |
|
| 516 |
-
| 4.6942 | 6600 | 0.0602 | 0.0472 | 0.8925 |
|
| 517 |
-
| 4.7653 | 6700 | 0.0541 | 0.0468 | 0.8922 |
|
| 518 |
-
| 4.8364 | 6800 | 0.0588 | 0.0468 | 0.8917 |
|
| 519 |
-
| 4.9075 | 6900 | 0.0588 | 0.0471 | 0.8920 |
|
| 520 |
-
| 4.9787 | 7000 | 0.0549 | 0.0469 | 0.8921 |
|
| 521 |
-
| 5.0498 | 7100 | 0.0522 | 0.0466 | 0.8920 |
|
| 522 |
-
| 5.1209 | 7200 | 0.0527 | 0.0462 | 0.8924 |
|
| 523 |
-
| 5.1920 | 7300 | 0.0519 | 0.0461 | 0.8924 |
|
| 524 |
-
| 5.2632 | 7400 | 0.0544 | 0.0459 | 0.8927 |
|
| 525 |
-
| 5.3343 | 7500 | 0.0549 | 0.0456 | 0.8925 |
|
| 526 |
-
| 5.4054 | 7600 | 0.0527 | 0.0460 | 0.8932 |
|
| 527 |
-
| 5.4765 | 7700 | 0.0519 | 0.0453 | 0.8920 |
|
| 528 |
-
| 5.5477 | 7800 | 0.0528 | 0.0455 | 0.8928 |
|
| 529 |
-
| 5.6188 | 7900 | 0.0525 | 0.0451 | 0.8929 |
|
| 530 |
-
| 5.6899 | 8000 | 0.0535 | 0.0454 | 0.8931 |
|
| 531 |
-
| 5.7610 | 8100 | 0.0526 | 0.0452 | 0.8931 |
|
| 532 |
-
| 5.8321 | 8200 | 0.0507 | 0.0454 | 0.8930 |
|
| 533 |
-
| 5.9033 | 8300 | 0.0511 | 0.0451 | 0.8932 |
|
| 534 |
-
| 5.9744 | 8400 | 0.0489 | 0.0451 | 0.8930 |
|
| 535 |
-
| 6.0455 | 8500 | 0.0509 | 0.0451 | 0.8929 |
|
| 536 |
-
| 6.1166 | 8600 | 0.0487 | 0.0447 | 0.8931 |
|
| 537 |
-
| 6.1878 | 8700 | 0.0494 | 0.0449 | 0.8932 |
|
| 538 |
-
| 6.2589 | 8800 | 0.0474 | 0.0444 | 0.8932 |
|
| 539 |
-
| 6.3300 | 8900 | 0.049 | 0.0448 | 0.8934 |
|
| 540 |
-
| 6.4011 | 9000 | 0.0492 | 0.0446 | 0.8934 |
|
| 541 |
-
| 6.4723 | 9100 | 0.0493 | 0.0443 | 0.8931 |
|
| 542 |
-
| 6.5434 | 9200 | 0.0517 | 0.0442 | 0.8931 |
|
| 543 |
-
| 6.6145 | 9300 | 0.0502 | 0.0445 | 0.8938 |
|
| 544 |
-
| 6.6856 | 9400 | 0.0501 | 0.0441 | 0.8935 |
|
| 545 |
-
| 6.7568 | 9500 | 0.0484 | 0.0439 | 0.8935 |
|
| 546 |
-
| 6.8279 | 9600 | 0.0472 | 0.0437 | 0.8935 |
|
| 547 |
-
| 6.8990 | 9700 | 0.0484 | 0.0435 | 0.8936 |
|
| 548 |
-
| 6.9701 | 9800 | 0.051 | 0.0433 | 0.8933 |
|
| 549 |
-
| 7.0413 | 9900 | 0.0496 | 0.0435 | 0.8935 |
|
| 550 |
-
| 7.1124 | 10000 | 0.0469 | 0.0434 | 0.8937 |
|
| 551 |
-
| 7.1835 | 10100 | 0.0479 | 0.0432 | 0.8935 |
|
| 552 |
-
| 7.2546 | 10200 | 0.0476 | 0.0430 | 0.8937 |
|
| 553 |
-
| 7.3257 | 10300 | 0.0454 | 0.0431 | 0.8934 |
|
| 554 |
-
| 7.3969 | 10400 | 0.0445 | 0.0430 | 0.8937 |
|
| 555 |
-
| 7.4680 | 10500 | 0.0471 | 0.0427 | 0.8936 |
|
| 556 |
-
| 7.5391 | 10600 | 0.0441 | 0.0429 | 0.8938 |
|
| 557 |
-
| 7.6102 | 10700 | 0.046 | 0.0429 | 0.8932 |
|
| 558 |
-
| 7.6814 | 10800 | 0.046 | 0.0428 | 0.8934 |
|
| 559 |
-
| 7.7525 | 10900 | 0.049 | 0.0428 | 0.8938 |
|
| 560 |
-
| 7.8236 | 11000 | 0.0476 | 0.0427 | 0.8939 |
|
| 561 |
-
| 7.8947 | 11100 | 0.0468 | 0.0425 | 0.8938 |
|
| 562 |
-
| 7.9659 | 11200 | 0.0465 | 0.0426 | 0.8940 |
|
| 563 |
-
| 8.0370 | 11300 | 0.048 | 0.0428 | 0.8938 |
|
| 564 |
-
| 8.1081 | 11400 | 0.0448 | 0.0425 | 0.8937 |
|
| 565 |
-
| 8.1792 | 11500 | 0.0431 | 0.0424 | 0.8939 |
|
| 566 |
-
| 8.2504 | 11600 | 0.0428 | 0.0424 | 0.8935 |
|
| 567 |
-
| 8.3215 | 11700 | 0.046 | 0.0424 | 0.8937 |
|
| 568 |
-
| 8.3926 | 11800 | 0.0471 | 0.0423 | 0.8938 |
|
| 569 |
-
| 8.4637 | 11900 | 0.0466 | 0.0424 | 0.8943 |
|
| 570 |
-
| 8.5349 | 12000 | 0.0431 | 0.0421 | 0.8941 |
|
| 571 |
-
| 8.6060 | 12100 | 0.0462 | 0.0421 | 0.8938 |
|
| 572 |
-
| 8.6771 | 12200 | 0.0425 | 0.0423 | 0.8941 |
|
| 573 |
-
| 8.7482 | 12300 | 0.0455 | 0.0421 | 0.8941 |
|
| 574 |
-
| 8.8193 | 12400 | 0.0445 | 0.0422 | 0.8940 |
|
| 575 |
-
| 8.8905 | 12500 | 0.0455 | 0.0422 | 0.8943 |
|
| 576 |
-
| 8.9616 | 12600 | 0.0448 | 0.0421 | 0.8941 |
|
| 577 |
-
| 9.0327 | 12700 | 0.0462 | 0.0421 | 0.8940 |
|
| 578 |
-
| 9.1038 | 12800 | 0.0429 | 0.0421 | 0.8939 |
|
| 579 |
-
| 9.1750 | 12900 | 0.0452 | 0.0421 | 0.8942 |
|
| 580 |
-
| 9.2461 | 13000 | 0.0439 | 0.0420 | 0.8943 |
|
| 581 |
-
| 9.3172 | 13100 | 0.0472 | 0.0420 | 0.8942 |
|
| 582 |
-
| 9.3883 | 13200 | 0.0447 | 0.0420 | 0.8943 |
|
| 583 |
-
| 9.4595 | 13300 | 0.0426 | 0.0420 | 0.8942 |
|
| 584 |
-
| 9.5306 | 13400 | 0.0445 | 0.0420 | 0.8942 |
|
| 585 |
-
| 9.6017 | 13500 | 0.0436 | 0.0419 | 0.8942 |
|
| 586 |
-
| 9.6728 | 13600 | 0.0445 | 0.0419 | 0.8943 |
|
| 587 |
-
| 9.7440 | 13700 | 0.0477 | 0.0419 | 0.8943 |
|
| 588 |
-
| 9.8151 | 13800 | 0.0439 | 0.0419 | 0.8942 |
|
| 589 |
-
| 9.8862 | 13900 | 0.0438 | 0.0419 | 0.8942 |
|
| 590 |
-
| 9.9573 | 14000 | 0.0468 | 0.0419 | 0.8942 |
|
| 591 |
|
| 592 |
-
</details>
|
| 593 |
|
| 594 |
### Framework Versions
|
| 595 |
- Python: 3.10.18
|
|
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
+
- dataset_size:100000
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
+
- source_sentence: How do I calculate IQ?
|
|
|
|
| 13 |
sentences:
|
| 14 |
+
- What is the easiest way to know my IQ?
|
| 15 |
+
- How do I calculate not IQ ?
|
| 16 |
+
- What are some creative and innovative business ideas with less investment in India?
|
| 17 |
+
- source_sentence: How can I learn martial arts in my home?
|
|
|
|
|
|
|
| 18 |
sentences:
|
| 19 |
+
- How can I learn martial arts by myself?
|
| 20 |
+
- What are the advantages and disadvantages of investing in gold?
|
| 21 |
+
- Can people see that I have looked at their pictures on instagram if I am not following
|
| 22 |
+
them?
|
| 23 |
+
- source_sentence: When Enterprise picks you up do you have to take them back?
|
| 24 |
sentences:
|
| 25 |
+
- Are there any software Training institute in Tuticorin?
|
| 26 |
+
- When Enterprise picks you up do you have to take them back?
|
| 27 |
+
- When Enterprise picks you up do them have to take youback?
|
| 28 |
+
- source_sentence: What are some non-capital goods?
|
|
|
|
|
|
|
| 29 |
sentences:
|
| 30 |
+
- What are capital goods?
|
| 31 |
+
- How is the value of [math]\pi[/math] calculated?
|
| 32 |
+
- What are some non-capital goods?
|
| 33 |
+
- source_sentence: What is the QuickBooks technical support phone number in New York?
|
|
|
|
|
|
|
|
|
|
| 34 |
sentences:
|
| 35 |
+
- What caused the Great Depression?
|
| 36 |
+
- Can I apply for PR in Canada?
|
| 37 |
+
- Which is the best QuickBooks Hosting Support Number in New York?
|
|
|
|
| 38 |
pipeline_tag: sentence-similarity
|
| 39 |
library_name: sentence-transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
---
|
| 41 |
|
| 42 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
|
|
| 85 |
from sentence_transformers import SentenceTransformer
|
| 86 |
|
| 87 |
# Download from the 🤗 Hub
|
| 88 |
+
model = SentenceTransformer("sentence_transformers_model_id")
|
| 89 |
# Run inference
|
| 90 |
sentences = [
|
| 91 |
+
'What is the QuickBooks technical support phone number in New York?',
|
| 92 |
+
'Which is the best QuickBooks Hosting Support Number in New York?',
|
| 93 |
+
'Can I apply for PR in Canada?',
|
| 94 |
]
|
| 95 |
embeddings = model.encode(sentences)
|
| 96 |
print(embeddings.shape)
|
|
|
|
| 99 |
# Get the similarity scores for the embeddings
|
| 100 |
similarities = model.similarity(embeddings, embeddings)
|
| 101 |
print(similarities)
|
| 102 |
+
# tensor([[1.0000, 0.8563, 0.0594],
|
| 103 |
+
# [0.8563, 1.0000, 0.1245],
|
| 104 |
+
# [0.0594, 0.1245, 1.0000]])
|
| 105 |
```
|
| 106 |
|
| 107 |
<!--
|
|
|
|
| 128 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 129 |
-->
|
| 130 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
<!--
|
| 132 |
## Bias, Risks and Limitations
|
| 133 |
|
|
|
|
| 146 |
|
| 147 |
#### Unnamed Dataset
|
| 148 |
|
| 149 |
+
* Size: 100,000 training samples
|
| 150 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 151 |
* Approximate statistics based on the first 1000 samples:
|
| 152 |
+
| | sentence_0 | sentence_1 | sentence_2 |
|
| 153 |
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
| 154 |
| type | string | string | string |
|
| 155 |
+
| details | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.68 tokens</li><li>max: 66 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 16.37 tokens</li><li>max: 67 tokens</li></ul> |
|
| 156 |
* Samples:
|
| 157 |
+
| sentence_0 | sentence_1 | sentence_2 |
|
| 158 |
+
|:-----------------------------------------------------------------|:-----------------------------------------------------------------|:----------------------------------------------------------------------------------|
|
| 159 |
+
| <code>Is masturbating bad for boys?</code> | <code>Is masturbating bad for boys?</code> | <code>How harmful or unhealthy is masturbation?</code> |
|
| 160 |
+
| <code>Does a train engine move in reverse?</code> | <code>Does a train engine move in reverse?</code> | <code>Time moves forward, not in reverse. Doesn't that make time a vector?</code> |
|
| 161 |
+
| <code>What is the most badass thing anyone has ever done?</code> | <code>What is the most badass thing anyone has ever done?</code> | <code>anyone is the most badass thing Whathas ever done?</code> |
|
| 162 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 163 |
```json
|
| 164 |
{
|
|
|
|
| 171 |
### Training Hyperparameters
|
| 172 |
#### Non-Default Hyperparameters
|
| 173 |
|
| 174 |
+
- `per_device_train_batch_size`: 64
|
| 175 |
+
- `per_device_eval_batch_size`: 64
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 176 |
- `fp16`: True
|
| 177 |
+
- `multi_dataset_batch_sampler`: round_robin
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 178 |
|
| 179 |
#### All Hyperparameters
|
| 180 |
<details><summary>Click to expand</summary>
|
| 181 |
|
| 182 |
- `overwrite_output_dir`: False
|
| 183 |
- `do_predict`: False
|
| 184 |
+
- `eval_strategy`: no
|
| 185 |
- `prediction_loss_only`: True
|
| 186 |
+
- `per_device_train_batch_size`: 64
|
| 187 |
+
- `per_device_eval_batch_size`: 64
|
| 188 |
- `per_gpu_train_batch_size`: None
|
| 189 |
- `per_gpu_eval_batch_size`: None
|
| 190 |
- `gradient_accumulation_steps`: 1
|
| 191 |
- `eval_accumulation_steps`: None
|
| 192 |
- `torch_empty_cache_steps`: None
|
| 193 |
+
- `learning_rate`: 5e-05
|
| 194 |
+
- `weight_decay`: 0.0
|
| 195 |
- `adam_beta1`: 0.9
|
| 196 |
- `adam_beta2`: 0.999
|
| 197 |
- `adam_epsilon`: 1e-08
|
| 198 |
+
- `max_grad_norm`: 1
|
| 199 |
+
- `num_train_epochs`: 3
|
| 200 |
+
- `max_steps`: -1
|
| 201 |
- `lr_scheduler_type`: linear
|
| 202 |
- `lr_scheduler_kwargs`: {}
|
| 203 |
+
- `warmup_ratio`: 0.0
|
| 204 |
- `warmup_steps`: 0
|
| 205 |
- `log_level`: passive
|
| 206 |
- `log_level_replica`: warning
|
|
|
|
| 228 |
- `tpu_num_cores`: None
|
| 229 |
- `tpu_metrics_debug`: False
|
| 230 |
- `debug`: []
|
| 231 |
+
- `dataloader_drop_last`: False
|
| 232 |
+
- `dataloader_num_workers`: 0
|
| 233 |
+
- `dataloader_prefetch_factor`: None
|
| 234 |
- `past_index`: -1
|
| 235 |
- `disable_tqdm`: False
|
| 236 |
- `remove_unused_columns`: True
|
| 237 |
- `label_names`: None
|
| 238 |
+
- `load_best_model_at_end`: False
|
| 239 |
- `ignore_data_skip`: False
|
| 240 |
- `fsdp`: []
|
| 241 |
- `fsdp_min_num_params`: 0
|
|
|
|
| 245 |
- `parallelism_config`: None
|
| 246 |
- `deepspeed`: None
|
| 247 |
- `label_smoothing_factor`: 0.0
|
| 248 |
+
- `optim`: adamw_torch_fused
|
| 249 |
- `optim_args`: None
|
| 250 |
- `adafactor`: False
|
| 251 |
- `group_by_length`: False
|
| 252 |
- `length_column_name`: length
|
| 253 |
- `project`: huggingface
|
| 254 |
- `trackio_space_id`: trackio
|
| 255 |
+
- `ddp_find_unused_parameters`: None
|
| 256 |
- `ddp_bucket_cap_mb`: None
|
| 257 |
- `ddp_broadcast_buffers`: False
|
| 258 |
- `dataloader_pin_memory`: True
|
| 259 |
- `dataloader_persistent_workers`: False
|
| 260 |
- `skip_memory_metrics`: True
|
| 261 |
- `use_legacy_prediction_loop`: False
|
| 262 |
+
- `push_to_hub`: False
|
| 263 |
- `resume_from_checkpoint`: None
|
| 264 |
+
- `hub_model_id`: None
|
| 265 |
- `hub_strategy`: every_save
|
| 266 |
- `hub_private_repo`: None
|
| 267 |
- `hub_always_push`: False
|
|
|
|
| 288 |
- `neftune_noise_alpha`: None
|
| 289 |
- `optim_target_modules`: None
|
| 290 |
- `batch_eval_metrics`: False
|
| 291 |
+
- `eval_on_start`: False
|
| 292 |
- `use_liger_kernel`: False
|
| 293 |
- `liger_kernel_config`: None
|
| 294 |
- `eval_use_gather_object`: False
|
| 295 |
- `average_tokens_across_devices`: True
|
| 296 |
- `prompts`: None
|
| 297 |
- `batch_sampler`: batch_sampler
|
| 298 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 299 |
- `router_mapping`: {}
|
| 300 |
- `learning_rate_mapping`: {}
|
| 301 |
|
| 302 |
</details>
|
| 303 |
|
| 304 |
### Training Logs
|
| 305 |
+
| Epoch | Step | Training Loss |
|
| 306 |
+
|:------:|:----:|:-------------:|
|
| 307 |
+
| 0.3199 | 500 | 0.4294 |
|
| 308 |
+
| 0.6398 | 1000 | 0.1268 |
|
| 309 |
+
| 0.9597 | 1500 | 0.1 |
|
| 310 |
+
| 1.2796 | 2000 | 0.0792 |
|
| 311 |
+
| 1.5995 | 2500 | 0.0706 |
|
| 312 |
+
| 1.9194 | 3000 | 0.0687 |
|
| 313 |
+
| 2.2393 | 3500 | 0.0584 |
|
| 314 |
+
| 2.5592 | 4000 | 0.057 |
|
| 315 |
+
| 2.8791 | 4500 | 0.0581 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 316 |
|
|
|
|
| 317 |
|
| 318 |
### Framework Versions
|
| 319 |
- Python: 3.10.18
|
eval/Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -179,3 +179,144 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
|
|
| 179 |
9.815078236130867,13800,0.827975,0.90265,0.931,0.827975,0.827975,0.3008833333333333,0.90265,0.18620000000000003,0.931,0.827975,0.8676762499999963,0.8720721230158671,0.894178246968981,0.8741864677767469
|
| 180 |
9.88620199146515,13900,0.828025,0.9026,0.9311,0.828025,0.828025,0.3008666666666666,0.9026,0.18622000000000002,0.9311,0.828025,0.867727916666663,0.8721113690476133,0.8942189714318465,0.8742215047241502
|
| 181 |
9.95732574679943,14000,0.828025,0.9027,0.931025,0.828025,0.828025,0.3008999999999999,0.9027,0.186205,0.931025,0.828025,0.8677179166666629,0.8721162896825339,0.8942284691055087,0.8742240723304836
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 179 |
9.815078236130867,13800,0.827975,0.90265,0.931,0.827975,0.827975,0.3008833333333333,0.90265,0.18620000000000003,0.931,0.827975,0.8676762499999963,0.8720721230158671,0.894178246968981,0.8741864677767469
|
| 180 |
9.88620199146515,13900,0.828025,0.9026,0.9311,0.828025,0.828025,0.3008666666666666,0.9026,0.18622000000000002,0.9311,0.828025,0.867727916666663,0.8721113690476133,0.8942189714318465,0.8742215047241502
|
| 181 |
9.95732574679943,14000,0.828025,0.9027,0.931025,0.828025,0.828025,0.3008999999999999,0.9027,0.186205,0.931025,0.828025,0.8677179166666629,0.8721162896825339,0.8942284691055087,0.8742240723304836
|
| 182 |
+
0,0,0.660625,0.78855,0.816875,0.660625,0.660625,0.26285,0.78855,0.163375,0.816875,0.660625,0.7265508333333278,0.7310880059523766,0.7604347668434622,0.7344176635988929
|
| 183 |
+
0.07112375533428165,100,0.65785,0.8221,0.855375,0.65785,0.65785,0.2740333333333333,0.8221,0.171075,0.855375,0.65785,0.7426349999999933,0.747496111111108,0.7830189713908015,0.7509685938326416
|
| 184 |
+
0.1422475106685633,200,0.76665,0.86435,0.8921,0.76665,0.76665,0.28811666666666663,0.86435,0.17842,0.8921,0.76665,0.8175337499999934,0.8216855654761861,0.8463620941294383,0.8244802043057975
|
| 185 |
+
0.21337126600284495,300,0.790825,0.87115,0.89655,0.790825,0.790825,0.2903833333333333,0.87115,0.17931,0.89655,0.790825,0.832905833333328,0.8370798809523765,0.8589576487020009,0.8396630602475953
|
| 186 |
+
0.2844950213371266,400,0.7981,0.8745,0.899575,0.7981,0.7981,0.2915,0.8745,0.17991500000000002,0.899575,0.7981,0.838206249999995,0.8422020833333296,0.8632435713976536,0.8448211843344123
|
| 187 |
+
0.35561877667140823,500,0.802025,0.878225,0.902325,0.802025,0.802025,0.2927416666666666,0.878225,0.18046500000000004,0.902325,0.802025,0.8418195833333277,0.845981617063489,0.8670208129259122,0.8485604470631575
|
| 188 |
+
0.4267425320056899,600,0.805275,0.880225,0.90495,0.805275,0.805275,0.2934083333333333,0.880225,0.18099000000000004,0.90495,0.805275,0.8446729166666621,0.848717420634917,0.8695488582496811,0.8513143530847598
|
| 189 |
+
0.49786628733997157,700,0.807725,0.88225,0.906575,0.807725,0.807725,0.2940833333333333,0.88225,0.181315,0.906575,0.807725,0.84687833333333,0.8510880654761879,0.8720327027175575,0.8536533089445467
|
| 190 |
+
0.5689900426742532,800,0.809625,0.881675,0.9072,0.809625,0.809625,0.29389166666666666,0.881675,0.18144000000000005,0.9072,0.809625,0.8480374999999966,0.8522258333333291,0.872984794383946,0.854857136239104
|
| 191 |
+
0.6401137980085349,900,0.810425,0.883875,0.908725,0.810425,0.810425,0.29462499999999997,0.883875,0.18174500000000002,0.908725,0.810425,0.8493383333333296,0.8535709623015844,0.8744516835300677,0.8561838754847994
|
| 192 |
+
0.7112375533428165,1000,0.812775,0.885475,0.910425,0.812775,0.812775,0.2951583333333333,0.885475,0.18208500000000002,0.910425,0.812775,0.8513895833333308,0.8556871527777754,0.8765193710353214,0.858272041629315
|
| 193 |
+
0.7823613086770982,1100,0.813225,0.8863,0.9114,0.813225,0.813225,0.2954333333333333,0.8863,0.18228000000000003,0.9114,0.813225,0.8519433333333298,0.8562210515872983,0.8771499843059728,0.8588020045901827
|
| 194 |
+
0.8534850640113798,1200,0.81475,0.88835,0.91365,0.81475,0.81475,0.2961166666666666,0.88835,0.18273000000000003,0.91365,0.81475,0.85359958333333,0.8578052579365052,0.8787946145827298,0.8603464162431504
|
| 195 |
+
0.9246088193456614,1300,0.815125,0.88775,0.913825,0.815125,0.815125,0.29591666666666666,0.88775,0.182765,0.913825,0.815125,0.853832916666663,0.8579933333333299,0.8788853951072472,0.8605746130331765
|
| 196 |
+
0.9957325746799431,1400,0.8174,0.88975,0.916,0.8174,0.8174,0.2965833333333333,0.88975,0.18320000000000003,0.916,0.8174,0.8559345833333298,0.8600880952380925,0.8810021639106901,0.8625984545198154
|
| 197 |
+
1.0668563300142249,1500,0.816325,0.8888,0.914975,0.816325,0.816325,0.2962666666666666,0.8888,0.18299500000000002,0.914975,0.816325,0.854826249999997,0.8589861408730142,0.8799266262551008,0.8615620862321841
|
| 198 |
+
1.1379800853485065,1600,0.818025,0.890975,0.91655,0.818025,0.818025,0.2969916666666667,0.890975,0.18331,0.91655,0.818025,0.8567195833333306,0.8609042559523788,0.8818126370189291,0.8634202065126004
|
| 199 |
+
1.209103840682788,1700,0.8187,0.8904,0.916775,0.8187,0.8187,0.2968,0.8904,0.18335500000000002,0.916775,0.8187,0.8569770833333301,0.861196865079362,0.8821120071786194,0.863703746171397
|
| 200 |
+
1.2802275960170697,1800,0.819325,0.890925,0.91765,0.819325,0.819325,0.296975,0.890925,0.18353000000000005,0.91765,0.819325,0.8575979166666624,0.8617786904761866,0.8827203211699732,0.8642523489926849
|
| 201 |
+
1.3513513513513513,1900,0.819575,0.89235,0.918925,0.819575,0.819575,0.29744999999999994,0.89235,0.183785,0.918925,0.819575,0.8583795833333298,0.8625744345238064,0.8836298047121208,0.8650300188501633
|
| 202 |
+
1.422475106685633,2000,0.82045,0.892125,0.91885,0.82045,0.82045,0.297375,0.892125,0.18377000000000002,0.91885,0.82045,0.8587412499999957,0.8629089980158691,0.8838312959674957,0.8653673262499637
|
| 203 |
+
1.4935988620199145,2100,0.8213,0.891925,0.91905,0.8213,0.8213,0.29730833333333323,0.891925,0.18381000000000003,0.91905,0.8213,0.8593224999999954,0.8635408134920594,0.8844500604469038,0.8659837345007254
|
| 204 |
+
1.5647226173541964,2200,0.82035,0.89215,0.919425,0.82035,0.82035,0.2973833333333333,0.89215,0.18388500000000002,0.919425,0.82035,0.8589304166666636,0.8631554464285682,0.884277860760647,0.8656085699448136
|
| 205 |
+
1.635846372688478,2300,0.820625,0.89235,0.91915,0.820625,0.820625,0.29744999999999994,0.89235,0.18383000000000002,0.91915,0.820625,0.8590399999999964,0.8633189781745999,0.8844012972181756,0.8657824496152007
|
| 206 |
+
1.7069701280227596,2400,0.8213,0.893175,0.920525,0.8213,0.8213,0.2977249999999999,0.893175,0.18410500000000005,0.920525,0.8213,0.8596737499999956,0.8639137103174555,0.8851440135349451,0.8663259037610367
|
| 207 |
+
1.7780938833570412,2500,0.82315,0.896175,0.92235,0.82315,0.82315,0.29872499999999996,0.896175,0.18447000000000002,0.92235,0.82315,0.8618333333333296,0.866077976190474,0.8872320902095131,0.8684474128011467
|
| 208 |
+
1.8492176386913228,2600,0.82145,0.89405,0.920575,0.82145,0.82145,0.2980166666666666,0.89405,0.18411500000000006,0.920575,0.82145,0.8601562499999962,0.8644306547619017,0.885607192551569,0.8668801137114076
|
| 209 |
+
1.9203413940256047,2700,0.823,0.894925,0.921975,0.823,0.823,0.2983083333333333,0.894925,0.1843950000000001,0.921975,0.823,0.8615154166666628,0.865800605158727,0.8869919207490155,0.868200699005933
|
| 210 |
+
1.991465149359886,2800,0.822525,0.8952,0.921975,0.822525,0.822525,0.2984,0.8952,0.18439500000000003,0.921975,0.822525,0.8612554166666628,0.8655788888888847,0.8868942891876812,0.8679661063701757
|
| 211 |
+
2.062588904694168,2900,0.82265,0.8948,0.921475,0.82265,0.82265,0.2982666666666666,0.8948,0.18429500000000001,0.921475,0.82265,0.8611083333333301,0.8654150793650768,0.8866298372806077,0.8678082206661994
|
| 212 |
+
2.1337126600284497,3000,0.8235,0.895625,0.92215,0.8235,0.8235,0.2985416666666666,0.895625,0.18442999999999998,0.92215,0.8235,0.8619112499999959,0.8662926984126956,0.8875291073406817,0.868677931078467
|
| 213 |
+
2.204836415362731,3100,0.822925,0.89535,0.9217,0.822925,0.822925,0.29844999999999994,0.89535,0.18434000000000003,0.9217,0.822925,0.8613891666666638,0.8658161507936486,0.887148424379975,0.8681836052644616
|
| 214 |
+
2.275960170697013,3200,0.824975,0.89685,0.922975,0.824975,0.824975,0.29894999999999994,0.89685,0.18459500000000004,0.922975,0.824975,0.8630554166666626,0.8674809722222193,0.8887076656677213,0.8697873502580414
|
| 215 |
+
2.3470839260312943,3300,0.824325,0.89665,0.922425,0.824325,0.824325,0.2988833333333333,0.89665,0.18448500000000004,0.922425,0.824325,0.862727916666663,0.8672325099206315,0.8884958531410477,0.8695684400453714
|
| 216 |
+
2.418207681365576,3400,0.824275,0.896175,0.922325,0.824275,0.824275,0.2987249999999999,0.896175,0.18446500000000005,0.922325,0.824275,0.8625166666666632,0.8669797718253927,0.8882555705244735,0.8693215056646579
|
| 217 |
+
2.4893314366998576,3500,0.824075,0.8962,0.922325,0.824075,0.824075,0.2987333333333333,0.8962,0.18446500000000005,0.922325,0.824075,0.8624966666666627,0.8669934424603137,0.8883243690177098,0.8693086070714775
|
| 218 |
+
2.5604551920341394,3600,0.8248,0.896775,0.923875,0.8248,0.8248,0.2989249999999999,0.896775,0.18477500000000002,0.923875,0.8248,0.8632529166666633,0.8675360714285681,0.8887711826852934,0.8698778746887647
|
| 219 |
+
2.6315789473684212,3700,0.825075,0.897175,0.923625,0.825075,0.825075,0.2990583333333333,0.897175,0.18472500000000003,0.923625,0.825075,0.8634116666666629,0.8677288293650759,0.8889119603422132,0.8701022749841466
|
| 220 |
+
2.7027027027027026,3800,0.824525,0.89705,0.923225,0.824525,0.824525,0.2990166666666666,0.89705,0.18464500000000003,0.923225,0.824525,0.8629341666666629,0.867378343253966,0.8887312117027028,0.8697003139888475
|
| 221 |
+
2.7738264580369845,3900,0.825025,0.897425,0.924375,0.825025,0.825025,0.2991416666666666,0.897425,0.184875,0.924375,0.825025,0.8637245833333296,0.8680381746031713,0.8893130999822193,0.8703772079610197
|
| 222 |
+
2.844950213371266,4000,0.823725,0.89655,0.92275,0.823725,0.823725,0.29885,0.89655,0.18455000000000002,0.92275,0.823725,0.8624545833333299,0.8669824801587267,0.8884952313868706,0.8693080141848417
|
| 223 |
+
2.9160739687055477,4100,0.8248,0.89735,0.92425,0.8248,0.8248,0.2991166666666666,0.89735,0.18485000000000001,0.92425,0.8248,0.8635024999999963,0.8679394246031711,0.8894314662912053,0.8702279980402775
|
| 224 |
+
2.987197724039829,4200,0.8259,0.897925,0.924975,0.8259,0.8259,0.2993083333333333,0.897925,0.18499500000000002,0.924975,0.8259,0.8644899999999974,0.8688425595238083,0.8901161301402605,0.8711507401738996
|
| 225 |
+
3.058321479374111,4300,0.824725,0.8973,0.92515,0.824725,0.824725,0.2991,0.8973,0.18503000000000003,0.92515,0.824725,0.8636933333333314,0.868004523809522,0.8894692161504967,0.8703365016119559
|
| 226 |
+
3.1294452347083928,4400,0.8249,0.897525,0.924325,0.8249,0.8249,0.2991749999999999,0.897525,0.184865,0.924325,0.8249,0.8636020833333312,0.868028988095237,0.8894864613534786,0.8703489640976624
|
| 227 |
+
3.200568990042674,4500,0.8258,0.8986,0.92565,0.8258,0.8258,0.2995333333333333,0.8986,0.18513000000000004,0.92565,0.8258,0.8646745833333307,0.8689872420634889,0.890385534258935,0.8712701053994162
|
| 228 |
+
3.271692745376956,4600,0.8257,0.8983,0.92535,0.8257,0.8257,0.29943333333333333,0.8983,0.18507000000000004,0.92535,0.8257,0.8644454166666634,0.8688889682539637,0.8904275201569494,0.8711486096369192
|
| 229 |
+
3.3428165007112374,4700,0.826325,0.89855,0.92535,0.826325,0.826325,0.29951666666666665,0.89855,0.18507000000000004,0.92535,0.826325,0.8648566666666646,0.8692932539682527,0.8906216945646305,0.8716049278778036
|
| 230 |
+
3.413940256045519,4800,0.8256,0.898475,0.92485,0.8256,0.8256,0.2994916666666666,0.898475,0.18497000000000002,0.92485,0.8256,0.8642804166666638,0.8687215277777753,0.890143837174269,0.8710299777162378
|
| 231 |
+
3.485064011379801,4900,0.8265,0.89945,0.9264,0.8265,0.8265,0.2998166666666666,0.89945,0.18528000000000003,0.9264,0.8265,0.8654299999999965,0.8697809325396798,0.8911726290639512,0.8720622159211568
|
| 232 |
+
3.5561877667140824,5000,0.8265,0.89915,0.926275,0.8265,0.8265,0.29971666666666663,0.89915,0.18525500000000003,0.926275,0.8265,0.8653016666666635,0.8696054067460289,0.8909265121877501,0.8719266555079046
|
| 233 |
+
3.6273115220483643,5100,0.82565,0.899175,0.9258,0.82565,0.82565,0.29972499999999996,0.899175,0.18516000000000005,0.9258,0.82565,0.8647320833333304,0.8691491269841237,0.8906552891143423,0.8714300727214882
|
| 234 |
+
3.6984352773826457,5200,0.8256,0.8991,0.9263,0.8256,0.8256,0.2996999999999999,0.8991,0.18526000000000004,0.9263,0.8256,0.8647333333333299,0.8690647321428537,0.8906112568228791,0.8713395023759158
|
| 235 |
+
3.7695590327169275,5300,0.8267,0.90005,0.927175,0.8267,0.8267,0.30001666666666665,0.90005,0.18543500000000004,0.927175,0.8267,0.8658445833333303,0.870133650793649,0.8915589357540593,0.8724021579521188
|
| 236 |
+
3.8406827880512093,5400,0.82635,0.89965,0.92665,0.82635,0.82635,0.2998833333333333,0.89965,0.18533000000000002,0.92665,0.82635,0.8654087499999968,0.8697750694444412,0.8913026933116265,0.8720366312771636
|
| 237 |
+
3.9118065433854907,5500,0.826675,0.900275,0.927025,0.826675,0.826675,0.30009166666666665,0.900275,0.18540500000000001,0.927025,0.826675,0.8657945833333306,0.8701155952380921,0.891577510224056,0.8723879285250398
|
| 238 |
+
3.9829302987197726,5600,0.82665,0.89945,0.9268,0.82665,0.82665,0.2998166666666666,0.89945,0.18536000000000002,0.9268,0.82665,0.8655083333333307,0.8698893154761871,0.8914393474383101,0.8721425772539715
|
| 239 |
+
4.054054054054054,5700,0.826,0.898975,0.926125,0.826,0.826,0.29965833333333325,0.898975,0.185225,0.926125,0.826,0.8650079166666638,0.8693917162698382,0.8908775275994638,0.8717032813673281
|
| 240 |
+
4.125177809388336,5800,0.82755,0.900225,0.92755,0.82755,0.82755,0.300075,0.900225,0.18551000000000004,0.92755,0.82755,0.8664683333333296,0.8708438392857104,0.8923375459234719,0.8730607380382754
|
| 241 |
+
4.196301564722617,5900,0.82705,0.90065,0.9282,0.82705,0.82705,0.30021666666666663,0.90065,0.18564,0.9282,0.82705,0.8664145833333298,0.8706986011904715,0.8922585677258182,0.8729291403486316
|
| 242 |
+
4.2674253200568995,6000,0.827125,0.899675,0.92705,0.827125,0.827125,0.2998916666666666,0.899675,0.18541000000000005,0.92705,0.827125,0.8659537499999962,0.8703908829365028,0.89196813784815,0.8726083427829648
|
| 243 |
+
4.338549075391181,6100,0.82685,0.90015,0.927425,0.82685,0.82685,0.30004999999999993,0.90015,0.185485,0.927425,0.82685,0.8659616666666632,0.8703435515872973,0.891964842926896,0.8725465070798555
|
| 244 |
+
4.409672830725462,6200,0.826875,0.900275,0.927075,0.826875,0.826875,0.3000916666666666,0.900275,0.185415,0.927075,0.826875,0.8658904166666626,0.8703097519841209,0.8918968662133886,0.8725315660787611
|
| 245 |
+
4.480796586059744,6300,0.82725,0.9012,0.927925,0.82725,0.82725,0.30039999999999994,0.9012,0.185585,0.927925,0.82725,0.8665112499999957,0.8708849999999949,0.8924496641197642,0.8731048780889112
|
| 246 |
+
4.551920341394026,6400,0.82685,0.900775,0.927275,0.82685,0.82685,0.3002583333333333,0.900775,0.185455,0.927275,0.82685,0.8660962499999955,0.8705262896825335,0.8921092548063507,0.8727407525006975
|
| 247 |
+
4.623044096728307,6500,0.827575,0.901075,0.928275,0.827575,0.827575,0.3003583333333333,0.901075,0.18565500000000001,0.928275,0.827575,0.8668083333333294,0.8711607341269799,0.8926932674759781,0.8733821469708742
|
| 248 |
+
4.694167852062589,6600,0.826725,0.90005,0.927225,0.826725,0.826725,0.3000166666666666,0.90005,0.18544500000000003,0.927225,0.826725,0.8657887499999959,0.8702033432539638,0.891834835753562,0.8724344252761722
|
| 249 |
+
4.76529160739687,6700,0.827225,0.900125,0.927525,0.827225,0.827225,0.3000416666666666,0.900125,0.18550500000000003,0.927525,0.827225,0.8662174999999963,0.8706892460317409,0.8923377513084956,0.8729005329969197
|
| 250 |
+
4.836415362731152,6800,0.827675,0.90055,0.9274,0.827675,0.827675,0.30018333333333325,0.90055,0.18548000000000003,0.9274,0.827675,0.8665483333333288,0.8710253075396777,0.8925563711068755,0.8732563983996522
|
| 251 |
+
4.907539118065434,6900,0.827575,0.9004,0.92825,0.827575,0.827575,0.30013333333333325,0.9004,0.18565,0.92825,0.827575,0.8666908333333294,0.871095198412693,0.8927540200153565,0.873288654653562
|
| 252 |
+
4.978662873399715,7000,0.82795,0.90095,0.9285,0.82795,0.82795,0.3003166666666666,0.90095,0.1857,0.9285,0.82795,0.8669474999999964,0.8713514285714237,0.8929598750593017,0.8735464299621104
|
| 253 |
+
5.049786628733997,7100,0.82815,0.90145,0.928925,0.82815,0.82815,0.30048333333333327,0.90145,0.18578500000000003,0.928925,0.82815,0.8673516666666624,0.8717274305555508,0.893290068636731,0.8739296073428533
|
| 254 |
+
5.120910384068279,7200,0.82795,0.90095,0.9286,0.82795,0.82795,0.3003166666666666,0.90095,0.18572000000000002,0.9286,0.82795,0.8670533333333291,0.8714805257936449,0.8931097492872178,0.8736657314635108
|
| 255 |
+
5.19203413940256,7300,0.82755,0.9014,0.92895,0.82755,0.82755,0.3004666666666666,0.9014,0.18579,0.92895,0.82755,0.8670079166666628,0.8714806150793606,0.8932898007413307,0.8736056338472569
|
| 256 |
+
5.2631578947368425,7400,0.827525,0.9012,0.9288,0.827525,0.827525,0.30039999999999994,0.9012,0.18576,0.9288,0.827525,0.8667895833333292,0.8712067261904715,0.8929619217995062,0.8733752014385081
|
| 257 |
+
5.334281650071124,7500,0.8274,0.900825,0.928425,0.8274,0.8274,0.300275,0.900825,0.18568500000000002,0.928425,0.8274,0.8666358333333299,0.871154603174598,0.8930023691933456,0.8732816124281766
|
| 258 |
+
5.405405405405405,7600,0.8278,0.90085,0.92875,0.8278,0.8278,0.3002833333333333,0.90085,0.18575000000000003,0.92875,0.8278,0.8669041666666626,0.8714118353174557,0.893248046324686,0.8735481954345135
|
| 259 |
+
5.476529160739687,7700,0.8275,0.9006,0.929025,0.8275,0.8275,0.3001999999999999,0.9006,0.18580500000000003,0.929025,0.8275,0.8667345833333302,0.8711018650793607,0.8928943868112501,0.873268752899655
|
| 260 |
+
5.547652916073969,7800,0.8282,0.90175,0.929125,0.8282,0.8282,0.3005833333333333,0.90175,0.185825,0.929125,0.8282,0.8674249999999957,0.8717897222222175,0.8934076888371761,0.8739759838230828
|
| 261 |
+
5.61877667140825,7900,0.828125,0.90145,0.9286,0.828125,0.828125,0.3004833333333333,0.90145,0.18572000000000002,0.9286,0.828125,0.8671795833333292,0.8716491468253924,0.8933186665620889,0.8738263228050317
|
| 262 |
+
5.689900426742532,8000,0.827775,0.901,0.928925,0.827775,0.827775,0.3003333333333333,0.901,0.18578500000000003,0.928925,0.827775,0.8669049999999959,0.8712585416666617,0.8929668333310461,0.8734331161309917
|
| 263 |
+
5.761024182076814,8100,0.828325,0.901925,0.92915,0.828325,0.828325,0.3006416666666666,0.901925,0.18583000000000005,0.92915,0.828325,0.8674870833333292,0.8719067162698362,0.8935731518881023,0.874073460724267
|
| 264 |
+
5.832147937411095,8200,0.82765,0.901475,0.9292,0.82765,0.82765,0.3004916666666666,0.901475,0.18584000000000006,0.9292,0.82765,0.867018749999996,0.8714473412698366,0.8932306586102523,0.8736176102556379
|
| 265 |
+
5.903271692745377,8300,0.8282,0.9018,0.9294,0.8282,0.8282,0.3006,0.9018,0.18588000000000002,0.9294,0.8282,0.8674824999999957,0.8719401686507885,0.8937213383114889,0.8740581043681251
|
| 266 |
+
5.974395448079658,8400,0.828325,0.90185,0.929925,0.828325,0.828325,0.30061666666666664,0.90185,0.185985,0.929925,0.828325,0.8677295833333308,0.8720828869047582,0.8937899946723228,0.8742234355394175
|
| 267 |
+
6.0455192034139404,8500,0.8286,0.9023,0.929925,0.8286,0.8286,0.3007666666666666,0.9023,0.18598500000000004,0.929925,0.8286,0.8680758333333293,0.8724603968253921,0.894128127198057,0.8745885827849681
|
| 268 |
+
6.116642958748222,8600,0.828175,0.90225,0.929325,0.828175,0.828175,0.30074999999999996,0.90225,0.185865,0.929325,0.828175,0.8675804166666634,0.8720532440476145,0.8938181845798283,0.8741853105958053
|
| 269 |
+
6.187766714082503,8700,0.82805,0.90215,0.929725,0.82805,0.82805,0.30071666666666663,0.90215,0.18594500000000003,0.929725,0.82805,0.8674470833333299,0.8718606547619014,0.8936594724961032,0.8739825128657174
|
| 270 |
+
6.2588904694167855,8800,0.82785,0.90205,0.9297,0.82785,0.82785,0.3006833333333333,0.90205,0.18594000000000002,0.9297,0.82785,0.8674191666666624,0.8718350892857094,0.893655148048901,0.873962454047865
|
| 271 |
+
6.330014224751067,8900,0.828525,0.90235,0.930125,0.828525,0.828525,0.3007833333333333,0.90235,0.18602500000000002,0.930125,0.828525,0.8680445833333297,0.8724303174603135,0.8941609142278779,0.8745462546782486
|
| 272 |
+
6.401137980085348,9000,0.828325,0.902,0.930025,0.828325,0.828325,0.3006666666666666,0.902,0.18600500000000003,0.930025,0.828325,0.8678520833333294,0.8722724900793611,0.8940775055650032,0.8743679763133748
|
| 273 |
+
6.472261735419631,9100,0.828475,0.902075,0.92965,0.828475,0.828475,0.3006916666666666,0.902075,0.18593,0.92965,0.828475,0.8678004166666627,0.8722169146825343,0.8939301588915514,0.8743476154747785
|
| 274 |
+
6.543385490753912,9200,0.8286,0.901525,0.929325,0.8286,0.8286,0.3005083333333333,0.901525,0.18586500000000003,0.929325,0.8286,0.867682083333329,0.8721559920634869,0.8938887495928067,0.8742804116459679
|
| 275 |
+
6.614509246088193,9300,0.82855,0.902625,0.930175,0.82855,0.82855,0.300875,0.902625,0.18603500000000003,0.930175,0.82855,0.8681324999999959,0.8725687400793595,0.8943725725633536,0.8746570790204005
|
| 276 |
+
6.685633001422475,9400,0.8288,0.90245,0.93045,0.8288,0.8288,0.3008166666666666,0.90245,0.18609,0.93045,0.8288,0.8681895833333286,0.8725463095238056,0.8942745973992838,0.8746616129494023
|
| 277 |
+
6.756756756756757,9500,0.82835,0.902875,0.930225,0.82835,0.82835,0.3009583333333333,0.902875,0.18604500000000002,0.930225,0.82835,0.8679937499999963,0.8723895138888845,0.8941721987188964,0.8745020848542968
|
| 278 |
+
6.827880512091038,9600,0.82875,0.9031,0.9302,0.82875,0.82875,0.3010333333333333,0.9031,0.18604,0.9302,0.82875,0.8681779166666623,0.8725444742063444,0.8942281345359215,0.8746847415751485
|
| 279 |
+
6.89900426742532,9700,0.828775,0.903275,0.929975,0.828775,0.828775,0.3010916666666666,0.903275,0.18599500000000002,0.929975,0.828775,0.8680683333333283,0.8724758234126926,0.8941863586244911,0.8745995624422745
|
| 280 |
+
6.970128022759602,9800,0.828725,0.90295,0.929725,0.828725,0.828725,0.3009833333333333,0.90295,0.18594500000000003,0.929725,0.828725,0.8680020833333291,0.8724609027777737,0.8941897885251688,0.8745898837951301
|
| 281 |
+
7.0412517780938835,9900,0.82885,0.902975,0.9297,0.82885,0.82885,0.3009916666666666,0.902975,0.18594000000000002,0.9297,0.82885,0.8680699999999961,0.8725679563492021,0.8943256778966812,0.8746741755899257
|
| 282 |
+
7.112375533428165,10000,0.828825,0.9032,0.930225,0.828825,0.828825,0.3010666666666666,0.9032,0.18604500000000002,0.930225,0.828825,0.8682116666666623,0.8726781349206311,0.8945003142115322,0.8747602011610026
|
| 283 |
+
7.183499288762446,10100,0.82925,0.903175,0.9307,0.82925,0.82925,0.3010583333333333,0.903175,0.18614,0.9307,0.82925,0.8686112499999957,0.8730120634920597,0.8947492895278029,0.8750990256783056
|
| 284 |
+
7.2546230440967285,10200,0.828975,0.9032,0.929975,0.828975,0.828975,0.30106666666666665,0.9032,0.185995,0.929975,0.828975,0.8683299999999957,0.8728390873015822,0.8946373730573125,0.8749157675276352
|
| 285 |
+
7.32574679943101,10300,0.829325,0.902925,0.930125,0.829325,0.829325,0.3009749999999999,0.902925,0.18602500000000005,0.930125,0.829325,0.8684108333333296,0.8728565773809487,0.89456862638456,0.8749638243818381
|
| 286 |
+
7.396870554765291,10400,0.82955,0.902825,0.930375,0.82955,0.82955,0.3009416666666666,0.902825,0.186075,0.930375,0.82955,0.8686049999999963,0.872999374999997,0.8946617306603606,0.8751100324975649
|
| 287 |
+
7.467994310099574,10500,0.828575,0.90255,0.930075,0.828575,0.828575,0.30085,0.90255,0.186015,0.930075,0.828575,0.8679862499999965,0.8724021329365041,0.89419947680479,0.8745111244014055
|
| 288 |
+
7.539118065433855,10600,0.82875,0.902775,0.93,0.82875,0.82875,0.300925,0.902775,0.186,0.93,0.82875,0.8681016666666623,0.8725284722222181,0.8942673037954326,0.874650465442734
|
| 289 |
+
7.610241820768136,10700,0.829025,0.902775,0.93065,0.829025,0.829025,0.300925,0.902775,0.18613000000000005,0.93065,0.829025,0.8684041666666626,0.8728070238095202,0.8945965165100355,0.8749042361097843
|
| 290 |
+
7.681365576102419,10800,0.828825,0.90305,0.930675,0.828825,0.828825,0.3010166666666666,0.90305,0.18613500000000002,0.930675,0.828825,0.8684024999999962,0.8727813988095199,0.8945637455567085,0.8748808984568961
|
| 291 |
+
7.7524893314367,10900,0.8292,0.903125,0.93095,0.8292,0.8292,0.3010416666666666,0.903125,0.18619000000000002,0.93095,0.8292,0.8686620833333288,0.8730319146825355,0.8948043753706577,0.8751241215815942
|
| 292 |
+
7.823613086770981,11000,0.829,0.9025,0.930925,0.829,0.829,0.3008333333333333,0.9025,0.18618500000000004,0.930925,0.829,0.8684387499999965,0.8728221924603132,0.8946611073435173,0.8748990293437295
|
| 293 |
+
7.894736842105263,11100,0.828975,0.90305,0.930825,0.828975,0.828975,0.3010166666666666,0.90305,0.18616500000000005,0.930825,0.828975,0.8684504166666628,0.872878621031742,0.894732783487921,0.874949510397515
|
| 294 |
+
7.965860597439545,11200,0.829025,0.90335,0.930975,0.829025,0.829025,0.3011166666666666,0.90335,0.18619500000000003,0.930975,0.829025,0.8685791666666625,0.8729389384920585,0.8947336103485135,0.8750321865169739
|
| 295 |
+
8.036984352773827,11300,0.829175,0.903225,0.93075,0.829175,0.829175,0.301075,0.903225,0.18615,0.93075,0.829175,0.8685495833333294,0.8729351289682501,0.8947091937766589,0.8750362483818597
|
| 296 |
+
8.108108108108109,11400,0.828775,0.902825,0.9307,0.828775,0.828775,0.30094166666666666,0.902825,0.18614000000000006,0.9307,0.828775,0.8683158333333291,0.8727185416666623,0.894565804979394,0.8748080562858987
|
| 297 |
+
8.17923186344239,11500,0.82925,0.9035,0.931125,0.82925,0.82925,0.30116666666666664,0.9035,0.18622500000000006,0.931125,0.82925,0.8688029166666628,0.8731468551587266,0.8948947495660065,0.8752462046242943
|
| 298 |
+
8.250355618776672,11600,0.82925,0.90315,0.930525,0.82925,0.82925,0.30105,0.90315,0.18610500000000002,0.930525,0.82925,0.8686154166666618,0.8730597619047572,0.8948103433261322,0.8751613059053036
|
| 299 |
+
8.321479374110954,11700,0.8288,0.903,0.9303,0.8288,0.8288,0.30099999999999993,0.903,0.18606,0.9303,0.8288,0.8682366666666615,0.8727099900793599,0.8945588779528776,0.8748019389609702
|
| 300 |
+
8.392603129445234,11800,0.82915,0.903,0.930975,0.82915,0.82915,0.30099999999999993,0.903,0.186195,0.930975,0.82915,0.8686054166666618,0.8730238888888842,0.8948571327655279,0.8750999528350607
|
| 301 |
+
8.463726884779517,11900,0.8292,0.90305,0.930975,0.8292,0.8292,0.30101666666666665,0.90305,0.18619500000000003,0.930975,0.8292,0.8686795833333283,0.8731174404761854,0.8949856631148606,0.8751770568215895
|
| 302 |
+
8.534850640113799,12000,0.82915,0.902825,0.930825,0.82915,0.82915,0.3009416666666666,0.902825,0.186165,0.930825,0.82915,0.8685429166666616,0.8729799206349158,0.8948425781107103,0.8750457194043624
|
| 303 |
+
8.60597439544808,12100,0.829275,0.9026,0.9306,0.829275,0.829275,0.3008666666666666,0.9026,0.18612000000000004,0.9306,0.829275,0.8685587499999947,0.8730033134920586,0.8947984660739079,0.8750927454028319
|
| 304 |
+
8.677098150782362,12200,0.829275,0.90285,0.930625,0.829275,0.829275,0.3009499999999999,0.90285,0.186125,0.930625,0.829275,0.8685366666666623,0.8729908829365044,0.894820570983912,0.8750702930735716
|
| 305 |
+
8.748221906116642,12300,0.829275,0.902775,0.931025,0.829275,0.829275,0.300925,0.902775,0.186205,0.931025,0.829275,0.8686212499999958,0.8730137202380908,0.8948325362646716,0.8750985303378986
|
| 306 |
+
8.819345661450924,12400,0.829275,0.90275,0.930925,0.829275,0.829275,0.3009166666666666,0.90275,0.18618500000000002,0.930925,0.829275,0.8686174999999962,0.8730114880952342,0.8948164169545031,0.8750965074481144
|
| 307 |
+
8.890469416785207,12500,0.829325,0.902975,0.931,0.829325,0.829325,0.30099166666666666,0.902975,0.18620000000000003,0.931,0.829325,0.8687383333333292,0.8731361408730117,0.8949238533713902,0.8752293018007289
|
| 308 |
+
8.961593172119487,12600,0.8295,0.902775,0.9313,0.8295,0.8295,0.30092499999999994,0.902775,0.18626000000000004,0.9313,0.8295,0.8689020833333292,0.8732708234126947,0.8950697708359244,0.8753445888295041
|
| 309 |
+
9.03271692745377,12700,0.82955,0.903225,0.931225,0.82955,0.82955,0.3010749999999999,0.903225,0.18624500000000002,0.931225,0.82955,0.8689562499999957,0.8733353571428528,0.8951241731895416,0.8754143164227475
|
| 310 |
+
9.103840682788052,12800,0.82935,0.903,0.931075,0.82935,0.82935,0.301,0.903,0.18621500000000002,0.931075,0.82935,0.8687095833333284,0.8731101884920585,0.8949432152346973,0.8751869179331543
|
| 311 |
+
9.174964438122332,12900,0.829275,0.90295,0.930825,0.829275,0.829275,0.30098333333333327,0.90295,0.186165,0.930825,0.829275,0.868580416666662,0.8730058630952335,0.8948285709625057,0.8750920981147012
|
| 312 |
+
9.246088193456615,13000,0.829225,0.902875,0.931025,0.829225,0.829225,0.3009583333333333,0.902875,0.186205,0.931025,0.829225,0.8686254166666617,0.8730646825396782,0.8949587209733254,0.8751212188013575
|
| 313 |
+
9.317211948790897,13100,0.829375,0.9032,0.931225,0.829375,0.829375,0.30106666666666665,0.9032,0.18624500000000002,0.931225,0.829375,0.8687699999999946,0.8731812896825357,0.8950327784818084,0.8752468123212304
|
| 314 |
+
9.388335704125177,13200,0.8296,0.90295,0.931375,0.8296,0.8296,0.3009833333333333,0.90295,0.186275,0.931375,0.8296,0.8689904166666618,0.8734019444444404,0.8952498050830631,0.8754484318300516
|
| 315 |
+
9.45945945945946,13300,0.82955,0.902775,0.93115,0.82955,0.82955,0.30092499999999994,0.902775,0.18623,0.93115,0.82955,0.8688566666666613,0.8732831845238046,0.895126620421129,0.8753385862163494
|
| 316 |
+
9.530583214793742,13400,0.82935,0.9029,0.931,0.82935,0.82935,0.3009666666666666,0.9029,0.1862,0.931,0.82935,0.8686979166666617,0.873155942460313,0.895046161583383,0.8751996221334275
|
| 317 |
+
9.601706970128022,13500,0.829275,0.90305,0.9311,0.829275,0.829275,0.30101666666666665,0.90305,0.18622000000000002,0.9311,0.829275,0.8686891666666617,0.8731191964285671,0.895002811736966,0.875171085337729
|
| 318 |
+
9.672830725462305,13600,0.82945,0.903025,0.931225,0.82945,0.82945,0.3010083333333333,0.903025,0.18624500000000002,0.931225,0.82945,0.8688266666666615,0.8732329166666623,0.8950792684099309,0.8752895149961647
|
| 319 |
+
9.743954480796585,13700,0.82945,0.90305,0.931125,0.82945,0.82945,0.3010166666666666,0.90305,0.18622500000000003,0.931125,0.82945,0.8687870833333281,0.8732131448412651,0.895073130294718,0.8752667124340483
|
| 320 |
+
9.815078236130867,13800,0.82945,0.9031,0.931025,0.82945,0.82945,0.3010333333333333,0.9031,0.186205,0.931025,0.82945,0.8687966666666614,0.8732481845238048,0.8951075415333131,0.8753000882548883
|
| 321 |
+
9.88620199146515,13900,0.82935,0.903025,0.931025,0.82935,0.82935,0.3010083333333333,0.903025,0.18620500000000004,0.931025,0.82935,0.8687354166666617,0.8731747619047575,0.895030096393872,0.8752349923053447
|
| 322 |
+
9.95732574679943,14000,0.82935,0.903025,0.9311,0.82935,0.82935,0.30100833333333327,0.903025,0.18622,0.9311,0.82935,0.8687558333333282,0.8731832242063449,0.8950372962037911,0.8752427301346968
|
final_metrics.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
{
|
| 2 |
-
"val_cosine_accuracy@1": 0.
|
| 3 |
-
"val_cosine_accuracy@3": 0.
|
| 4 |
-
"val_cosine_accuracy@5": 0.
|
| 5 |
-
"val_cosine_precision@1": 0.
|
| 6 |
-
"val_cosine_precision@3": 0.
|
| 7 |
-
"val_cosine_precision@5": 0.
|
| 8 |
-
"val_cosine_recall@1": 0.
|
| 9 |
-
"val_cosine_recall@3": 0.
|
| 10 |
-
"val_cosine_recall@5": 0.
|
| 11 |
-
"val_cosine_ndcg@10": 0.
|
| 12 |
-
"val_cosine_mrr@1": 0.
|
| 13 |
-
"val_cosine_mrr@5": 0.
|
| 14 |
-
"val_cosine_mrr@10": 0.
|
| 15 |
-
"val_cosine_map@100": 0.
|
| 16 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"val_cosine_accuracy@1": 0.8281,
|
| 3 |
+
"val_cosine_accuracy@3": 0.9026,
|
| 4 |
+
"val_cosine_accuracy@5": 0.93105,
|
| 5 |
+
"val_cosine_precision@1": 0.8281,
|
| 6 |
+
"val_cosine_precision@3": 0.3008666666666666,
|
| 7 |
+
"val_cosine_precision@5": 0.18621000000000004,
|
| 8 |
+
"val_cosine_recall@1": 0.8281,
|
| 9 |
+
"val_cosine_recall@3": 0.9026,
|
| 10 |
+
"val_cosine_recall@5": 0.93105,
|
| 11 |
+
"val_cosine_ndcg@10": 0.8942437004811851,
|
| 12 |
+
"val_cosine_mrr@1": 0.8281,
|
| 13 |
+
"val_cosine_mrr@5": 0.8677437499999962,
|
| 14 |
+
"val_cosine_mrr@10": 0.8721381249999942,
|
| 15 |
+
"val_cosine_map@100": 0.874246358340888
|
| 16 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 114011616
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:07288e0fdf774af44164e803f88a32086d4b380ab0234eb9fc654ddb063c8150
|
| 3 |
size 114011616
|