Training in progress, step 14060
Browse files- Information-Retrieval_evaluation_val_results.csv +1 -0
- README.md +73 -349
- eval/Information-Retrieval_evaluation_val_results.csv +141 -0
- final_metrics.json +14 -14
- model.safetensors +1 -1
Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -1,3 +1,4 @@
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.9208,0.9698,0.9842,0.9208,0.9208,0.3232666666666667,0.9698,0.19684,0.9842,0.9208,0.9460899999999998,0.9476021428571432,0.9593212690041523,0.9479260307963899
|
| 3 |
-1,-1,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451366666666663,0.9466038095238087,0.9586270476620361,0.946959374340519
|
|
|
|
|
|
| 1 |
epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precision@1,cosine-Recall@1,cosine-Precision@3,cosine-Recall@3,cosine-Precision@5,cosine-Recall@5,cosine-MRR@1,cosine-MRR@5,cosine-MRR@10,cosine-NDCG@10,cosine-MAP@100
|
| 2 |
-1,-1,0.9208,0.9698,0.9842,0.9208,0.9208,0.3232666666666667,0.9698,0.19684,0.9842,0.9208,0.9460899999999998,0.9476021428571432,0.9593212690041523,0.9479260307963899
|
| 3 |
-1,-1,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451366666666663,0.9466038095238087,0.9586270476620361,0.946959374340519
|
| 4 |
+
-1,-1,0.829575,0.9048,0.9324,0.829575,0.829575,0.3016,0.9048,0.18648000000000003,0.9324,0.829575,0.8693266666666628,0.873717658730154,0.8957411186558171,0.8757871539962314
|
README.md
CHANGED
|
@@ -5,114 +5,38 @@ tags:
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
-
- dataset_size:
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
-
- source_sentence:
|
| 13 |
-
for one that's not married? Which one is for what?
|
| 14 |
sentences:
|
| 15 |
-
-
|
| 16 |
-
|
| 17 |
-
-
|
| 18 |
-
|
| 19 |
-
- source_sentence: Which ointment is applied to the face of UFC fighters at the commencement
|
| 20 |
-
of a bout? What does it do?
|
| 21 |
sentences:
|
| 22 |
-
-
|
| 23 |
-
-
|
| 24 |
-
|
| 25 |
-
-
|
| 26 |
-
- source_sentence:
|
| 27 |
sentences:
|
| 28 |
-
-
|
| 29 |
-
-
|
| 30 |
-
-
|
| 31 |
-
|
| 32 |
-
- source_sentence: Can I do shoulder and triceps workout on same day? What other combinations
|
| 33 |
-
like this can I do?
|
| 34 |
sentences:
|
| 35 |
-
-
|
| 36 |
-
|
| 37 |
-
- How
|
| 38 |
-
|
| 39 |
-
this can I do?
|
| 40 |
-
- source_sentence: Ordered food on Swiggy 3 days ago.After accepting my money, said
|
| 41 |
-
no more on Menu! When if ever will I atleast get refund in cr card a/c?
|
| 42 |
sentences:
|
| 43 |
-
-
|
| 44 |
-
-
|
| 45 |
-
-
|
| 46 |
-
When if ever will I atleast get refund in cr card a/c?
|
| 47 |
pipeline_tag: sentence-similarity
|
| 48 |
library_name: sentence-transformers
|
| 49 |
-
metrics:
|
| 50 |
-
- cosine_accuracy@1
|
| 51 |
-
- cosine_accuracy@3
|
| 52 |
-
- cosine_accuracy@5
|
| 53 |
-
- cosine_precision@1
|
| 54 |
-
- cosine_precision@3
|
| 55 |
-
- cosine_precision@5
|
| 56 |
-
- cosine_recall@1
|
| 57 |
-
- cosine_recall@3
|
| 58 |
-
- cosine_recall@5
|
| 59 |
-
- cosine_ndcg@10
|
| 60 |
-
- cosine_mrr@1
|
| 61 |
-
- cosine_mrr@5
|
| 62 |
-
- cosine_mrr@10
|
| 63 |
-
- cosine_map@100
|
| 64 |
-
model-index:
|
| 65 |
-
- name: SentenceTransformer based on prajjwal1/bert-small
|
| 66 |
-
results:
|
| 67 |
-
- task:
|
| 68 |
-
type: information-retrieval
|
| 69 |
-
name: Information Retrieval
|
| 70 |
-
dataset:
|
| 71 |
-
name: val
|
| 72 |
-
type: val
|
| 73 |
-
metrics:
|
| 74 |
-
- type: cosine_accuracy@1
|
| 75 |
-
value: 0.829675
|
| 76 |
-
name: Cosine Accuracy@1
|
| 77 |
-
- type: cosine_accuracy@3
|
| 78 |
-
value: 0.9048
|
| 79 |
-
name: Cosine Accuracy@3
|
| 80 |
-
- type: cosine_accuracy@5
|
| 81 |
-
value: 0.93245
|
| 82 |
-
name: Cosine Accuracy@5
|
| 83 |
-
- type: cosine_precision@1
|
| 84 |
-
value: 0.829675
|
| 85 |
-
name: Cosine Precision@1
|
| 86 |
-
- type: cosine_precision@3
|
| 87 |
-
value: 0.3016
|
| 88 |
-
name: Cosine Precision@3
|
| 89 |
-
- type: cosine_precision@5
|
| 90 |
-
value: 0.18649000000000004
|
| 91 |
-
name: Cosine Precision@5
|
| 92 |
-
- type: cosine_recall@1
|
| 93 |
-
value: 0.829675
|
| 94 |
-
name: Cosine Recall@1
|
| 95 |
-
- type: cosine_recall@3
|
| 96 |
-
value: 0.9048
|
| 97 |
-
name: Cosine Recall@3
|
| 98 |
-
- type: cosine_recall@5
|
| 99 |
-
value: 0.93245
|
| 100 |
-
name: Cosine Recall@5
|
| 101 |
-
- type: cosine_ndcg@10
|
| 102 |
-
value: 0.8957919450437679
|
| 103 |
-
name: Cosine Ndcg@10
|
| 104 |
-
- type: cosine_mrr@1
|
| 105 |
-
value: 0.829675
|
| 106 |
-
name: Cosine Mrr@1
|
| 107 |
-
- type: cosine_mrr@5
|
| 108 |
-
value: 0.8693824999999958
|
| 109 |
-
name: Cosine Mrr@5
|
| 110 |
-
- type: cosine_mrr@10
|
| 111 |
-
value: 0.8737714285714238
|
| 112 |
-
name: Cosine Mrr@10
|
| 113 |
-
- type: cosine_map@100
|
| 114 |
-
value: 0.8758361833602419
|
| 115 |
-
name: Cosine Map@100
|
| 116 |
---
|
| 117 |
|
| 118 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
@@ -161,12 +85,12 @@ Then you can load this model and run inference.
|
|
| 161 |
from sentence_transformers import SentenceTransformer
|
| 162 |
|
| 163 |
# Download from the 🤗 Hub
|
| 164 |
-
model = SentenceTransformer("
|
| 165 |
# Run inference
|
| 166 |
sentences = [
|
| 167 |
-
'
|
| 168 |
-
'
|
| 169 |
-
'
|
| 170 |
]
|
| 171 |
embeddings = model.encode(sentences)
|
| 172 |
print(embeddings.shape)
|
|
@@ -175,9 +99,9 @@ print(embeddings.shape)
|
|
| 175 |
# Get the similarity scores for the embeddings
|
| 176 |
similarities = model.similarity(embeddings, embeddings)
|
| 177 |
print(similarities)
|
| 178 |
-
# tensor([[1.0000,
|
| 179 |
-
# [
|
| 180 |
-
# [0.
|
| 181 |
```
|
| 182 |
|
| 183 |
<!--
|
|
@@ -204,32 +128,6 @@ You can finetune this model on your own dataset.
|
|
| 204 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 205 |
-->
|
| 206 |
|
| 207 |
-
## Evaluation
|
| 208 |
-
|
| 209 |
-
### Metrics
|
| 210 |
-
|
| 211 |
-
#### Information Retrieval
|
| 212 |
-
|
| 213 |
-
* Dataset: `val`
|
| 214 |
-
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
| 215 |
-
|
| 216 |
-
| Metric | Value |
|
| 217 |
-
|:-------------------|:-----------|
|
| 218 |
-
| cosine_accuracy@1 | 0.8297 |
|
| 219 |
-
| cosine_accuracy@3 | 0.9048 |
|
| 220 |
-
| cosine_accuracy@5 | 0.9325 |
|
| 221 |
-
| cosine_precision@1 | 0.8297 |
|
| 222 |
-
| cosine_precision@3 | 0.3016 |
|
| 223 |
-
| cosine_precision@5 | 0.1865 |
|
| 224 |
-
| cosine_recall@1 | 0.8297 |
|
| 225 |
-
| cosine_recall@3 | 0.9048 |
|
| 226 |
-
| cosine_recall@5 | 0.9325 |
|
| 227 |
-
| **cosine_ndcg@10** | **0.8958** |
|
| 228 |
-
| cosine_mrr@1 | 0.8297 |
|
| 229 |
-
| cosine_mrr@5 | 0.8694 |
|
| 230 |
-
| cosine_mrr@10 | 0.8738 |
|
| 231 |
-
| cosine_map@100 | 0.8758 |
|
| 232 |
-
|
| 233 |
<!--
|
| 234 |
## Bias, Risks and Limitations
|
| 235 |
|
|
@@ -248,45 +146,19 @@ You can finetune this model on your own dataset.
|
|
| 248 |
|
| 249 |
#### Unnamed Dataset
|
| 250 |
|
| 251 |
-
* Size:
|
| 252 |
-
* Columns: <code>
|
| 253 |
-
* Approximate statistics based on the first 1000 samples:
|
| 254 |
-
| | anchor | positive | negative |
|
| 255 |
-
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
|
| 256 |
-
| type | string | string | string |
|
| 257 |
-
| details | <ul><li>min: 4 tokens</li><li>mean: 15.46 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 15.52 tokens</li><li>max: 49 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 17.07 tokens</li><li>max: 128 tokens</li></ul> |
|
| 258 |
-
* Samples:
|
| 259 |
-
| anchor | positive | negative |
|
| 260 |
-
|:--------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
|
| 261 |
-
| <code>Shall I upgrade my iPhone 5s to iOS 10 final version?</code> | <code>Should I upgrade an iPhone 5s to iOS 10?</code> | <code>How, if at all, is the accent, pitch and voice of gay men different than that of straight men and how accurate is voice in determining sexual orientation?</code> |
|
| 262 |
-
| <code>Is Donald Trump really going to be the president of United States?</code> | <code>Do you think Donald Trump could conceivably be the next President of the United States?</code> | <code>Can a 15-year-old boy and an 18-year-old girl have sex?</code> |
|
| 263 |
-
| <code>What are real tips to improve work life balance?</code> | <code>What are the best ways to create a work life balance?</code> | <code>How far is Miami from Fort Lauderdale?</code> |
|
| 264 |
-
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 265 |
-
```json
|
| 266 |
-
{
|
| 267 |
-
"scale": 20.0,
|
| 268 |
-
"similarity_fct": "cos_sim",
|
| 269 |
-
"gather_across_devices": false
|
| 270 |
-
}
|
| 271 |
-
```
|
| 272 |
-
|
| 273 |
-
### Evaluation Dataset
|
| 274 |
-
|
| 275 |
-
#### Unnamed Dataset
|
| 276 |
-
|
| 277 |
-
* Size: 40,000 evaluation samples
|
| 278 |
-
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
| 279 |
* Approximate statistics based on the first 1000 samples:
|
| 280 |
-
| |
|
| 281 |
-
|
| 282 |
-
| type | string | string
|
| 283 |
-
| details | <ul><li>min:
|
| 284 |
* Samples:
|
| 285 |
-
|
|
| 286 |
-
|
| 287 |
-
| <code>
|
| 288 |
-
| <code>
|
| 289 |
-
| <code>
|
| 290 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 291 |
```json
|
| 292 |
{
|
|
@@ -299,49 +171,36 @@ You can finetune this model on your own dataset.
|
|
| 299 |
### Training Hyperparameters
|
| 300 |
#### Non-Default Hyperparameters
|
| 301 |
|
| 302 |
-
- `
|
| 303 |
-
- `
|
| 304 |
-
- `per_device_eval_batch_size`: 256
|
| 305 |
-
- `learning_rate`: 2e-05
|
| 306 |
-
- `weight_decay`: 0.001
|
| 307 |
-
- `max_steps`: 14060
|
| 308 |
-
- `warmup_ratio`: 0.1
|
| 309 |
- `fp16`: True
|
| 310 |
-
- `
|
| 311 |
-
- `dataloader_num_workers`: 1
|
| 312 |
-
- `dataloader_prefetch_factor`: 1
|
| 313 |
-
- `load_best_model_at_end`: True
|
| 314 |
-
- `optim`: adamw_torch
|
| 315 |
-
- `ddp_find_unused_parameters`: False
|
| 316 |
-
- `push_to_hub`: True
|
| 317 |
-
- `hub_model_id`: redis/model-a-baseline
|
| 318 |
-
- `eval_on_start`: True
|
| 319 |
|
| 320 |
#### All Hyperparameters
|
| 321 |
<details><summary>Click to expand</summary>
|
| 322 |
|
| 323 |
- `overwrite_output_dir`: False
|
| 324 |
- `do_predict`: False
|
| 325 |
-
- `eval_strategy`:
|
| 326 |
- `prediction_loss_only`: True
|
| 327 |
-
- `per_device_train_batch_size`:
|
| 328 |
-
- `per_device_eval_batch_size`:
|
| 329 |
- `per_gpu_train_batch_size`: None
|
| 330 |
- `per_gpu_eval_batch_size`: None
|
| 331 |
- `gradient_accumulation_steps`: 1
|
| 332 |
- `eval_accumulation_steps`: None
|
| 333 |
- `torch_empty_cache_steps`: None
|
| 334 |
-
- `learning_rate`:
|
| 335 |
-
- `weight_decay`: 0.
|
| 336 |
- `adam_beta1`: 0.9
|
| 337 |
- `adam_beta2`: 0.999
|
| 338 |
- `adam_epsilon`: 1e-08
|
| 339 |
-
- `max_grad_norm`: 1
|
| 340 |
-
- `num_train_epochs`: 3
|
| 341 |
-
- `max_steps`:
|
| 342 |
- `lr_scheduler_type`: linear
|
| 343 |
- `lr_scheduler_kwargs`: {}
|
| 344 |
-
- `warmup_ratio`: 0.
|
| 345 |
- `warmup_steps`: 0
|
| 346 |
- `log_level`: passive
|
| 347 |
- `log_level_replica`: warning
|
|
@@ -369,14 +228,14 @@ You can finetune this model on your own dataset.
|
|
| 369 |
- `tpu_num_cores`: None
|
| 370 |
- `tpu_metrics_debug`: False
|
| 371 |
- `debug`: []
|
| 372 |
-
- `dataloader_drop_last`:
|
| 373 |
-
- `dataloader_num_workers`:
|
| 374 |
-
- `dataloader_prefetch_factor`:
|
| 375 |
- `past_index`: -1
|
| 376 |
- `disable_tqdm`: False
|
| 377 |
- `remove_unused_columns`: True
|
| 378 |
- `label_names`: None
|
| 379 |
-
- `load_best_model_at_end`:
|
| 380 |
- `ignore_data_skip`: False
|
| 381 |
- `fsdp`: []
|
| 382 |
- `fsdp_min_num_params`: 0
|
|
@@ -386,23 +245,23 @@ You can finetune this model on your own dataset.
|
|
| 386 |
- `parallelism_config`: None
|
| 387 |
- `deepspeed`: None
|
| 388 |
- `label_smoothing_factor`: 0.0
|
| 389 |
-
- `optim`:
|
| 390 |
- `optim_args`: None
|
| 391 |
- `adafactor`: False
|
| 392 |
- `group_by_length`: False
|
| 393 |
- `length_column_name`: length
|
| 394 |
- `project`: huggingface
|
| 395 |
- `trackio_space_id`: trackio
|
| 396 |
-
- `ddp_find_unused_parameters`:
|
| 397 |
- `ddp_bucket_cap_mb`: None
|
| 398 |
- `ddp_broadcast_buffers`: False
|
| 399 |
- `dataloader_pin_memory`: True
|
| 400 |
- `dataloader_persistent_workers`: False
|
| 401 |
- `skip_memory_metrics`: True
|
| 402 |
- `use_legacy_prediction_loop`: False
|
| 403 |
-
- `push_to_hub`:
|
| 404 |
- `resume_from_checkpoint`: None
|
| 405 |
-
- `hub_model_id`:
|
| 406 |
- `hub_strategy`: every_save
|
| 407 |
- `hub_private_repo`: None
|
| 408 |
- `hub_always_push`: False
|
|
@@ -429,167 +288,32 @@ You can finetune this model on your own dataset.
|
|
| 429 |
- `neftune_noise_alpha`: None
|
| 430 |
- `optim_target_modules`: None
|
| 431 |
- `batch_eval_metrics`: False
|
| 432 |
-
- `eval_on_start`:
|
| 433 |
- `use_liger_kernel`: False
|
| 434 |
- `liger_kernel_config`: None
|
| 435 |
- `eval_use_gather_object`: False
|
| 436 |
- `average_tokens_across_devices`: True
|
| 437 |
- `prompts`: None
|
| 438 |
- `batch_sampler`: batch_sampler
|
| 439 |
-
- `multi_dataset_batch_sampler`:
|
| 440 |
- `router_mapping`: {}
|
| 441 |
- `learning_rate_mapping`: {}
|
| 442 |
|
| 443 |
</details>
|
| 444 |
|
| 445 |
### Training Logs
|
| 446 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 447 |
|
| 448 |
-
| Epoch | Step | Training Loss | Validation Loss | val_cosine_ndcg@10 |
|
| 449 |
-
|:------:|:-----:|:-------------:|:---------------:|:------------------:|
|
| 450 |
-
| 0 | 0 | - | 1.6256 | 0.8046 |
|
| 451 |
-
| 0.0711 | 100 | 1.9658 | 0.3149 | 0.8458 |
|
| 452 |
-
| 0.1422 | 200 | 0.4018 | 0.1192 | 0.8670 |
|
| 453 |
-
| 0.2134 | 300 | 0.1963 | 0.0916 | 0.8733 |
|
| 454 |
-
| 0.2845 | 400 | 0.163 | 0.0803 | 0.8766 |
|
| 455 |
-
| 0.3556 | 500 | 0.1397 | 0.0729 | 0.8783 |
|
| 456 |
-
| 0.4267 | 600 | 0.1261 | 0.0681 | 0.8798 |
|
| 457 |
-
| 0.4979 | 700 | 0.1181 | 0.0640 | 0.8810 |
|
| 458 |
-
| 0.5690 | 800 | 0.1166 | 0.0614 | 0.8818 |
|
| 459 |
-
| 0.6401 | 900 | 0.1068 | 0.0586 | 0.8829 |
|
| 460 |
-
| 0.7112 | 1000 | 0.1001 | 0.0564 | 0.8835 |
|
| 461 |
-
| 0.7824 | 1100 | 0.097 | 0.0549 | 0.8842 |
|
| 462 |
-
| 0.8535 | 1200 | 0.0941 | 0.0530 | 0.8844 |
|
| 463 |
-
| 0.9246 | 1300 | 0.0934 | 0.0515 | 0.8854 |
|
| 464 |
-
| 0.9957 | 1400 | 0.086 | 0.0499 | 0.8859 |
|
| 465 |
-
| 1.0669 | 1500 | 0.081 | 0.0482 | 0.8864 |
|
| 466 |
-
| 1.1380 | 1600 | 0.0778 | 0.0476 | 0.8868 |
|
| 467 |
-
| 1.2091 | 1700 | 0.0734 | 0.0459 | 0.8870 |
|
| 468 |
-
| 1.2802 | 1800 | 0.0744 | 0.0459 | 0.8872 |
|
| 469 |
-
| 1.3514 | 1900 | 0.0734 | 0.0447 | 0.8880 |
|
| 470 |
-
| 1.4225 | 2000 | 0.0695 | 0.0440 | 0.8883 |
|
| 471 |
-
| 1.4936 | 2100 | 0.0671 | 0.0438 | 0.8880 |
|
| 472 |
-
| 1.5647 | 2200 | 0.0703 | 0.0435 | 0.8884 |
|
| 473 |
-
| 1.6358 | 2300 | 0.0658 | 0.0425 | 0.8886 |
|
| 474 |
-
| 1.7070 | 2400 | 0.0695 | 0.0416 | 0.8894 |
|
| 475 |
-
| 1.7781 | 2500 | 0.0665 | 0.0413 | 0.8893 |
|
| 476 |
-
| 1.8492 | 2600 | 0.0648 | 0.0407 | 0.8896 |
|
| 477 |
-
| 1.9203 | 2700 | 0.0668 | 0.0405 | 0.8894 |
|
| 478 |
-
| 1.9915 | 2800 | 0.0636 | 0.0398 | 0.8902 |
|
| 479 |
-
| 2.0626 | 2900 | 0.0583 | 0.0394 | 0.8903 |
|
| 480 |
-
| 2.1337 | 3000 | 0.0583 | 0.0390 | 0.8903 |
|
| 481 |
-
| 2.2048 | 3100 | 0.0563 | 0.0385 | 0.8908 |
|
| 482 |
-
| 2.2760 | 3200 | 0.0588 | 0.0384 | 0.8910 |
|
| 483 |
-
| 2.3471 | 3300 | 0.0586 | 0.0383 | 0.8910 |
|
| 484 |
-
| 2.4182 | 3400 | 0.0557 | 0.0380 | 0.8907 |
|
| 485 |
-
| 2.4893 | 3500 | 0.0561 | 0.0378 | 0.8910 |
|
| 486 |
-
| 2.5605 | 3600 | 0.0557 | 0.0375 | 0.8912 |
|
| 487 |
-
| 2.6316 | 3700 | 0.0553 | 0.0371 | 0.8915 |
|
| 488 |
-
| 2.7027 | 3800 | 0.053 | 0.0371 | 0.8913 |
|
| 489 |
-
| 2.7738 | 3900 | 0.0562 | 0.0369 | 0.8916 |
|
| 490 |
-
| 2.8450 | 4000 | 0.0531 | 0.0368 | 0.8914 |
|
| 491 |
-
| 2.9161 | 4100 | 0.0522 | 0.0363 | 0.8918 |
|
| 492 |
-
| 2.9872 | 4200 | 0.0555 | 0.0363 | 0.8919 |
|
| 493 |
-
| 3.0583 | 4300 | 0.0524 | 0.0363 | 0.8921 |
|
| 494 |
-
| 3.1294 | 4400 | 0.0512 | 0.0356 | 0.8924 |
|
| 495 |
-
| 3.2006 | 4500 | 0.0503 | 0.0356 | 0.8922 |
|
| 496 |
-
| 3.2717 | 4600 | 0.0523 | 0.0355 | 0.8925 |
|
| 497 |
-
| 3.3428 | 4700 | 0.0524 | 0.0351 | 0.8925 |
|
| 498 |
-
| 3.4139 | 4800 | 0.0478 | 0.0351 | 0.8925 |
|
| 499 |
-
| 3.4851 | 4900 | 0.0506 | 0.0347 | 0.8929 |
|
| 500 |
-
| 3.5562 | 5000 | 0.0486 | 0.0344 | 0.8929 |
|
| 501 |
-
| 3.6273 | 5100 | 0.0496 | 0.0344 | 0.8929 |
|
| 502 |
-
| 3.6984 | 5200 | 0.0486 | 0.0345 | 0.8928 |
|
| 503 |
-
| 3.7696 | 5300 | 0.0441 | 0.0343 | 0.8928 |
|
| 504 |
-
| 3.8407 | 5400 | 0.0502 | 0.0342 | 0.8929 |
|
| 505 |
-
| 3.9118 | 5500 | 0.0498 | 0.0341 | 0.8931 |
|
| 506 |
-
| 3.9829 | 5600 | 0.0499 | 0.0342 | 0.8932 |
|
| 507 |
-
| 4.0541 | 5700 | 0.0483 | 0.0339 | 0.8933 |
|
| 508 |
-
| 4.1252 | 5800 | 0.046 | 0.0338 | 0.8933 |
|
| 509 |
-
| 4.1963 | 5900 | 0.0481 | 0.0337 | 0.8933 |
|
| 510 |
-
| 4.2674 | 6000 | 0.0435 | 0.0335 | 0.8936 |
|
| 511 |
-
| 4.3385 | 6100 | 0.0459 | 0.0335 | 0.8935 |
|
| 512 |
-
| 4.4097 | 6200 | 0.0467 | 0.0335 | 0.8933 |
|
| 513 |
-
| 4.4808 | 6300 | 0.0452 | 0.0334 | 0.8937 |
|
| 514 |
-
| 4.5519 | 6400 | 0.0436 | 0.0330 | 0.8940 |
|
| 515 |
-
| 4.6230 | 6500 | 0.0447 | 0.0329 | 0.8940 |
|
| 516 |
-
| 4.6942 | 6600 | 0.0474 | 0.0328 | 0.8940 |
|
| 517 |
-
| 4.7653 | 6700 | 0.0419 | 0.0328 | 0.8940 |
|
| 518 |
-
| 4.8364 | 6800 | 0.0456 | 0.0327 | 0.8939 |
|
| 519 |
-
| 4.9075 | 6900 | 0.0464 | 0.0328 | 0.8940 |
|
| 520 |
-
| 4.9787 | 7000 | 0.0432 | 0.0326 | 0.8940 |
|
| 521 |
-
| 5.0498 | 7100 | 0.0409 | 0.0326 | 0.8939 |
|
| 522 |
-
| 5.1209 | 7200 | 0.0411 | 0.0324 | 0.8941 |
|
| 523 |
-
| 5.1920 | 7300 | 0.0415 | 0.0326 | 0.8940 |
|
| 524 |
-
| 5.2632 | 7400 | 0.0424 | 0.0324 | 0.8943 |
|
| 525 |
-
| 5.3343 | 7500 | 0.0436 | 0.0324 | 0.8942 |
|
| 526 |
-
| 5.4054 | 7600 | 0.0431 | 0.0325 | 0.8942 |
|
| 527 |
-
| 5.4765 | 7700 | 0.0433 | 0.0324 | 0.8941 |
|
| 528 |
-
| 5.5477 | 7800 | 0.0421 | 0.0323 | 0.8943 |
|
| 529 |
-
| 5.6188 | 7900 | 0.0423 | 0.0321 | 0.8944 |
|
| 530 |
-
| 5.6899 | 8000 | 0.0427 | 0.0322 | 0.8947 |
|
| 531 |
-
| 5.7610 | 8100 | 0.0426 | 0.0321 | 0.8946 |
|
| 532 |
-
| 5.8321 | 8200 | 0.0415 | 0.0323 | 0.8944 |
|
| 533 |
-
| 5.9033 | 8300 | 0.0415 | 0.0320 | 0.8946 |
|
| 534 |
-
| 5.9744 | 8400 | 0.0403 | 0.0321 | 0.8947 |
|
| 535 |
-
| 6.0455 | 8500 | 0.0416 | 0.0318 | 0.8949 |
|
| 536 |
-
| 6.1166 | 8600 | 0.0391 | 0.0317 | 0.8949 |
|
| 537 |
-
| 6.1878 | 8700 | 0.0408 | 0.0316 | 0.8949 |
|
| 538 |
-
| 6.2589 | 8800 | 0.0405 | 0.0316 | 0.8950 |
|
| 539 |
-
| 6.3300 | 8900 | 0.041 | 0.0316 | 0.8950 |
|
| 540 |
-
| 6.4011 | 9000 | 0.041 | 0.0316 | 0.8947 |
|
| 541 |
-
| 6.4723 | 9100 | 0.0396 | 0.0315 | 0.8949 |
|
| 542 |
-
| 6.5434 | 9200 | 0.0416 | 0.0315 | 0.8949 |
|
| 543 |
-
| 6.6145 | 9300 | 0.0411 | 0.0315 | 0.8947 |
|
| 544 |
-
| 6.6856 | 9400 | 0.0387 | 0.0315 | 0.8948 |
|
| 545 |
-
| 6.7568 | 9500 | 0.0393 | 0.0315 | 0.8951 |
|
| 546 |
-
| 6.8279 | 9600 | 0.0379 | 0.0314 | 0.8951 |
|
| 547 |
-
| 6.8990 | 9700 | 0.0409 | 0.0313 | 0.8952 |
|
| 548 |
-
| 6.9701 | 9800 | 0.0417 | 0.0312 | 0.8952 |
|
| 549 |
-
| 7.0413 | 9900 | 0.0412 | 0.0312 | 0.8950 |
|
| 550 |
-
| 7.1124 | 10000 | 0.0386 | 0.0312 | 0.8951 |
|
| 551 |
-
| 7.1835 | 10100 | 0.0397 | 0.0312 | 0.8952 |
|
| 552 |
-
| 7.2546 | 10200 | 0.0396 | 0.0311 | 0.8953 |
|
| 553 |
-
| 7.3257 | 10300 | 0.0385 | 0.0312 | 0.8952 |
|
| 554 |
-
| 7.3969 | 10400 | 0.0364 | 0.0310 | 0.8952 |
|
| 555 |
-
| 7.4680 | 10500 | 0.0387 | 0.0310 | 0.8952 |
|
| 556 |
-
| 7.5391 | 10600 | 0.0356 | 0.0309 | 0.8953 |
|
| 557 |
-
| 7.6102 | 10700 | 0.0384 | 0.0310 | 0.8951 |
|
| 558 |
-
| 7.6814 | 10800 | 0.0381 | 0.0308 | 0.8953 |
|
| 559 |
-
| 7.7525 | 10900 | 0.0407 | 0.0309 | 0.8954 |
|
| 560 |
-
| 7.8236 | 11000 | 0.0398 | 0.0308 | 0.8955 |
|
| 561 |
-
| 7.8947 | 11100 | 0.0396 | 0.0307 | 0.8955 |
|
| 562 |
-
| 7.9659 | 11200 | 0.0381 | 0.0307 | 0.8957 |
|
| 563 |
-
| 8.0370 | 11300 | 0.0411 | 0.0308 | 0.8955 |
|
| 564 |
-
| 8.1081 | 11400 | 0.0377 | 0.0307 | 0.8955 |
|
| 565 |
-
| 8.1792 | 11500 | 0.0369 | 0.0307 | 0.8954 |
|
| 566 |
-
| 8.2504 | 11600 | 0.0355 | 0.0307 | 0.8956 |
|
| 567 |
-
| 8.3215 | 11700 | 0.0395 | 0.0307 | 0.8954 |
|
| 568 |
-
| 8.3926 | 11800 | 0.0376 | 0.0306 | 0.8956 |
|
| 569 |
-
| 8.4637 | 11900 | 0.0384 | 0.0307 | 0.8957 |
|
| 570 |
-
| 8.5349 | 12000 | 0.0369 | 0.0306 | 0.8957 |
|
| 571 |
-
| 8.6060 | 12100 | 0.0371 | 0.0306 | 0.8957 |
|
| 572 |
-
| 8.6771 | 12200 | 0.0343 | 0.0306 | 0.8957 |
|
| 573 |
-
| 8.7482 | 12300 | 0.0374 | 0.0305 | 0.8957 |
|
| 574 |
-
| 8.8193 | 12400 | 0.0376 | 0.0305 | 0.8957 |
|
| 575 |
-
| 8.8905 | 12500 | 0.0365 | 0.0305 | 0.8957 |
|
| 576 |
-
| 8.9616 | 12600 | 0.0373 | 0.0304 | 0.8957 |
|
| 577 |
-
| 9.0327 | 12700 | 0.0381 | 0.0305 | 0.8957 |
|
| 578 |
-
| 9.1038 | 12800 | 0.0356 | 0.0305 | 0.8956 |
|
| 579 |
-
| 9.1750 | 12900 | 0.0384 | 0.0305 | 0.8955 |
|
| 580 |
-
| 9.2461 | 13000 | 0.0369 | 0.0304 | 0.8956 |
|
| 581 |
-
| 9.3172 | 13100 | 0.0399 | 0.0304 | 0.8956 |
|
| 582 |
-
| 9.3883 | 13200 | 0.0382 | 0.0304 | 0.8957 |
|
| 583 |
-
| 9.4595 | 13300 | 0.0357 | 0.0304 | 0.8956 |
|
| 584 |
-
| 9.5306 | 13400 | 0.0375 | 0.0304 | 0.8957 |
|
| 585 |
-
| 9.6017 | 13500 | 0.0362 | 0.0304 | 0.8956 |
|
| 586 |
-
| 9.6728 | 13600 | 0.0374 | 0.0304 | 0.8957 |
|
| 587 |
-
| 9.7440 | 13700 | 0.0397 | 0.0304 | 0.8957 |
|
| 588 |
-
| 9.8151 | 13800 | 0.0385 | 0.0304 | 0.8957 |
|
| 589 |
-
| 9.8862 | 13900 | 0.0383 | 0.0304 | 0.8957 |
|
| 590 |
-
| 9.9573 | 14000 | 0.0384 | 0.0304 | 0.8958 |
|
| 591 |
-
|
| 592 |
-
</details>
|
| 593 |
|
| 594 |
### Framework Versions
|
| 595 |
- Python: 3.10.18
|
|
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
+
- dataset_size:100000
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
+
- source_sentence: How do I polish my English skills?
|
|
|
|
| 13 |
sentences:
|
| 14 |
+
- How can we polish English skills?
|
| 15 |
+
- Why should I move to Israel as a Jew?
|
| 16 |
+
- What are vitamins responsible for?
|
| 17 |
+
- source_sentence: Can I use the Kozuka Gothic Pro font as a font-face on my web site?
|
|
|
|
|
|
|
| 18 |
sentences:
|
| 19 |
+
- Can I use the Kozuka Gothic Pro font as a font-face on my web site?
|
| 20 |
+
- Why are Google, Facebook, YouTube and other social networking sites banned in
|
| 21 |
+
China?
|
| 22 |
+
- What font is used in Bloomberg Terminal?
|
| 23 |
+
- source_sentence: Is Quora the best Q&A site?
|
| 24 |
sentences:
|
| 25 |
+
- What was the best Quora question ever?
|
| 26 |
+
- Is Quora the best inquiry site?
|
| 27 |
+
- Where do I buy Oway hair products online?
|
| 28 |
+
- source_sentence: How can I customize my walking speed on Google Maps?
|
|
|
|
|
|
|
| 29 |
sentences:
|
| 30 |
+
- How do I bring back Google maps icon in my home screen?
|
| 31 |
+
- How many pages are there in all the Harry Potter books combined?
|
| 32 |
+
- How can I customize my walking speed on Google Maps?
|
| 33 |
+
- source_sentence: DId something exist before the Big Bang?
|
|
|
|
|
|
|
|
|
|
| 34 |
sentences:
|
| 35 |
+
- How can I improve my memory problem?
|
| 36 |
+
- Where can I buy Fairy Tail Manga?
|
| 37 |
+
- Is there a scientific name for what existed before the Big Bang?
|
|
|
|
| 38 |
pipeline_tag: sentence-similarity
|
| 39 |
library_name: sentence-transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
---
|
| 41 |
|
| 42 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
|
|
| 85 |
from sentence_transformers import SentenceTransformer
|
| 86 |
|
| 87 |
# Download from the 🤗 Hub
|
| 88 |
+
model = SentenceTransformer("sentence_transformers_model_id")
|
| 89 |
# Run inference
|
| 90 |
sentences = [
|
| 91 |
+
'DId something exist before the Big Bang?',
|
| 92 |
+
'Is there a scientific name for what existed before the Big Bang?',
|
| 93 |
+
'Where can I buy Fairy Tail Manga?',
|
| 94 |
]
|
| 95 |
embeddings = model.encode(sentences)
|
| 96 |
print(embeddings.shape)
|
|
|
|
| 99 |
# Get the similarity scores for the embeddings
|
| 100 |
similarities = model.similarity(embeddings, embeddings)
|
| 101 |
print(similarities)
|
| 102 |
+
# tensor([[ 1.0000, 0.7596, -0.0398],
|
| 103 |
+
# [ 0.7596, 1.0000, -0.0308],
|
| 104 |
+
# [-0.0398, -0.0308, 1.0000]])
|
| 105 |
```
|
| 106 |
|
| 107 |
<!--
|
|
|
|
| 128 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 129 |
-->
|
| 130 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
<!--
|
| 132 |
## Bias, Risks and Limitations
|
| 133 |
|
|
|
|
| 146 |
|
| 147 |
#### Unnamed Dataset
|
| 148 |
|
| 149 |
+
* Size: 100,000 training samples
|
| 150 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 151 |
* Approximate statistics based on the first 1000 samples:
|
| 152 |
+
| | sentence_0 | sentence_1 | sentence_2 |
|
| 153 |
+
|:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
|
| 154 |
+
| type | string | string | string |
|
| 155 |
+
| details | <ul><li>min: 3 tokens</li><li>mean: 15.53 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 15.5 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.87 tokens</li><li>max: 128 tokens</li></ul> |
|
| 156 |
* Samples:
|
| 157 |
+
| sentence_0 | sentence_1 | sentence_2 |
|
| 158 |
+
|:----------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------|:-----------------------------------------------------------------------|
|
| 159 |
+
| <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>How much is the airport tax in bogota?</code> |
|
| 160 |
+
| <code>Which concept is more important: good planning or hard work?</code> | <code>Which concept is more important: good planning or hard work?</code> | <code>What is important in life: luck or hard work?</code> |
|
| 161 |
+
| <code>What is the most efficient way to make money?</code> | <code>How can I make my money make money?</code> | <code>What can one learn about Quantum Mechanics in 10 minutes?</code> |
|
| 162 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 163 |
```json
|
| 164 |
{
|
|
|
|
| 171 |
### Training Hyperparameters
|
| 172 |
#### Non-Default Hyperparameters
|
| 173 |
|
| 174 |
+
- `per_device_train_batch_size`: 64
|
| 175 |
+
- `per_device_eval_batch_size`: 64
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 176 |
- `fp16`: True
|
| 177 |
+
- `multi_dataset_batch_sampler`: round_robin
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 178 |
|
| 179 |
#### All Hyperparameters
|
| 180 |
<details><summary>Click to expand</summary>
|
| 181 |
|
| 182 |
- `overwrite_output_dir`: False
|
| 183 |
- `do_predict`: False
|
| 184 |
+
- `eval_strategy`: no
|
| 185 |
- `prediction_loss_only`: True
|
| 186 |
+
- `per_device_train_batch_size`: 64
|
| 187 |
+
- `per_device_eval_batch_size`: 64
|
| 188 |
- `per_gpu_train_batch_size`: None
|
| 189 |
- `per_gpu_eval_batch_size`: None
|
| 190 |
- `gradient_accumulation_steps`: 1
|
| 191 |
- `eval_accumulation_steps`: None
|
| 192 |
- `torch_empty_cache_steps`: None
|
| 193 |
+
- `learning_rate`: 5e-05
|
| 194 |
+
- `weight_decay`: 0.0
|
| 195 |
- `adam_beta1`: 0.9
|
| 196 |
- `adam_beta2`: 0.999
|
| 197 |
- `adam_epsilon`: 1e-08
|
| 198 |
+
- `max_grad_norm`: 1
|
| 199 |
+
- `num_train_epochs`: 3
|
| 200 |
+
- `max_steps`: -1
|
| 201 |
- `lr_scheduler_type`: linear
|
| 202 |
- `lr_scheduler_kwargs`: {}
|
| 203 |
+
- `warmup_ratio`: 0.0
|
| 204 |
- `warmup_steps`: 0
|
| 205 |
- `log_level`: passive
|
| 206 |
- `log_level_replica`: warning
|
|
|
|
| 228 |
- `tpu_num_cores`: None
|
| 229 |
- `tpu_metrics_debug`: False
|
| 230 |
- `debug`: []
|
| 231 |
+
- `dataloader_drop_last`: False
|
| 232 |
+
- `dataloader_num_workers`: 0
|
| 233 |
+
- `dataloader_prefetch_factor`: None
|
| 234 |
- `past_index`: -1
|
| 235 |
- `disable_tqdm`: False
|
| 236 |
- `remove_unused_columns`: True
|
| 237 |
- `label_names`: None
|
| 238 |
+
- `load_best_model_at_end`: False
|
| 239 |
- `ignore_data_skip`: False
|
| 240 |
- `fsdp`: []
|
| 241 |
- `fsdp_min_num_params`: 0
|
|
|
|
| 245 |
- `parallelism_config`: None
|
| 246 |
- `deepspeed`: None
|
| 247 |
- `label_smoothing_factor`: 0.0
|
| 248 |
+
- `optim`: adamw_torch_fused
|
| 249 |
- `optim_args`: None
|
| 250 |
- `adafactor`: False
|
| 251 |
- `group_by_length`: False
|
| 252 |
- `length_column_name`: length
|
| 253 |
- `project`: huggingface
|
| 254 |
- `trackio_space_id`: trackio
|
| 255 |
+
- `ddp_find_unused_parameters`: None
|
| 256 |
- `ddp_bucket_cap_mb`: None
|
| 257 |
- `ddp_broadcast_buffers`: False
|
| 258 |
- `dataloader_pin_memory`: True
|
| 259 |
- `dataloader_persistent_workers`: False
|
| 260 |
- `skip_memory_metrics`: True
|
| 261 |
- `use_legacy_prediction_loop`: False
|
| 262 |
+
- `push_to_hub`: False
|
| 263 |
- `resume_from_checkpoint`: None
|
| 264 |
+
- `hub_model_id`: None
|
| 265 |
- `hub_strategy`: every_save
|
| 266 |
- `hub_private_repo`: None
|
| 267 |
- `hub_always_push`: False
|
|
|
|
| 288 |
- `neftune_noise_alpha`: None
|
| 289 |
- `optim_target_modules`: None
|
| 290 |
- `batch_eval_metrics`: False
|
| 291 |
+
- `eval_on_start`: False
|
| 292 |
- `use_liger_kernel`: False
|
| 293 |
- `liger_kernel_config`: None
|
| 294 |
- `eval_use_gather_object`: False
|
| 295 |
- `average_tokens_across_devices`: True
|
| 296 |
- `prompts`: None
|
| 297 |
- `batch_sampler`: batch_sampler
|
| 298 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 299 |
- `router_mapping`: {}
|
| 300 |
- `learning_rate_mapping`: {}
|
| 301 |
|
| 302 |
</details>
|
| 303 |
|
| 304 |
### Training Logs
|
| 305 |
+
| Epoch | Step | Training Loss |
|
| 306 |
+
|:------:|:----:|:-------------:|
|
| 307 |
+
| 0.3199 | 500 | 0.2284 |
|
| 308 |
+
| 0.6398 | 1000 | 0.0571 |
|
| 309 |
+
| 0.9597 | 1500 | 0.0486 |
|
| 310 |
+
| 1.2796 | 2000 | 0.0378 |
|
| 311 |
+
| 1.5995 | 2500 | 0.0367 |
|
| 312 |
+
| 1.9194 | 3000 | 0.0338 |
|
| 313 |
+
| 2.2393 | 3500 | 0.0327 |
|
| 314 |
+
| 2.5592 | 4000 | 0.0285 |
|
| 315 |
+
| 2.8791 | 4500 | 0.0285 |
|
| 316 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 317 |
|
| 318 |
### Framework Versions
|
| 319 |
- Python: 3.10.18
|
eval/Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -180,3 +180,144 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
|
|
| 180 |
9.815078236130867,13800,0.8296,0.904725,0.932475,0.8296,0.8296,0.301575,0.904725,0.18649500000000002,0.932475,0.8296,0.8693120833333294,0.8736897916666616,0.8957134946126672,0.8757585053341062
|
| 181 |
9.88620199146515,13900,0.829475,0.904825,0.932425,0.829475,0.829475,0.3016083333333333,0.904825,0.186485,0.932425,0.829475,0.8692816666666627,0.8736694444444394,0.8957054405004242,0.875738183071688
|
| 182 |
9.95732574679943,14000,0.829675,0.9048,0.93245,0.829675,0.829675,0.3016,0.9048,0.18649000000000004,0.93245,0.829675,0.8693824999999958,0.8737714285714238,0.8957919450437679,0.8758361833602419
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 180 |
9.815078236130867,13800,0.8296,0.904725,0.932475,0.8296,0.8296,0.301575,0.904725,0.18649500000000002,0.932475,0.8296,0.8693120833333294,0.8736897916666616,0.8957134946126672,0.8757585053341062
|
| 181 |
9.88620199146515,13900,0.829475,0.904825,0.932425,0.829475,0.829475,0.3016083333333333,0.904825,0.186485,0.932425,0.829475,0.8692816666666627,0.8736694444444394,0.8957054405004242,0.875738183071688
|
| 182 |
9.95732574679943,14000,0.829675,0.9048,0.93245,0.829675,0.829675,0.3016,0.9048,0.18649000000000004,0.93245,0.829675,0.8693824999999958,0.8737714285714238,0.8957919450437679,0.8758361833602419
|
| 183 |
+
0,0,0.754525,0.807575,0.830125,0.754525,0.754525,0.26919166666666666,0.807575,0.166025,0.830125,0.754525,0.7830570833333289,0.787073373015871,0.8044871056225573,0.7902066872230316
|
| 184 |
+
0.07112375533428165,100,0.788175,0.8515,0.8754,0.788175,0.788175,0.2838333333333333,0.8515,0.17507999999999999,0.8754,0.788175,0.8216874999999941,0.8261519940476153,0.8459150549027301,0.829169181087181
|
| 185 |
+
0.1422475106685633,200,0.8078,0.875,0.898975,0.8078,0.8078,0.29166666666666663,0.875,0.17979500000000004,0.898975,0.8078,0.8431537499999955,0.847442658730155,0.8675041742945899,0.85001676579445
|
| 186 |
+
0.21337126600284495,300,0.8139,0.882075,0.9063,0.8139,0.8139,0.294025,0.882075,0.18126,0.9063,0.8139,0.8497174999999954,0.8537709821428534,0.8737150227497443,0.8563985748208767
|
| 187 |
+
0.2844950213371266,400,0.816375,0.884475,0.9097,0.816375,0.816375,0.294825,0.884475,0.18194000000000002,0.9097,0.816375,0.8523545833333286,0.85645184523809,0.8766246185595526,0.8589986601030838
|
| 188 |
+
0.35561877667140823,500,0.817925,0.8864,0.911825,0.817925,0.817925,0.2954666666666666,0.8864,0.18236500000000005,0.911825,0.817925,0.8540712499999951,0.8581809722222173,0.8784583584690699,0.8607180206672261
|
| 189 |
+
0.4267425320056899,600,0.81905,0.887875,0.9133,0.81905,0.81905,0.29595833333333327,0.887875,0.18266000000000004,0.9133,0.81905,0.8553095833333285,0.859498859126979,0.8799370224897283,0.861996070829766
|
| 190 |
+
0.49786628733997157,700,0.820825,0.8887,0.91495,0.820825,0.820825,0.29623333333333335,0.8887,0.18299000000000004,0.91495,0.820825,0.8568612499999957,0.8609963591269792,0.8813434328983816,0.8635262113099532
|
| 191 |
+
0.5689900426742532,800,0.8212,0.889525,0.9163,0.8212,0.8212,0.2965083333333333,0.889525,0.18325999999999998,0.9163,0.8212,0.8576537499999951,0.8617598908730091,0.882234040479893,0.8642600344938268
|
| 192 |
+
0.6401137980085349,900,0.821825,0.890425,0.91765,0.821825,0.821825,0.2968083333333333,0.890425,0.18353,0.91765,0.821825,0.8585424999999958,0.8626341865079316,0.8831597899495076,0.8651357816461163
|
| 193 |
+
0.7112375533428165,1000,0.821875,0.891225,0.917725,0.821875,0.821875,0.297075,0.891225,0.183545,0.917725,0.821875,0.8586737499999942,0.8628899305555491,0.8836079830828143,0.8653478838731942
|
| 194 |
+
0.7823613086770982,1100,0.8221,0.891925,0.91895,0.8221,0.8221,0.2973083333333333,0.891925,0.18379000000000004,0.91895,0.8221,0.8592887499999946,0.8633686904761843,0.8840734765399635,0.8658691525588665
|
| 195 |
+
0.8534850640113798,1200,0.822925,0.892725,0.919175,0.822925,0.822925,0.297575,0.892725,0.18383500000000003,0.919175,0.822925,0.8599345833333281,0.8640485912698367,0.8846586524909154,0.8665777843003274
|
| 196 |
+
0.9246088193456614,1300,0.82345,0.8935,0.92055,0.82345,0.82345,0.2978333333333333,0.8935,0.18411,0.92055,0.82345,0.8607987499999945,0.8649032638888835,0.8856106027310329,0.867403577329114
|
| 197 |
+
0.9957325746799431,1400,0.82325,0.8933,0.921125,0.82325,0.82325,0.2977666666666666,0.8933,0.18422500000000003,0.921125,0.82325,0.8608366666666618,0.8649197718253918,0.885785240388492,0.8673886311195982
|
| 198 |
+
1.0668563300142249,1500,0.8239,0.894675,0.9212,0.8239,0.8239,0.298225,0.894675,0.18424000000000004,0.9212,0.8239,0.861378749999995,0.8656084027777738,0.886533866338354,0.868050543099101
|
| 199 |
+
1.1379800853485065,1600,0.8239,0.89485,0.9205,0.8239,0.8239,0.2982833333333333,0.89485,0.18409999999999999,0.9205,0.8239,0.8612483333333286,0.8656391964285669,0.8866256611000325,0.8680736081907602
|
| 200 |
+
1.209103840682788,1700,0.82405,0.89515,0.92115,0.82405,0.82405,0.2983833333333333,0.89515,0.18423,0.92115,0.82405,0.8615316666666606,0.8659026388888836,0.8869176027034992,0.8683301973092413
|
| 201 |
+
1.2802275960170697,1800,0.824175,0.896225,0.922375,0.824175,0.824175,0.2987416666666666,0.896225,0.18447500000000003,0.922375,0.824175,0.8621304166666607,0.8663353273809463,0.8873448262106999,0.8687668641747599
|
| 202 |
+
1.3513513513513513,1900,0.825525,0.896575,0.92315,0.825525,0.825525,0.2988583333333333,0.896575,0.18463000000000002,0.92315,0.825525,0.8630337499999942,0.8672058829365027,0.8881349504555975,0.8696378148874042
|
| 203 |
+
1.422475106685633,2000,0.824875,0.89615,0.922225,0.824875,0.824875,0.29871666666666663,0.89615,0.184445,0.922225,0.824875,0.8624995833333283,0.8668498412698372,0.8879244763281516,0.8692683038849478
|
| 204 |
+
1.4935988620199145,2100,0.8251,0.89675,0.922925,0.8251,0.8251,0.29891666666666666,0.89675,0.18458500000000003,0.922925,0.8251,0.8628412499999946,0.8671035813492023,0.8881754016915514,0.8695244951133804
|
| 205 |
+
1.5647226173541964,2200,0.824925,0.8969,0.92355,0.824925,0.824925,0.2989666666666666,0.8969,0.18470999999999999,0.92355,0.824925,0.8630016666666616,0.8671947222222179,0.8882835417851939,0.8696206929447456
|
| 206 |
+
1.635846372688478,2300,0.82495,0.89735,0.923825,0.82495,0.82495,0.29911666666666664,0.89735,0.18476500000000004,0.923825,0.82495,0.8631245833333274,0.8673548611111046,0.888537638953492,0.8697591451729704
|
| 207 |
+
1.7069701280227596,2400,0.826225,0.8979,0.924375,0.826225,0.826225,0.2992999999999999,0.8979,0.184875,0.924375,0.826225,0.8640779166666614,0.868321726190472,0.8894215247719068,0.8707139327932544
|
| 208 |
+
1.7780938833570412,2500,0.82545,0.898275,0.924175,0.82545,0.82545,0.29942499999999994,0.898275,0.18483500000000003,0.924175,0.82545,0.8636241666666614,0.8679304464285673,0.8891594573590612,0.8703335724958975
|
| 209 |
+
1.8492176386913228,2600,0.826175,0.898175,0.9247,0.826175,0.826175,0.29939166666666667,0.898175,0.18494000000000002,0.9247,0.826175,0.8641804166666623,0.8684414682539631,0.8896458867227734,0.870847238427157
|
| 210 |
+
1.9203413940256047,2700,0.825875,0.897825,0.924175,0.825875,0.825875,0.299275,0.897825,0.18483500000000003,0.924175,0.825875,0.8638574999999944,0.868217509920631,0.8894948799781416,0.8706327376670759
|
| 211 |
+
1.991465149359886,2800,0.826225,0.8983,0.924475,0.826225,0.826225,0.29943333333333333,0.8983,0.18489500000000003,0.924475,0.826225,0.8641595833333282,0.8685288888888845,0.889811094671074,0.8709200811674378
|
| 212 |
+
2.062588904694168,2900,0.826575,0.89885,0.92485,0.826575,0.826575,0.2996166666666666,0.89885,0.18497000000000002,0.92485,0.826575,0.8645774999999943,0.8689782341269785,0.8903225599754984,0.8713206279473252
|
| 213 |
+
2.1337126600284497,3000,0.8266,0.89905,0.925175,0.8266,0.8266,0.2996833333333333,0.89905,0.18503500000000003,0.925175,0.8266,0.8647558333333282,0.869107787698408,0.8904131932586173,0.8714774361878306
|
| 214 |
+
2.204836415362731,3100,0.826675,0.8997,0.925225,0.826675,0.826675,0.2999,0.8997,0.18504500000000002,0.925225,0.826675,0.8648191666666607,0.8691860912698357,0.8904989946578933,0.871553217284866
|
| 215 |
+
2.275960170697013,3200,0.826725,0.89955,0.9256,0.826725,0.826725,0.2998499999999999,0.89955,0.18511999999999998,0.9256,0.826725,0.8649791666666601,0.8693465277777699,0.8907582295982882,0.8716640608002595
|
| 216 |
+
2.3470839260312943,3300,0.826575,0.89985,0.925625,0.826575,0.826575,0.29994999999999994,0.89985,0.185125,0.925625,0.826575,0.8650091666666611,0.869368204365073,0.8907657483756646,0.8717004467651088
|
| 217 |
+
2.418207681365576,3400,0.826575,0.8997,0.925775,0.826575,0.826575,0.2998999999999999,0.8997,0.18515500000000001,0.925775,0.826575,0.8649995833333284,0.8693408432539614,0.8907417130181404,0.8716965684052127
|
| 218 |
+
2.4893314366998576,3500,0.82665,0.899475,0.9261,0.82665,0.82665,0.29982499999999995,0.899475,0.18522000000000002,0.9261,0.82665,0.8649920833333282,0.8692922817460256,0.8907358632847491,0.871649860200826
|
| 219 |
+
2.5604551920341394,3600,0.826875,0.9003,0.92605,0.826875,0.826875,0.3001,0.9003,0.18521000000000004,0.92605,0.826875,0.8652491666666605,0.8695896031745965,0.8909979289659293,0.8719302926120457
|
| 220 |
+
2.6315789473684212,3700,0.827025,0.900275,0.926075,0.827025,0.827025,0.30009166666666665,0.900275,0.18521500000000002,0.926075,0.827025,0.865409583333329,0.8697828075396781,0.8911913854619483,0.8721236055235071
|
| 221 |
+
2.7027027027027026,3800,0.827025,0.90095,0.92645,0.827025,0.827025,0.3003166666666666,0.90095,0.18529000000000004,0.92645,0.827025,0.8655795833333275,0.8698699603174542,0.8912290647496708,0.8722296272752561
|
| 222 |
+
2.7738264580369845,3900,0.826775,0.900975,0.926525,0.826775,0.826775,0.3003249999999999,0.900975,0.18530500000000003,0.926525,0.826775,0.8655420833333285,0.8699055654761841,0.8914045716439474,0.8722080443535084
|
| 223 |
+
2.844950213371266,4000,0.8272,0.90065,0.926825,0.8272,0.8272,0.30021666666666663,0.90065,0.18536500000000003,0.926825,0.8272,0.8658591666666623,0.8701858333333292,0.8916375973389772,0.8724953320587089
|
| 224 |
+
2.9160739687055477,4100,0.8273,0.90065,0.92755,0.8273,0.8273,0.30021666666666663,0.90065,0.18551,0.92755,0.8273,0.8660504166666627,0.8703939980158677,0.8919791384990317,0.8726592726636881
|
| 225 |
+
2.987197724039829,4200,0.827725,0.90045,0.92775,0.827725,0.827725,0.30015,0.90045,0.18555,0.92775,0.827725,0.8662579166666624,0.8705597023809468,0.892066241132213,0.8728541618998119
|
| 226 |
+
3.058321479374111,4300,0.827475,0.901175,0.928025,0.827475,0.827475,0.3003916666666666,0.901175,0.18560500000000002,0.928025,0.827475,0.866314583333329,0.8705805654761841,0.8921197780063829,0.8728613054421528
|
| 227 |
+
3.1294452347083928,4400,0.8277,0.901175,0.92825,0.8277,0.8277,0.3003916666666666,0.901175,0.18565,0.92825,0.8277,0.8665124999999955,0.8708128670634864,0.8924373846941944,0.8730523461044413
|
| 228 |
+
3.200568990042674,4500,0.827575,0.901225,0.928375,0.827575,0.827575,0.3004083333333333,0.901225,0.18567500000000003,0.928375,0.827575,0.8664383333333281,0.8708214682539619,0.8925953907373283,0.8730156219725579
|
| 229 |
+
3.271692745376956,4600,0.8279,0.901575,0.928375,0.8279,0.8279,0.300525,0.901575,0.185675,0.928375,0.8279,0.866585833333329,0.8709510019841202,0.8926942737918313,0.8731458638145055
|
| 230 |
+
3.3428165007112374,4700,0.827575,0.9014,0.92795,0.827575,0.827575,0.3004666666666666,0.9014,0.18559000000000006,0.92795,0.827575,0.8664058333333281,0.8708573710317392,0.8926411564059361,0.8730650536407063
|
| 231 |
+
3.413940256045519,4800,0.8276,0.901575,0.92875,0.8276,0.8276,0.300525,0.901575,0.18575,0.92875,0.8276,0.8665841666666624,0.8708900198412636,0.8926305600650463,0.873122633964442
|
| 232 |
+
3.485064011379801,4900,0.82815,0.9011,0.92885,0.82815,0.82815,0.3003666666666666,0.9011,0.18577,0.92885,0.82815,0.8668041666666617,0.8711553373015805,0.8929320907320707,0.8733508626597645
|
| 233 |
+
3.5561877667140824,5000,0.827675,0.901125,0.92895,0.827675,0.827675,0.300375,0.901125,0.18579,0.92895,0.827675,0.8665837499999955,0.8708607738095179,0.8925810943314898,0.8731088176524069
|
| 234 |
+
3.6273115220483643,5100,0.8279,0.9018,0.9286,0.8279,0.8279,0.3006,0.9018,0.18572000000000002,0.9286,0.8279,0.8668145833333285,0.8711948015872957,0.8929008508637697,0.8734305622412212
|
| 235 |
+
3.6984352773826457,5200,0.8277,0.90215,0.9287,0.8277,0.8277,0.30071666666666663,0.90215,0.18574000000000004,0.9287,0.8277,0.8667604166666624,0.8711313293650733,0.8928740678326429,0.8733567486761654
|
| 236 |
+
3.7695590327169275,5300,0.82845,0.9025,0.929125,0.82845,0.82845,0.3008333333333333,0.9025,0.18582500000000002,0.929125,0.82845,0.8674124999999964,0.8718316964285665,0.8935972664353634,0.8740016594916253
|
| 237 |
+
3.8406827880512093,5400,0.827975,0.90155,0.92865,0.827975,0.827975,0.3005166666666666,0.90155,0.18573000000000003,0.92865,0.827975,0.8668745833333286,0.8712864880952323,0.8930123747026218,0.8735193873832935
|
| 238 |
+
3.9118065433854907,5500,0.8277,0.90215,0.9289,0.8277,0.8277,0.30071666666666663,0.90215,0.18578,0.9289,0.8277,0.8668504166666622,0.8712548115079313,0.8930900566464374,0.873443092121
|
| 239 |
+
3.9829302987197726,5600,0.8274,0.90235,0.928875,0.8274,0.8274,0.3007833333333333,0.90235,0.18577500000000005,0.928875,0.8274,0.8667458333333284,0.8712413492063439,0.8932113985600477,0.8733958591837214
|
| 240 |
+
4.054054054054054,5700,0.8284,0.902525,0.92915,0.8284,0.8284,0.3008416666666666,0.902525,0.18583000000000002,0.92915,0.8284,0.8673966666666616,0.8718133630952317,0.8935767431048043,0.8740015491385773
|
| 241 |
+
4.125177809388336,5800,0.828425,0.902825,0.9296,0.828425,0.828425,0.3009416666666666,0.902825,0.18592000000000003,0.9296,0.828425,0.8675812499999948,0.871918124999994,0.893666432010205,0.8741087426233196
|
| 242 |
+
4.196301564722617,5900,0.828525,0.9022,0.92915,0.828525,0.828525,0.3007333333333333,0.9022,0.18583000000000002,0.92915,0.828525,0.8673833333333288,0.8718712202380895,0.8937141847404675,0.8740192110160396
|
| 243 |
+
4.2674253200568995,6000,0.828875,0.90235,0.92975,0.828875,0.828875,0.30078333333333324,0.90235,0.18595000000000003,0.92975,0.828875,0.8678129166666622,0.8721811706349157,0.8938989321601007,0.8743745063266631
|
| 244 |
+
4.338549075391181,6100,0.82845,0.9026,0.9291,0.82845,0.82845,0.3008666666666666,0.9026,0.18582,0.9291,0.82845,0.8673166666666617,0.8717749999999929,0.8936243154363203,0.8739455546849367
|
| 245 |
+
4.409672830725462,6200,0.8284,0.903075,0.9296,0.8284,0.8284,0.30102499999999993,0.903075,0.18592000000000003,0.9296,0.8284,0.8676320833333286,0.8720804365079297,0.8939257947166456,0.8742395675408973
|
| 246 |
+
4.480796586059744,6300,0.828225,0.90275,0.929325,0.828225,0.828225,0.3009166666666666,0.90275,0.18586500000000003,0.929325,0.828225,0.8674999999999947,0.8719905158730092,0.8938322598813815,0.8741705169023509
|
| 247 |
+
4.551920341394026,6400,0.8285,0.90335,0.92975,0.8285,0.8285,0.3011166666666666,0.90335,0.18595000000000003,0.92975,0.8285,0.8678516666666617,0.8723455853174541,0.8942491439779721,0.874474011966291
|
| 248 |
+
4.623044096728307,6500,0.829175,0.902875,0.929875,0.829175,0.829175,0.30095833333333327,0.902875,0.18597500000000006,0.929875,0.829175,0.8680937499999958,0.8725534623015822,0.894368923703291,0.8746872189322042
|
| 249 |
+
4.694167852062589,6600,0.82905,0.903075,0.93015,0.82905,0.82905,0.301025,0.903075,0.18603000000000003,0.93015,0.82905,0.8681745833333278,0.8725917361111044,0.8944163915781091,0.8747310389447597
|
| 250 |
+
4.76529160739687,6700,0.828925,0.90325,0.929775,0.828925,0.828925,0.3010833333333333,0.90325,0.185955,0.929775,0.828925,0.8679954166666616,0.8724790773809465,0.8943439212494719,0.8746128881249259
|
| 251 |
+
4.836415362731152,6800,0.828575,0.9031,0.9304,0.828575,0.828575,0.3010333333333333,0.9031,0.18608000000000002,0.9304,0.828575,0.8679383333333275,0.8723482936507868,0.8942892449456445,0.8744742271170953
|
| 252 |
+
4.907539118065434,6900,0.82885,0.9029,0.9307,0.82885,0.82885,0.3009666666666666,0.9029,0.18614,0.9307,0.82885,0.8680495833333282,0.8724709027777715,0.8944710513562503,0.8745647676257946
|
| 253 |
+
4.978662873399715,7000,0.8283,0.9031,0.93085,0.8283,0.8283,0.3010333333333333,0.9031,0.18617,0.93085,0.8283,0.8678516666666619,0.8722358730158676,0.8943000182461698,0.8743369199623621
|
| 254 |
+
5.049786628733997,7100,0.8279,0.903125,0.9307,0.8279,0.8279,0.3010416666666666,0.903125,0.18614000000000003,0.9307,0.8279,0.8676270833333279,0.8719915376984065,0.8940120699532174,0.8741380897523008
|
| 255 |
+
5.120910384068279,7200,0.828275,0.90295,0.930025,0.828275,0.828275,0.3009833333333333,0.90295,0.18600500000000003,0.930025,0.828275,0.8676429166666617,0.8721463293650742,0.8941977596786901,0.8742604553436768
|
| 256 |
+
5.19203413940256,7300,0.828425,0.903025,0.930275,0.828425,0.828425,0.3010083333333333,0.903025,0.18605500000000003,0.930275,0.828425,0.8678812499999949,0.872354652777772,0.8943000793022067,0.8745114126364067
|
| 257 |
+
5.2631578947368425,7400,0.828225,0.9028,0.9301,0.828225,0.828225,0.30093333333333333,0.9028,0.18602000000000002,0.9301,0.828225,0.8675679166666612,0.872012182539677,0.8939575105681979,0.8741988719154433
|
| 258 |
+
5.334281650071124,7500,0.82785,0.903125,0.930575,0.82785,0.82785,0.30104166666666665,0.903125,0.186115,0.930575,0.82785,0.8676316666666626,0.8721601488095195,0.8943613633651454,0.874238288502371
|
| 259 |
+
5.405405405405405,7600,0.828475,0.902975,0.93095,0.828475,0.828475,0.3009916666666666,0.902975,0.18619000000000002,0.93095,0.828475,0.8680033333333292,0.8724383134920588,0.8944984330360523,0.8745428567045579
|
| 260 |
+
5.476529160739687,7700,0.82855,0.90335,0.9307,0.82855,0.82855,0.30111666666666664,0.90335,0.18614000000000003,0.9307,0.82855,0.8680608333333294,0.8725198611111069,0.8945047825331323,0.8746575985799696
|
| 261 |
+
5.547652916073969,7800,0.828875,0.903275,0.93085,0.828875,0.828875,0.3010916666666666,0.903275,0.18617000000000006,0.93085,0.828875,0.8682458333333294,0.872676269841266,0.8946389217721571,0.8747955260908377
|
| 262 |
+
5.61877667140825,7900,0.82905,0.903375,0.93105,0.82905,0.82905,0.301125,0.903375,0.18621,0.93105,0.82905,0.8683870833333295,0.8727912202380916,0.8947269456300353,0.8749224068854644
|
| 263 |
+
5.689900426742532,8000,0.82915,0.903325,0.930675,0.82915,0.82915,0.3011083333333333,0.903325,0.18613500000000002,0.930675,0.82915,0.8684083333333289,0.8729126091269789,0.8948820398516149,0.8750130251353759
|
| 264 |
+
5.761024182076814,8100,0.829075,0.904075,0.9312,0.829075,0.829075,0.3013583333333333,0.904075,0.18624000000000004,0.9312,0.829075,0.8685508333333278,0.873021924603169,0.8950310737741233,0.8751092075074616
|
| 265 |
+
5.832147937411095,8200,0.828825,0.90365,0.9305,0.828825,0.828825,0.30121666666666663,0.90365,0.18610000000000002,0.9305,0.828825,0.8682554166666617,0.8728638095238044,0.8949489700564521,0.8749277099562706
|
| 266 |
+
5.903271692745377,8300,0.82895,0.903725,0.931225,0.82895,0.82895,0.30124166666666663,0.903725,0.18624500000000002,0.931225,0.82895,0.8684424999999952,0.8729208234126938,0.8949867471707442,0.8749962923229312
|
| 267 |
+
5.974395448079658,8400,0.82885,0.903675,0.931025,0.82885,0.82885,0.30122499999999997,0.903675,0.18620499999999998,0.931025,0.82885,0.8683604166666622,0.8728612103174557,0.8949169009398232,0.8749454784815862
|
| 268 |
+
6.0455192034139404,8500,0.82895,0.903775,0.931475,0.82895,0.82895,0.3012583333333333,0.903775,0.18629500000000002,0.931475,0.82895,0.8685329166666627,0.8729780257936461,0.8950478966468592,0.8750537538993218
|
| 269 |
+
6.116642958748222,8600,0.82905,0.90415,0.93205,0.82905,0.82905,0.3013833333333333,0.90415,0.18641000000000005,0.93205,0.82905,0.8687849999999957,0.8731482539682496,0.8951754441957878,0.8752259309496394
|
| 270 |
+
6.187766714082503,8700,0.828825,0.9041,0.93175,0.828825,0.828825,0.30136666666666667,0.9041,0.18635,0.93175,0.828825,0.8685724999999961,0.872876924603171,0.8947828539779766,0.8750292561714247
|
| 271 |
+
6.2588904694167855,8800,0.829075,0.903925,0.931425,0.829075,0.829075,0.3013083333333333,0.903925,0.186285,0.931425,0.829075,0.868536666666662,0.8729551984126942,0.8949195971526226,0.8750767734715837
|
| 272 |
+
6.330014224751067,8900,0.828975,0.90415,0.931325,0.828975,0.828975,0.3013833333333333,0.90415,0.18626500000000004,0.931325,0.828975,0.8684991666666616,0.8729706845238042,0.8949951561857655,0.8750641724121062
|
| 273 |
+
6.401137980085348,9000,0.829075,0.903975,0.93165,0.829075,0.829075,0.301325,0.903975,0.18633000000000005,0.93165,0.829075,0.8685995833333287,0.872990962301583,0.8949819996202747,0.8750914105522226
|
| 274 |
+
6.472261735419631,9100,0.829225,0.903925,0.932075,0.829225,0.829225,0.3013083333333333,0.903925,0.18641500000000003,0.932075,0.829225,0.868800416666662,0.8731110317460269,0.895053660038409,0.8752295993536805
|
| 275 |
+
6.543385490753912,9200,0.82855,0.9041,0.931375,0.82855,0.82855,0.30136666666666667,0.9041,0.18627500000000002,0.931375,0.82855,0.8682741666666616,0.8726703174603115,0.8946713862569544,0.8748099752974711
|
| 276 |
+
6.614509246088193,9300,0.828775,0.904225,0.9317,0.828775,0.828775,0.3014083333333333,0.904225,0.18634,0.9317,0.828775,0.8684645833333281,0.8728734226190421,0.8949274319558564,0.8749753112232658
|
| 277 |
+
6.685633001422475,9400,0.82835,0.90395,0.932025,0.82835,0.82835,0.3013166666666666,0.90395,0.18640500000000002,0.932025,0.82835,0.8683237499999945,0.8726935218253904,0.8948281974118523,0.8747854079129327
|
| 278 |
+
6.756756756756757,9500,0.828475,0.904075,0.932075,0.828475,0.828475,0.3013583333333333,0.904075,0.18641500000000003,0.932075,0.828475,0.8684204166666614,0.8727985515872957,0.8949274062023101,0.8748895020570487
|
| 279 |
+
6.827880512091038,9600,0.8288,0.904225,0.93185,0.8288,0.8288,0.3014083333333333,0.904225,0.18637000000000004,0.93185,0.8288,0.8685758333333276,0.873004513888882,0.8951136713393641,0.8750801043302465
|
| 280 |
+
6.89900426742532,9700,0.828825,0.904375,0.9318,0.828825,0.828825,0.30145833333333333,0.904375,0.18636000000000003,0.9318,0.828825,0.8686991666666614,0.8731213591269789,0.8951596018402795,0.8752201609740874
|
| 281 |
+
6.970128022759602,9800,0.829025,0.904525,0.932225,0.829025,0.829025,0.30150833333333327,0.904525,0.18644500000000003,0.932225,0.829025,0.8689174999999952,0.8732630853174554,0.8952619483692564,0.8753690315163684
|
| 282 |
+
7.0412517780938835,9900,0.828875,0.904425,0.93255,0.828875,0.828875,0.301475,0.904425,0.18651000000000004,0.93255,0.828875,0.8688845833333287,0.873162787698407,0.8951766391067307,0.8752720359442444
|
| 283 |
+
7.112375533428165,10000,0.828575,0.904625,0.932375,0.828575,0.828575,0.30154166666666665,0.904625,0.186475,0.932375,0.828575,0.8687079166666616,0.8730274900793593,0.8950797790185748,0.8751374732802933
|
| 284 |
+
7.183499288762446,10100,0.8286,0.905075,0.932175,0.8286,0.8286,0.3016916666666666,0.905075,0.18643500000000005,0.932175,0.8286,0.8687141666666612,0.8730440476190419,0.8950594465802302,0.8751640577624185
|
| 285 |
+
7.2546230440967285,10200,0.8296,0.904775,0.9323,0.8296,0.8296,0.3015916666666666,0.904775,0.18646000000000001,0.9323,0.8296,0.8691612499999947,0.873514434523804,0.8954897884983242,0.8756084914792746
|
| 286 |
+
7.32574679943101,10300,0.829075,0.904875,0.93235,0.829075,0.829075,0.301625,0.904875,0.18647000000000002,0.93235,0.829075,0.868976249999995,0.8732967956349148,0.8952704782589539,0.8754151233921735
|
| 287 |
+
7.396870554765291,10400,0.82915,0.905075,0.932425,0.82915,0.82915,0.3016916666666666,0.905075,0.186485,0.932425,0.82915,0.8690495833333287,0.8734093849206295,0.8954447584237889,0.8755008479404327
|
| 288 |
+
7.467994310099574,10500,0.829425,0.904825,0.932525,0.829425,0.829425,0.30160833333333326,0.904825,0.18650500000000003,0.932525,0.829425,0.8691991666666623,0.8734998115079301,0.8954660672180369,0.8756100523433119
|
| 289 |
+
7.539118065433855,10600,0.829125,0.905075,0.932475,0.829125,0.829125,0.3016916666666666,0.905075,0.18649500000000002,0.932475,0.829125,0.8691154166666616,0.8734564682539616,0.8954709669225793,0.8755484892898732
|
| 290 |
+
7.610241820768136,10700,0.82905,0.9048,0.932375,0.82905,0.82905,0.3016,0.9048,0.18647500000000006,0.932375,0.82905,0.8689229166666611,0.8732853769841208,0.8953490062056495,0.875372576803479
|
| 291 |
+
7.681365576102419,10800,0.829025,0.905325,0.932625,0.829025,0.829025,0.30177499999999996,0.905325,0.18652500000000005,0.932625,0.829025,0.869070416666661,0.8734086011904688,0.8954652430846454,0.8754917937310882
|
| 292 |
+
7.7524893314367,10900,0.82875,0.904775,0.932125,0.82875,0.82875,0.3015916666666666,0.904775,0.18642500000000004,0.932125,0.82875,0.868665416666662,0.8730924007936451,0.8952425511491899,0.8751620499741801
|
| 293 |
+
7.823613086770981,11000,0.8295,0.90505,0.932275,0.8295,0.8295,0.3016833333333333,0.90505,0.186455,0.932275,0.8295,0.8692354166666619,0.8736657539682472,0.8957464957352415,0.8757095548251062
|
| 294 |
+
7.894736842105263,11100,0.828925,0.90485,0.932575,0.828925,0.828925,0.3016166666666666,0.90485,0.18651500000000001,0.932575,0.828925,0.8689179166666623,0.8732802380952318,0.8954065650616195,0.8753463124648857
|
| 295 |
+
7.965860597439545,11200,0.829025,0.904825,0.932675,0.829025,0.829025,0.3016083333333333,0.904825,0.186535,0.932675,0.829025,0.8690079166666621,0.8733515476190415,0.8954423259492094,0.875432665821568
|
| 296 |
+
8.036984352773827,11300,0.8292,0.904725,0.93255,0.8292,0.8292,0.3015749999999999,0.904725,0.18650999999999998,0.93255,0.8292,0.869073749999995,0.8734561210317406,0.8955415674350273,0.8755229302308101
|
| 297 |
+
8.108108108108109,11400,0.829325,0.904975,0.93275,0.829325,0.829325,0.3016583333333333,0.904975,0.18655000000000005,0.93275,0.829325,0.8692104166666624,0.8735678571428518,0.8956531906883425,0.8756297665813315
|
| 298 |
+
8.17923186344239,11500,0.82925,0.90495,0.932375,0.82925,0.82925,0.30165,0.90495,0.186475,0.932375,0.82925,0.8690454166666621,0.8734416964285652,0.8955104391817192,0.8755223363162908
|
| 299 |
+
8.250355618776672,11600,0.828875,0.9047,0.93215,0.828875,0.828875,0.3015666666666666,0.9047,0.18643,0.93215,0.828875,0.8687787499999952,0.8732215476190417,0.8953592242644414,0.8752882594576469
|
| 300 |
+
8.321479374110954,11700,0.828775,0.90415,0.9323,0.828775,0.828775,0.3013833333333333,0.90415,0.18646000000000001,0.9323,0.828775,0.8686216666666616,0.8730347222222162,0.8951978059378991,0.8751061946163587
|
| 301 |
+
8.392603129445234,11800,0.828925,0.90455,0.932375,0.828925,0.828925,0.3015166666666666,0.90455,0.18647500000000003,0.932375,0.828925,0.8688104166666617,0.8731834126984072,0.8952621827829491,0.8752830033195543
|
| 302 |
+
8.463726884779517,11900,0.82875,0.90425,0.932075,0.82875,0.82875,0.3014166666666666,0.90425,0.186415,0.932075,0.82875,0.8686170833333284,0.8730646230158685,0.8952314018142266,0.8751380575962121
|
| 303 |
+
8.534850640113799,12000,0.829075,0.9047,0.932425,0.829075,0.829075,0.30156666666666665,0.9047,0.18648500000000004,0.932425,0.829075,0.8689379166666622,0.8733352777777729,0.8954426784378953,0.8754119442285108
|
| 304 |
+
8.60597439544808,12100,0.8291,0.90465,0.932525,0.8291,0.8291,0.30155,0.90465,0.18650500000000003,0.932525,0.8291,0.8689716666666616,0.8733671428571368,0.8955036503963157,0.8754304331343709
|
| 305 |
+
8.677098150782362,12200,0.82895,0.90485,0.932375,0.82895,0.82895,0.3016166666666666,0.90485,0.186475,0.932375,0.82895,0.868958749999995,0.873387123015867,0.8955303325344552,0.8754507768463161
|
| 306 |
+
8.748221906116642,12300,0.829475,0.904725,0.93275,0.829475,0.829475,0.301575,0.904725,0.18655,0.93275,0.829475,0.8692883333333291,0.873639414682535,0.8956905693084858,0.8757137507599527
|
| 307 |
+
8.819345661450924,12400,0.8292,0.9051,0.932725,0.8292,0.8292,0.30169999999999997,0.9051,0.18654500000000002,0.932725,0.8292,0.8692066666666622,0.8735561309523755,0.8956121742695453,0.8756422684505465
|
| 308 |
+
8.890469416785207,12500,0.82945,0.90485,0.93265,0.82945,0.82945,0.3016166666666666,0.90485,0.18653000000000003,0.93265,0.82945,0.8692741666666625,0.8736324999999945,0.8956449964020973,0.8757278374859732
|
| 309 |
+
8.961593172119487,12600,0.8295,0.904875,0.9325,0.8295,0.8295,0.301625,0.904875,0.18650000000000005,0.9325,0.8295,0.8692387499999957,0.8736341964285653,0.8956903497814453,0.8757079138167887
|
| 310 |
+
9.03271692745377,12700,0.829525,0.9047,0.9324,0.829525,0.829525,0.3015666666666666,0.9047,0.18648,0.9324,0.829525,0.8692474999999956,0.8736550694444389,0.8956994527858438,0.8757306836256025
|
| 311 |
+
9.103840682788052,12800,0.829375,0.904375,0.9324,0.829375,0.829375,0.3014583333333333,0.904375,0.18648,0.9324,0.829375,0.8690412499999959,0.8734550595238041,0.8955588176732403,0.875525147366692
|
| 312 |
+
9.174964438122332,12900,0.829125,0.90475,0.9324,0.829125,0.829125,0.3015833333333333,0.90475,0.18648,0.9324,0.829125,0.868967916666662,0.8733762797618992,0.8954799203066371,0.8754609487406843
|
| 313 |
+
9.246088193456615,13000,0.829125,0.90485,0.9324,0.829125,0.829125,0.3016166666666666,0.90485,0.18648,0.9324,0.829125,0.868992083333329,0.8734280654761849,0.8955737083799604,0.8754895469283179
|
| 314 |
+
9.317211948790897,13100,0.8293,0.904775,0.9327,0.8293,0.8293,0.3015916666666666,0.904775,0.18654,0.9327,0.8293,0.8691408333333289,0.873514990079359,0.8956181445893849,0.8755874727685248
|
| 315 |
+
9.388335704125177,13200,0.829325,0.905025,0.932525,0.829325,0.829325,0.301675,0.905025,0.18650500000000003,0.932525,0.829325,0.8691233333333288,0.8735301587301532,0.8956296717143643,0.8756034523955281
|
| 316 |
+
9.45945945945946,13300,0.829325,0.905125,0.932525,0.829325,0.829325,0.3017083333333333,0.905125,0.186505,0.932525,0.829325,0.8691458333333285,0.8735614980158675,0.8956763840293377,0.8756260378815925
|
| 317 |
+
9.530583214793742,13400,0.82915,0.90495,0.9326,0.82915,0.82915,0.30165,0.90495,0.18652000000000002,0.9326,0.82915,0.8690558333333281,0.8734571130952323,0.8955881800163638,0.8755247191912429
|
| 318 |
+
9.601706970128022,13500,0.82925,0.90495,0.932525,0.82925,0.82925,0.3016499999999999,0.90495,0.18650500000000003,0.932525,0.82925,0.8690795833333285,0.8734938392857088,0.8956149222360327,0.8755655926205971
|
| 319 |
+
9.672830725462305,13600,0.8292,0.904925,0.9327,0.8292,0.8292,0.30164166666666664,0.904925,0.18654,0.9327,0.8292,0.8691233333333286,0.8735162896825339,0.895645553478534,0.8755826049757932
|
| 320 |
+
9.743954480796585,13700,0.829325,0.90495,0.932775,0.829325,0.829325,0.30165,0.90495,0.18655500000000003,0.932775,0.829325,0.8692137499999953,0.8735839880952322,0.8956849160109633,0.875656327707471
|
| 321 |
+
9.815078236130867,13800,0.82925,0.905,0.932825,0.82925,0.82925,0.3016666666666666,0.905,0.186565,0.932825,0.82925,0.8691904166666616,0.873566468253962,0.8956952492720289,0.8756302411100991
|
| 322 |
+
9.88620199146515,13900,0.829175,0.905025,0.93285,0.829175,0.829175,0.3016749999999999,0.905025,0.18657,0.93285,0.829175,0.869152916666662,0.8735199007936449,0.895645535201997,0.8755899948508465
|
| 323 |
+
9.95732574679943,14000,0.8292,0.905075,0.932925,0.8292,0.8292,0.3016916666666666,0.905075,0.18658500000000003,0.932925,0.8292,0.869192916666662,0.8735491567460258,0.895673602678825,0.8756171762848609
|
final_metrics.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
{
|
| 2 |
-
"val_cosine_accuracy@1": 0.
|
| 3 |
-
"val_cosine_accuracy@3": 0.
|
| 4 |
-
"val_cosine_accuracy@5": 0.
|
| 5 |
-
"val_cosine_precision@1": 0.
|
| 6 |
-
"val_cosine_precision@3": 0.
|
| 7 |
-
"val_cosine_precision@5": 0.
|
| 8 |
-
"val_cosine_recall@1": 0.
|
| 9 |
-
"val_cosine_recall@3": 0.
|
| 10 |
-
"val_cosine_recall@5": 0.
|
| 11 |
-
"val_cosine_ndcg@10": 0.
|
| 12 |
-
"val_cosine_mrr@1": 0.
|
| 13 |
-
"val_cosine_mrr@5": 0.
|
| 14 |
-
"val_cosine_mrr@10": 0.
|
| 15 |
-
"val_cosine_map@100": 0.
|
| 16 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"val_cosine_accuracy@1": 0.829575,
|
| 3 |
+
"val_cosine_accuracy@3": 0.9048,
|
| 4 |
+
"val_cosine_accuracy@5": 0.9324,
|
| 5 |
+
"val_cosine_precision@1": 0.829575,
|
| 6 |
+
"val_cosine_precision@3": 0.3016,
|
| 7 |
+
"val_cosine_precision@5": 0.18648000000000003,
|
| 8 |
+
"val_cosine_recall@1": 0.829575,
|
| 9 |
+
"val_cosine_recall@3": 0.9048,
|
| 10 |
+
"val_cosine_recall@5": 0.9324,
|
| 11 |
+
"val_cosine_ndcg@10": 0.8957411186558171,
|
| 12 |
+
"val_cosine_mrr@1": 0.829575,
|
| 13 |
+
"val_cosine_mrr@5": 0.8693266666666628,
|
| 14 |
+
"val_cosine_mrr@10": 0.873717658730154,
|
| 15 |
+
"val_cosine_map@100": 0.8757871539962314
|
| 16 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 114011616
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:36f07360daa85c59707827836aa470469164811479c2802fa0e2cd50f554b85f
|
| 3 |
size 114011616
|