Training in progress, step 5000
Browse files- Information-Retrieval_evaluation_val_results.csv +1 -0
- README.md +74 -346
- eval/Information-Retrieval_evaluation_val_results.csv +144 -0
- final_metrics.json +14 -14
- model.safetensors +1 -1
- training_args.bin +1 -1
Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -3,3 +3,4 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
|
|
| 3 |
-1,-1,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451366666666663,0.9466038095238087,0.9586270476620361,0.946959374340519
|
| 4 |
-1,-1,0.829575,0.9048,0.9324,0.829575,0.829575,0.3016,0.9048,0.18648000000000003,0.9324,0.829575,0.8693266666666628,0.873717658730154,0.8957411186558171,0.8757871539962314
|
| 5 |
-1,-1,0.829275,0.9051,0.9329,0.829275,0.829275,0.30169999999999997,0.9051,0.18658000000000002,0.9329,0.829275,0.8692179166666618,0.8735753373015815,0.8956869608914538,0.8756452160249361
|
|
|
|
|
|
| 3 |
-1,-1,0.9184,0.97,0.9852,0.9184,0.9184,0.3233333333333333,0.97,0.19703999999999997,0.9852,0.9184,0.9451366666666663,0.9466038095238087,0.9586270476620361,0.946959374340519
|
| 4 |
-1,-1,0.829575,0.9048,0.9324,0.829575,0.829575,0.3016,0.9048,0.18648000000000003,0.9324,0.829575,0.8693266666666628,0.873717658730154,0.8957411186558171,0.8757871539962314
|
| 5 |
-1,-1,0.829275,0.9051,0.9329,0.829275,0.829275,0.30169999999999997,0.9051,0.18658000000000002,0.9329,0.829275,0.8692179166666618,0.8735753373015815,0.8956869608914538,0.8756452160249361
|
| 6 |
+
-1,-1,0.76415,0.822175,0.84665,0.76415,0.76415,0.2740583333333333,0.822175,0.16933,0.84665,0.76415,0.7953370833333295,0.8000376587301573,0.8195085862842543,0.8034052122492592
|
README.md
CHANGED
|
@@ -5,110 +5,38 @@ tags:
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
-
- dataset_size:
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
-
- source_sentence:
|
| 13 |
-
for one that's not married? Which one is for what?
|
| 14 |
sentences:
|
| 15 |
-
-
|
| 16 |
-
|
| 17 |
-
-
|
| 18 |
-
|
| 19 |
-
- source_sentence: Which ointment is applied to the face of UFC fighters at the commencement
|
| 20 |
-
of a bout? What does it do?
|
| 21 |
sentences:
|
| 22 |
-
-
|
| 23 |
-
-
|
| 24 |
-
|
| 25 |
-
-
|
| 26 |
-
- source_sentence:
|
| 27 |
sentences:
|
| 28 |
-
-
|
| 29 |
-
-
|
| 30 |
-
-
|
| 31 |
-
- source_sentence:
|
| 32 |
-
no more on Menu! When if ever will I atleast get refund in cr card a/c?
|
| 33 |
sentences:
|
| 34 |
-
-
|
| 35 |
-
-
|
| 36 |
-
-
|
| 37 |
-
|
| 38 |
-
- source_sentence: How do you earn money on Quora?
|
| 39 |
sentences:
|
| 40 |
-
-
|
| 41 |
-
-
|
| 42 |
-
-
|
| 43 |
pipeline_tag: sentence-similarity
|
| 44 |
library_name: sentence-transformers
|
| 45 |
-
metrics:
|
| 46 |
-
- cosine_accuracy@1
|
| 47 |
-
- cosine_accuracy@3
|
| 48 |
-
- cosine_accuracy@5
|
| 49 |
-
- cosine_precision@1
|
| 50 |
-
- cosine_precision@3
|
| 51 |
-
- cosine_precision@5
|
| 52 |
-
- cosine_recall@1
|
| 53 |
-
- cosine_recall@3
|
| 54 |
-
- cosine_recall@5
|
| 55 |
-
- cosine_ndcg@10
|
| 56 |
-
- cosine_mrr@1
|
| 57 |
-
- cosine_mrr@5
|
| 58 |
-
- cosine_mrr@10
|
| 59 |
-
- cosine_map@100
|
| 60 |
-
model-index:
|
| 61 |
-
- name: SentenceTransformer based on prajjwal1/bert-small
|
| 62 |
-
results:
|
| 63 |
-
- task:
|
| 64 |
-
type: information-retrieval
|
| 65 |
-
name: Information Retrieval
|
| 66 |
-
dataset:
|
| 67 |
-
name: val
|
| 68 |
-
type: val
|
| 69 |
-
metrics:
|
| 70 |
-
- type: cosine_accuracy@1
|
| 71 |
-
value: 0.764
|
| 72 |
-
name: Cosine Accuracy@1
|
| 73 |
-
- type: cosine_accuracy@3
|
| 74 |
-
value: 0.8221
|
| 75 |
-
name: Cosine Accuracy@3
|
| 76 |
-
- type: cosine_accuracy@5
|
| 77 |
-
value: 0.84635
|
| 78 |
-
name: Cosine Accuracy@5
|
| 79 |
-
- type: cosine_precision@1
|
| 80 |
-
value: 0.764
|
| 81 |
-
name: Cosine Precision@1
|
| 82 |
-
- type: cosine_precision@3
|
| 83 |
-
value: 0.27403333333333335
|
| 84 |
-
name: Cosine Precision@3
|
| 85 |
-
- type: cosine_precision@5
|
| 86 |
-
value: 0.16927000000000003
|
| 87 |
-
name: Cosine Precision@5
|
| 88 |
-
- type: cosine_recall@1
|
| 89 |
-
value: 0.764
|
| 90 |
-
name: Cosine Recall@1
|
| 91 |
-
- type: cosine_recall@3
|
| 92 |
-
value: 0.8221
|
| 93 |
-
name: Cosine Recall@3
|
| 94 |
-
- type: cosine_recall@5
|
| 95 |
-
value: 0.84635
|
| 96 |
-
name: Cosine Recall@5
|
| 97 |
-
- type: cosine_ndcg@10
|
| 98 |
-
value: 0.8194586562525387
|
| 99 |
-
name: Cosine Ndcg@10
|
| 100 |
-
- type: cosine_mrr@1
|
| 101 |
-
value: 0.764
|
| 102 |
-
name: Cosine Mrr@1
|
| 103 |
-
- type: cosine_mrr@5
|
| 104 |
-
value: 0.7952029166666628
|
| 105 |
-
name: Cosine Mrr@5
|
| 106 |
-
- type: cosine_mrr@10
|
| 107 |
-
value: 0.7999395039682529
|
| 108 |
-
name: Cosine Mrr@10
|
| 109 |
-
- type: cosine_map@100
|
| 110 |
-
value: 0.8032950373722874
|
| 111 |
-
name: Cosine Map@100
|
| 112 |
---
|
| 113 |
|
| 114 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
@@ -157,12 +85,12 @@ Then you can load this model and run inference.
|
|
| 157 |
from sentence_transformers import SentenceTransformer
|
| 158 |
|
| 159 |
# Download from the 🤗 Hub
|
| 160 |
-
model = SentenceTransformer("
|
| 161 |
# Run inference
|
| 162 |
sentences = [
|
| 163 |
-
'
|
| 164 |
-
'
|
| 165 |
-
'
|
| 166 |
]
|
| 167 |
embeddings = model.encode(sentences)
|
| 168 |
print(embeddings.shape)
|
|
@@ -171,9 +99,9 @@ print(embeddings.shape)
|
|
| 171 |
# Get the similarity scores for the embeddings
|
| 172 |
similarities = model.similarity(embeddings, embeddings)
|
| 173 |
print(similarities)
|
| 174 |
-
# tensor([[
|
| 175 |
-
# [ 0.
|
| 176 |
-
# [-0.
|
| 177 |
```
|
| 178 |
|
| 179 |
<!--
|
|
@@ -200,32 +128,6 @@ You can finetune this model on your own dataset.
|
|
| 200 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 201 |
-->
|
| 202 |
|
| 203 |
-
## Evaluation
|
| 204 |
-
|
| 205 |
-
### Metrics
|
| 206 |
-
|
| 207 |
-
#### Information Retrieval
|
| 208 |
-
|
| 209 |
-
* Dataset: `val`
|
| 210 |
-
* Evaluated with [<code>InformationRetrievalEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.InformationRetrievalEvaluator)
|
| 211 |
-
|
| 212 |
-
| Metric | Value |
|
| 213 |
-
|:-------------------|:-----------|
|
| 214 |
-
| cosine_accuracy@1 | 0.764 |
|
| 215 |
-
| cosine_accuracy@3 | 0.8221 |
|
| 216 |
-
| cosine_accuracy@5 | 0.8464 |
|
| 217 |
-
| cosine_precision@1 | 0.764 |
|
| 218 |
-
| cosine_precision@3 | 0.274 |
|
| 219 |
-
| cosine_precision@5 | 0.1693 |
|
| 220 |
-
| cosine_recall@1 | 0.764 |
|
| 221 |
-
| cosine_recall@3 | 0.8221 |
|
| 222 |
-
| cosine_recall@5 | 0.8464 |
|
| 223 |
-
| **cosine_ndcg@10** | **0.8195** |
|
| 224 |
-
| cosine_mrr@1 | 0.764 |
|
| 225 |
-
| cosine_mrr@5 | 0.7952 |
|
| 226 |
-
| cosine_mrr@10 | 0.7999 |
|
| 227 |
-
| cosine_map@100 | 0.8033 |
|
| 228 |
-
|
| 229 |
<!--
|
| 230 |
## Bias, Risks and Limitations
|
| 231 |
|
|
@@ -244,49 +146,23 @@ You can finetune this model on your own dataset.
|
|
| 244 |
|
| 245 |
#### Unnamed Dataset
|
| 246 |
|
| 247 |
-
* Size:
|
| 248 |
-
* Columns: <code>
|
| 249 |
* Approximate statistics based on the first 1000 samples:
|
| 250 |
-
| |
|
| 251 |
-
|
| 252 |
-
| type | string | string
|
| 253 |
-
| details | <ul><li>min:
|
| 254 |
* Samples:
|
| 255 |
-
|
|
| 256 |
-
|
| 257 |
-
| <code>
|
| 258 |
-
| <code>
|
| 259 |
-
| <code>What
|
| 260 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 261 |
```json
|
| 262 |
{
|
| 263 |
-
"scale":
|
| 264 |
-
"similarity_fct": "cos_sim",
|
| 265 |
-
"gather_across_devices": false
|
| 266 |
-
}
|
| 267 |
-
```
|
| 268 |
-
|
| 269 |
-
### Evaluation Dataset
|
| 270 |
-
|
| 271 |
-
#### Unnamed Dataset
|
| 272 |
-
|
| 273 |
-
* Size: 40,000 evaluation samples
|
| 274 |
-
* Columns: <code>anchor</code>, <code>positive</code>, and <code>negative</code>
|
| 275 |
-
* Approximate statistics based on the first 1000 samples:
|
| 276 |
-
| | anchor | positive | negative |
|
| 277 |
-
|:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|
|
| 278 |
-
| type | string | string | string |
|
| 279 |
-
| details | <ul><li>min: 6 tokens</li><li>mean: 15.71 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 15.79 tokens</li><li>max: 65 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.8 tokens</li><li>max: 78 tokens</li></ul> |
|
| 280 |
-
* Samples:
|
| 281 |
-
| anchor | positive | negative |
|
| 282 |
-
|:------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------|
|
| 283 |
-
| <code>Why were feathered dinosaur fossils only found in the last 20 years?</code> | <code>Why were feathered dinosaur fossils only found in the last 20 years?</code> | <code>Why are only few people aware that many dinosaurs had feathers?</code> |
|
| 284 |
-
| <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>If FOX News is the conservative news station, which cable news network is for liberals/progressives?</code> | <code>How much did Fox News and conservative leaning media networks stoke the anger that contributed to Donald Trump's popularity?</code> |
|
| 285 |
-
| <code>How can guys last longer during sex?</code> | <code>How do I last longer in sex?</code> | <code>How do you get over the fear of death while fighting a war?</code> |
|
| 286 |
-
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 287 |
-
```json
|
| 288 |
-
{
|
| 289 |
-
"scale": 1.0,
|
| 290 |
"similarity_fct": "cos_sim",
|
| 291 |
"gather_across_devices": false
|
| 292 |
}
|
|
@@ -295,49 +171,36 @@ You can finetune this model on your own dataset.
|
|
| 295 |
### Training Hyperparameters
|
| 296 |
#### Non-Default Hyperparameters
|
| 297 |
|
| 298 |
-
- `
|
| 299 |
-
- `
|
| 300 |
-
- `per_device_eval_batch_size`: 256
|
| 301 |
-
- `learning_rate`: 2e-05
|
| 302 |
-
- `weight_decay`: 0.001
|
| 303 |
-
- `max_steps`: 14060
|
| 304 |
-
- `warmup_ratio`: 0.1
|
| 305 |
- `fp16`: True
|
| 306 |
-
- `
|
| 307 |
-
- `dataloader_num_workers`: 1
|
| 308 |
-
- `dataloader_prefetch_factor`: 1
|
| 309 |
-
- `load_best_model_at_end`: True
|
| 310 |
-
- `optim`: adamw_torch
|
| 311 |
-
- `ddp_find_unused_parameters`: False
|
| 312 |
-
- `push_to_hub`: True
|
| 313 |
-
- `hub_model_id`: redis/model-a-baseline
|
| 314 |
-
- `eval_on_start`: True
|
| 315 |
|
| 316 |
#### All Hyperparameters
|
| 317 |
<details><summary>Click to expand</summary>
|
| 318 |
|
| 319 |
- `overwrite_output_dir`: False
|
| 320 |
- `do_predict`: False
|
| 321 |
-
- `eval_strategy`:
|
| 322 |
- `prediction_loss_only`: True
|
| 323 |
-
- `per_device_train_batch_size`:
|
| 324 |
-
- `per_device_eval_batch_size`:
|
| 325 |
- `per_gpu_train_batch_size`: None
|
| 326 |
- `per_gpu_eval_batch_size`: None
|
| 327 |
- `gradient_accumulation_steps`: 1
|
| 328 |
- `eval_accumulation_steps`: None
|
| 329 |
- `torch_empty_cache_steps`: None
|
| 330 |
-
- `learning_rate`:
|
| 331 |
-
- `weight_decay`: 0.
|
| 332 |
- `adam_beta1`: 0.9
|
| 333 |
- `adam_beta2`: 0.999
|
| 334 |
- `adam_epsilon`: 1e-08
|
| 335 |
-
- `max_grad_norm`: 1
|
| 336 |
-
- `num_train_epochs`: 3
|
| 337 |
-
- `max_steps`:
|
| 338 |
- `lr_scheduler_type`: linear
|
| 339 |
- `lr_scheduler_kwargs`: {}
|
| 340 |
-
- `warmup_ratio`: 0.
|
| 341 |
- `warmup_steps`: 0
|
| 342 |
- `log_level`: passive
|
| 343 |
- `log_level_replica`: warning
|
|
@@ -365,14 +228,14 @@ You can finetune this model on your own dataset.
|
|
| 365 |
- `tpu_num_cores`: None
|
| 366 |
- `tpu_metrics_debug`: False
|
| 367 |
- `debug`: []
|
| 368 |
-
- `dataloader_drop_last`:
|
| 369 |
-
- `dataloader_num_workers`:
|
| 370 |
-
- `dataloader_prefetch_factor`:
|
| 371 |
- `past_index`: -1
|
| 372 |
- `disable_tqdm`: False
|
| 373 |
- `remove_unused_columns`: True
|
| 374 |
- `label_names`: None
|
| 375 |
-
- `load_best_model_at_end`:
|
| 376 |
- `ignore_data_skip`: False
|
| 377 |
- `fsdp`: []
|
| 378 |
- `fsdp_min_num_params`: 0
|
|
@@ -382,23 +245,23 @@ You can finetune this model on your own dataset.
|
|
| 382 |
- `parallelism_config`: None
|
| 383 |
- `deepspeed`: None
|
| 384 |
- `label_smoothing_factor`: 0.0
|
| 385 |
-
- `optim`:
|
| 386 |
- `optim_args`: None
|
| 387 |
- `adafactor`: False
|
| 388 |
- `group_by_length`: False
|
| 389 |
- `length_column_name`: length
|
| 390 |
- `project`: huggingface
|
| 391 |
- `trackio_space_id`: trackio
|
| 392 |
-
- `ddp_find_unused_parameters`:
|
| 393 |
- `ddp_bucket_cap_mb`: None
|
| 394 |
- `ddp_broadcast_buffers`: False
|
| 395 |
- `dataloader_pin_memory`: True
|
| 396 |
- `dataloader_persistent_workers`: False
|
| 397 |
- `skip_memory_metrics`: True
|
| 398 |
- `use_legacy_prediction_loop`: False
|
| 399 |
-
- `push_to_hub`:
|
| 400 |
- `resume_from_checkpoint`: None
|
| 401 |
-
- `hub_model_id`:
|
| 402 |
- `hub_strategy`: every_save
|
| 403 |
- `hub_private_repo`: None
|
| 404 |
- `hub_always_push`: False
|
|
@@ -425,167 +288,32 @@ You can finetune this model on your own dataset.
|
|
| 425 |
- `neftune_noise_alpha`: None
|
| 426 |
- `optim_target_modules`: None
|
| 427 |
- `batch_eval_metrics`: False
|
| 428 |
-
- `eval_on_start`:
|
| 429 |
- `use_liger_kernel`: False
|
| 430 |
- `liger_kernel_config`: None
|
| 431 |
- `eval_use_gather_object`: False
|
| 432 |
- `average_tokens_across_devices`: True
|
| 433 |
- `prompts`: None
|
| 434 |
- `batch_sampler`: batch_sampler
|
| 435 |
-
- `multi_dataset_batch_sampler`:
|
| 436 |
- `router_mapping`: {}
|
| 437 |
- `learning_rate_mapping`: {}
|
| 438 |
|
| 439 |
</details>
|
| 440 |
|
| 441 |
### Training Logs
|
| 442 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 443 |
|
| 444 |
-
| Epoch | Step | Training Loss | Validation Loss | val_cosine_ndcg@10 |
|
| 445 |
-
|:------:|:-----:|:-------------:|:---------------:|:------------------:|
|
| 446 |
-
| 0 | 0 | - | 5.9519 | 0.8045 |
|
| 447 |
-
| 0.0711 | 100 | 5.9667 | 5.7035 | 0.8414 |
|
| 448 |
-
| 0.1422 | 200 | 5.7144 | 5.4232 | 0.8556 |
|
| 449 |
-
| 0.2134 | 300 | 5.5593 | 5.3409 | 0.8496 |
|
| 450 |
-
| 0.2845 | 400 | 5.4925 | 5.3198 | 0.8407 |
|
| 451 |
-
| 0.3556 | 500 | 5.457 | 5.3155 | 0.8334 |
|
| 452 |
-
| 0.4267 | 600 | 5.4353 | 5.3145 | 0.8287 |
|
| 453 |
-
| 0.4979 | 700 | 5.4205 | 5.3114 | 0.8248 |
|
| 454 |
-
| 0.5690 | 800 | 5.4107 | 5.3092 | 0.8222 |
|
| 455 |
-
| 0.6401 | 900 | 5.4 | 5.3071 | 0.8192 |
|
| 456 |
-
| 0.7112 | 1000 | 5.3927 | 5.3057 | 0.8176 |
|
| 457 |
-
| 0.7824 | 1100 | 5.3867 | 5.3042 | 0.8173 |
|
| 458 |
-
| 0.8535 | 1200 | 5.3815 | 5.3008 | 0.8171 |
|
| 459 |
-
| 0.9246 | 1300 | 5.3757 | 5.2991 | 0.8163 |
|
| 460 |
-
| 0.9957 | 1400 | 5.3711 | 5.2976 | 0.8156 |
|
| 461 |
-
| 1.0669 | 1500 | 5.3671 | 5.2953 | 0.8160 |
|
| 462 |
-
| 1.1380 | 1600 | 5.3641 | 5.2942 | 0.8162 |
|
| 463 |
-
| 1.2091 | 1700 | 5.3611 | 5.2941 | 0.8166 |
|
| 464 |
-
| 1.2802 | 1800 | 5.3588 | 5.2927 | 0.8167 |
|
| 465 |
-
| 1.3514 | 1900 | 5.3549 | 5.2916 | 0.8162 |
|
| 466 |
-
| 1.4225 | 2000 | 5.3542 | 5.2915 | 0.8162 |
|
| 467 |
-
| 1.4936 | 2100 | 5.3512 | 5.2903 | 0.8164 |
|
| 468 |
-
| 1.5647 | 2200 | 5.3502 | 5.2898 | 0.8166 |
|
| 469 |
-
| 1.6358 | 2300 | 5.3479 | 5.2898 | 0.8162 |
|
| 470 |
-
| 1.7070 | 2400 | 5.346 | 5.2889 | 0.8161 |
|
| 471 |
-
| 1.7781 | 2500 | 5.3444 | 5.2881 | 0.8161 |
|
| 472 |
-
| 1.8492 | 2600 | 5.3431 | 5.2879 | 0.8164 |
|
| 473 |
-
| 1.9203 | 2700 | 5.3414 | 5.2876 | 0.8165 |
|
| 474 |
-
| 1.9915 | 2800 | 5.34 | 5.2869 | 0.8166 |
|
| 475 |
-
| 2.0626 | 2900 | 5.3384 | 5.2865 | 0.8166 |
|
| 476 |
-
| 2.1337 | 3000 | 5.3365 | 5.2858 | 0.8163 |
|
| 477 |
-
| 2.2048 | 3100 | 5.3356 | 5.2859 | 0.8168 |
|
| 478 |
-
| 2.2760 | 3200 | 5.3346 | 5.2857 | 0.8163 |
|
| 479 |
-
| 2.3471 | 3300 | 5.3338 | 5.2850 | 0.8166 |
|
| 480 |
-
| 2.4182 | 3400 | 5.3324 | 5.2851 | 0.8171 |
|
| 481 |
-
| 2.4893 | 3500 | 5.3323 | 5.2845 | 0.8170 |
|
| 482 |
-
| 2.5605 | 3600 | 5.33 | 5.2837 | 0.8172 |
|
| 483 |
-
| 2.6316 | 3700 | 5.3297 | 5.2841 | 0.8174 |
|
| 484 |
-
| 2.7027 | 3800 | 5.3281 | 5.2838 | 0.8173 |
|
| 485 |
-
| 2.7738 | 3900 | 5.3273 | 5.2835 | 0.8175 |
|
| 486 |
-
| 2.8450 | 4000 | 5.3259 | 5.2829 | 0.8176 |
|
| 487 |
-
| 2.9161 | 4100 | 5.3258 | 5.2834 | 0.8182 |
|
| 488 |
-
| 2.9872 | 4200 | 5.3245 | 5.2827 | 0.8181 |
|
| 489 |
-
| 3.0583 | 4300 | 5.3229 | 5.2821 | 0.8184 |
|
| 490 |
-
| 3.1294 | 4400 | 5.3224 | 5.2820 | 0.8181 |
|
| 491 |
-
| 3.2006 | 4500 | 5.3217 | 5.2819 | 0.8179 |
|
| 492 |
-
| 3.2717 | 4600 | 5.3212 | 5.2820 | 0.8181 |
|
| 493 |
-
| 3.3428 | 4700 | 5.3208 | 5.2814 | 0.8186 |
|
| 494 |
-
| 3.4139 | 4800 | 5.3197 | 5.2819 | 0.8181 |
|
| 495 |
-
| 3.4851 | 4900 | 5.3196 | 5.2815 | 0.8183 |
|
| 496 |
-
| 3.5562 | 5000 | 5.3189 | 5.2811 | 0.8184 |
|
| 497 |
-
| 3.6273 | 5100 | 5.3183 | 5.2807 | 0.8181 |
|
| 498 |
-
| 3.6984 | 5200 | 5.3169 | 5.2808 | 0.8182 |
|
| 499 |
-
| 3.7696 | 5300 | 5.3168 | 5.2805 | 0.8182 |
|
| 500 |
-
| 3.8407 | 5400 | 5.3163 | 5.2806 | 0.8180 |
|
| 501 |
-
| 3.9118 | 5500 | 5.3157 | 5.2806 | 0.8180 |
|
| 502 |
-
| 3.9829 | 5600 | 5.3153 | 5.2803 | 0.8184 |
|
| 503 |
-
| 4.0541 | 5700 | 5.3147 | 5.2801 | 0.8184 |
|
| 504 |
-
| 4.1252 | 5800 | 5.3134 | 5.2800 | 0.8183 |
|
| 505 |
-
| 4.1963 | 5900 | 5.313 | 5.2796 | 0.8185 |
|
| 506 |
-
| 4.2674 | 6000 | 5.3131 | 5.2797 | 0.8186 |
|
| 507 |
-
| 4.3385 | 6100 | 5.3118 | 5.2795 | 0.8184 |
|
| 508 |
-
| 4.4097 | 6200 | 5.3115 | 5.2792 | 0.8183 |
|
| 509 |
-
| 4.4808 | 6300 | 5.3111 | 5.2790 | 0.8186 |
|
| 510 |
-
| 4.5519 | 6400 | 5.3112 | 5.2793 | 0.8188 |
|
| 511 |
-
| 4.6230 | 6500 | 5.3103 | 5.2792 | 0.8190 |
|
| 512 |
-
| 4.6942 | 6600 | 5.3104 | 5.2788 | 0.8192 |
|
| 513 |
-
| 4.7653 | 6700 | 5.3093 | 5.2792 | 0.8192 |
|
| 514 |
-
| 4.8364 | 6800 | 5.309 | 5.2784 | 0.8189 |
|
| 515 |
-
| 4.9075 | 6900 | 5.3095 | 5.2788 | 0.8191 |
|
| 516 |
-
| 4.9787 | 7000 | 5.3083 | 5.2781 | 0.8192 |
|
| 517 |
-
| 5.0498 | 7100 | 5.3077 | 5.2784 | 0.8193 |
|
| 518 |
-
| 5.1209 | 7200 | 5.3071 | 5.2783 | 0.8191 |
|
| 519 |
-
| 5.1920 | 7300 | 5.307 | 5.2781 | 0.8189 |
|
| 520 |
-
| 5.2632 | 7400 | 5.3069 | 5.2781 | 0.8192 |
|
| 521 |
-
| 5.3343 | 7500 | 5.3064 | 5.2784 | 0.8193 |
|
| 522 |
-
| 5.4054 | 7600 | 5.3061 | 5.2778 | 0.8193 |
|
| 523 |
-
| 5.4765 | 7700 | 5.3054 | 5.2777 | 0.8193 |
|
| 524 |
-
| 5.5477 | 7800 | 5.3056 | 5.2781 | 0.8193 |
|
| 525 |
-
| 5.6188 | 7900 | 5.3054 | 5.2778 | 0.8192 |
|
| 526 |
-
| 5.6899 | 8000 | 5.305 | 5.2774 | 0.8191 |
|
| 527 |
-
| 5.7610 | 8100 | 5.3049 | 5.2773 | 0.8191 |
|
| 528 |
-
| 5.8321 | 8200 | 5.3041 | 5.2774 | 0.8188 |
|
| 529 |
-
| 5.9033 | 8300 | 5.3041 | 5.2771 | 0.8191 |
|
| 530 |
-
| 5.9744 | 8400 | 5.3039 | 5.2775 | 0.8191 |
|
| 531 |
-
| 6.0455 | 8500 | 5.3031 | 5.2770 | 0.8190 |
|
| 532 |
-
| 6.1166 | 8600 | 5.3028 | 5.2771 | 0.8191 |
|
| 533 |
-
| 6.1878 | 8700 | 5.3026 | 5.2770 | 0.8189 |
|
| 534 |
-
| 6.2589 | 8800 | 5.302 | 5.2772 | 0.8189 |
|
| 535 |
-
| 6.3300 | 8900 | 5.3025 | 5.2772 | 0.8190 |
|
| 536 |
-
| 6.4011 | 9000 | 5.3025 | 5.2768 | 0.8189 |
|
| 537 |
-
| 6.4723 | 9100 | 5.3022 | 5.2765 | 0.8191 |
|
| 538 |
-
| 6.5434 | 9200 | 5.3021 | 5.2765 | 0.8193 |
|
| 539 |
-
| 6.6145 | 9300 | 5.3014 | 5.2765 | 0.8196 |
|
| 540 |
-
| 6.6856 | 9400 | 5.3012 | 5.2766 | 0.8194 |
|
| 541 |
-
| 6.7568 | 9500 | 5.3008 | 5.2763 | 0.8194 |
|
| 542 |
-
| 6.8279 | 9600 | 5.3009 | 5.2764 | 0.8193 |
|
| 543 |
-
| 6.8990 | 9700 | 5.3009 | 5.2762 | 0.8195 |
|
| 544 |
-
| 6.9701 | 9800 | 5.3008 | 5.2762 | 0.8192 |
|
| 545 |
-
| 7.0413 | 9900 | 5.3001 | 5.2763 | 0.8193 |
|
| 546 |
-
| 7.1124 | 10000 | 5.2998 | 5.2762 | 0.8194 |
|
| 547 |
-
| 7.1835 | 10100 | 5.2996 | 5.2760 | 0.8195 |
|
| 548 |
-
| 7.2546 | 10200 | 5.2996 | 5.2763 | 0.8192 |
|
| 549 |
-
| 7.3257 | 10300 | 5.2992 | 5.2761 | 0.8194 |
|
| 550 |
-
| 7.3969 | 10400 | 5.2995 | 5.2762 | 0.8193 |
|
| 551 |
-
| 7.4680 | 10500 | 5.2994 | 5.2760 | 0.8191 |
|
| 552 |
-
| 7.5391 | 10600 | 5.2986 | 5.2758 | 0.8192 |
|
| 553 |
-
| 7.6102 | 10700 | 5.2985 | 5.2760 | 0.8190 |
|
| 554 |
-
| 7.6814 | 10800 | 5.2988 | 5.2758 | 0.8193 |
|
| 555 |
-
| 7.7525 | 10900 | 5.2993 | 5.2757 | 0.8190 |
|
| 556 |
-
| 7.8236 | 11000 | 5.2988 | 5.2758 | 0.8194 |
|
| 557 |
-
| 7.8947 | 11100 | 5.2989 | 5.2757 | 0.8192 |
|
| 558 |
-
| 7.9659 | 11200 | 5.2987 | 5.2757 | 0.8195 |
|
| 559 |
-
| 8.0370 | 11300 | 5.2983 | 5.2756 | 0.8194 |
|
| 560 |
-
| 8.1081 | 11400 | 5.2981 | 5.2756 | 0.8196 |
|
| 561 |
-
| 8.1792 | 11500 | 5.2981 | 5.2756 | 0.8192 |
|
| 562 |
-
| 8.2504 | 11600 | 5.2977 | 5.2757 | 0.8194 |
|
| 563 |
-
| 8.3215 | 11700 | 5.2979 | 5.2756 | 0.8194 |
|
| 564 |
-
| 8.3926 | 11800 | 5.2975 | 5.2754 | 0.8195 |
|
| 565 |
-
| 8.4637 | 11900 | 5.2976 | 5.2756 | 0.8195 |
|
| 566 |
-
| 8.5349 | 12000 | 5.2973 | 5.2755 | 0.8194 |
|
| 567 |
-
| 8.6060 | 12100 | 5.2969 | 5.2754 | 0.8195 |
|
| 568 |
-
| 8.6771 | 12200 | 5.2979 | 5.2755 | 0.8193 |
|
| 569 |
-
| 8.7482 | 12300 | 5.2976 | 5.2754 | 0.8194 |
|
| 570 |
-
| 8.8193 | 12400 | 5.2977 | 5.2753 | 0.8194 |
|
| 571 |
-
| 8.8905 | 12500 | 5.2973 | 5.2754 | 0.8193 |
|
| 572 |
-
| 8.9616 | 12600 | 5.297 | 5.2753 | 0.8195 |
|
| 573 |
-
| 9.0327 | 12700 | 5.2964 | 5.2754 | 0.8195 |
|
| 574 |
-
| 9.1038 | 12800 | 5.2971 | 5.2754 | 0.8193 |
|
| 575 |
-
| 9.1750 | 12900 | 5.2969 | 5.2753 | 0.8194 |
|
| 576 |
-
| 9.2461 | 13000 | 5.2969 | 5.2752 | 0.8193 |
|
| 577 |
-
| 9.3172 | 13100 | 5.2966 | 5.2753 | 0.8194 |
|
| 578 |
-
| 9.3883 | 13200 | 5.2968 | 5.2752 | 0.8195 |
|
| 579 |
-
| 9.4595 | 13300 | 5.2972 | 5.2753 | 0.8194 |
|
| 580 |
-
| 9.5306 | 13400 | 5.2968 | 5.2752 | 0.8194 |
|
| 581 |
-
| 9.6017 | 13500 | 5.2963 | 5.2751 | 0.8196 |
|
| 582 |
-
| 9.6728 | 13600 | 5.2968 | 5.2751 | 0.8196 |
|
| 583 |
-
| 9.7440 | 13700 | 5.2967 | 5.2752 | 0.8197 |
|
| 584 |
-
| 9.8151 | 13800 | 5.2965 | 5.2752 | 0.8195 |
|
| 585 |
-
| 9.8862 | 13900 | 5.2962 | 5.2751 | 0.8193 |
|
| 586 |
-
| 9.9573 | 14000 | 5.2964 | 5.2751 | 0.8195 |
|
| 587 |
-
|
| 588 |
-
</details>
|
| 589 |
|
| 590 |
### Framework Versions
|
| 591 |
- Python: 3.10.18
|
|
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
+
- dataset_size:100000
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
base_model: prajjwal1/bert-small
|
| 11 |
widget:
|
| 12 |
+
- source_sentence: How do I polish my English skills?
|
|
|
|
| 13 |
sentences:
|
| 14 |
+
- How can we polish English skills?
|
| 15 |
+
- Why should I move to Israel as a Jew?
|
| 16 |
+
- What are vitamins responsible for?
|
| 17 |
+
- source_sentence: Can I use the Kozuka Gothic Pro font as a font-face on my web site?
|
|
|
|
|
|
|
| 18 |
sentences:
|
| 19 |
+
- Can I use the Kozuka Gothic Pro font as a font-face on my web site?
|
| 20 |
+
- Why are Google, Facebook, YouTube and other social networking sites banned in
|
| 21 |
+
China?
|
| 22 |
+
- What font is used in Bloomberg Terminal?
|
| 23 |
+
- source_sentence: Is Quora the best Q&A site?
|
| 24 |
sentences:
|
| 25 |
+
- What was the best Quora question ever?
|
| 26 |
+
- Is Quora the best inquiry site?
|
| 27 |
+
- Where do I buy Oway hair products online?
|
| 28 |
+
- source_sentence: How can I customize my walking speed on Google Maps?
|
|
|
|
| 29 |
sentences:
|
| 30 |
+
- How do I bring back Google maps icon in my home screen?
|
| 31 |
+
- How many pages are there in all the Harry Potter books combined?
|
| 32 |
+
- How can I customize my walking speed on Google Maps?
|
| 33 |
+
- source_sentence: DId something exist before the Big Bang?
|
|
|
|
| 34 |
sentences:
|
| 35 |
+
- How can I improve my memory problem?
|
| 36 |
+
- Where can I buy Fairy Tail Manga?
|
| 37 |
+
- Is there a scientific name for what existed before the Big Bang?
|
| 38 |
pipeline_tag: sentence-similarity
|
| 39 |
library_name: sentence-transformers
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 40 |
---
|
| 41 |
|
| 42 |
# SentenceTransformer based on prajjwal1/bert-small
|
|
|
|
| 85 |
from sentence_transformers import SentenceTransformer
|
| 86 |
|
| 87 |
# Download from the 🤗 Hub
|
| 88 |
+
model = SentenceTransformer("sentence_transformers_model_id")
|
| 89 |
# Run inference
|
| 90 |
sentences = [
|
| 91 |
+
'DId something exist before the Big Bang?',
|
| 92 |
+
'Is there a scientific name for what existed before the Big Bang?',
|
| 93 |
+
'Where can I buy Fairy Tail Manga?',
|
| 94 |
]
|
| 95 |
embeddings = model.encode(sentences)
|
| 96 |
print(embeddings.shape)
|
|
|
|
| 99 |
# Get the similarity scores for the embeddings
|
| 100 |
similarities = model.similarity(embeddings, embeddings)
|
| 101 |
print(similarities)
|
| 102 |
+
# tensor([[ 1.0000, 0.7596, -0.0398],
|
| 103 |
+
# [ 0.7596, 1.0000, -0.0308],
|
| 104 |
+
# [-0.0398, -0.0308, 1.0000]])
|
| 105 |
```
|
| 106 |
|
| 107 |
<!--
|
|
|
|
| 128 |
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
|
| 129 |
-->
|
| 130 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
<!--
|
| 132 |
## Bias, Risks and Limitations
|
| 133 |
|
|
|
|
| 146 |
|
| 147 |
#### Unnamed Dataset
|
| 148 |
|
| 149 |
+
* Size: 100,000 training samples
|
| 150 |
+
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>sentence_2</code>
|
| 151 |
* Approximate statistics based on the first 1000 samples:
|
| 152 |
+
| | sentence_0 | sentence_1 | sentence_2 |
|
| 153 |
+
|:--------|:----------------------------------------------------------------------------------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|
|
| 154 |
+
| type | string | string | string |
|
| 155 |
+
| details | <ul><li>min: 3 tokens</li><li>mean: 15.53 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 15.5 tokens</li><li>max: 59 tokens</li></ul> | <ul><li>min: 6 tokens</li><li>mean: 16.87 tokens</li><li>max: 128 tokens</li></ul> |
|
| 156 |
* Samples:
|
| 157 |
+
| sentence_0 | sentence_1 | sentence_2 |
|
| 158 |
+
|:----------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------|:-----------------------------------------------------------------------|
|
| 159 |
+
| <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>Is there visitor entry facility in Jaipur airport. How much is the ticket?</code> | <code>How much is the airport tax in bogota?</code> |
|
| 160 |
+
| <code>Which concept is more important: good planning or hard work?</code> | <code>Which concept is more important: good planning or hard work?</code> | <code>What is important in life: luck or hard work?</code> |
|
| 161 |
+
| <code>What is the most efficient way to make money?</code> | <code>How can I make my money make money?</code> | <code>What can one learn about Quantum Mechanics in 10 minutes?</code> |
|
| 162 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 163 |
```json
|
| 164 |
{
|
| 165 |
+
"scale": 20.0,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 166 |
"similarity_fct": "cos_sim",
|
| 167 |
"gather_across_devices": false
|
| 168 |
}
|
|
|
|
| 171 |
### Training Hyperparameters
|
| 172 |
#### Non-Default Hyperparameters
|
| 173 |
|
| 174 |
+
- `per_device_train_batch_size`: 64
|
| 175 |
+
- `per_device_eval_batch_size`: 64
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 176 |
- `fp16`: True
|
| 177 |
+
- `multi_dataset_batch_sampler`: round_robin
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 178 |
|
| 179 |
#### All Hyperparameters
|
| 180 |
<details><summary>Click to expand</summary>
|
| 181 |
|
| 182 |
- `overwrite_output_dir`: False
|
| 183 |
- `do_predict`: False
|
| 184 |
+
- `eval_strategy`: no
|
| 185 |
- `prediction_loss_only`: True
|
| 186 |
+
- `per_device_train_batch_size`: 64
|
| 187 |
+
- `per_device_eval_batch_size`: 64
|
| 188 |
- `per_gpu_train_batch_size`: None
|
| 189 |
- `per_gpu_eval_batch_size`: None
|
| 190 |
- `gradient_accumulation_steps`: 1
|
| 191 |
- `eval_accumulation_steps`: None
|
| 192 |
- `torch_empty_cache_steps`: None
|
| 193 |
+
- `learning_rate`: 5e-05
|
| 194 |
+
- `weight_decay`: 0.0
|
| 195 |
- `adam_beta1`: 0.9
|
| 196 |
- `adam_beta2`: 0.999
|
| 197 |
- `adam_epsilon`: 1e-08
|
| 198 |
+
- `max_grad_norm`: 1
|
| 199 |
+
- `num_train_epochs`: 3
|
| 200 |
+
- `max_steps`: -1
|
| 201 |
- `lr_scheduler_type`: linear
|
| 202 |
- `lr_scheduler_kwargs`: {}
|
| 203 |
+
- `warmup_ratio`: 0.0
|
| 204 |
- `warmup_steps`: 0
|
| 205 |
- `log_level`: passive
|
| 206 |
- `log_level_replica`: warning
|
|
|
|
| 228 |
- `tpu_num_cores`: None
|
| 229 |
- `tpu_metrics_debug`: False
|
| 230 |
- `debug`: []
|
| 231 |
+
- `dataloader_drop_last`: False
|
| 232 |
+
- `dataloader_num_workers`: 0
|
| 233 |
+
- `dataloader_prefetch_factor`: None
|
| 234 |
- `past_index`: -1
|
| 235 |
- `disable_tqdm`: False
|
| 236 |
- `remove_unused_columns`: True
|
| 237 |
- `label_names`: None
|
| 238 |
+
- `load_best_model_at_end`: False
|
| 239 |
- `ignore_data_skip`: False
|
| 240 |
- `fsdp`: []
|
| 241 |
- `fsdp_min_num_params`: 0
|
|
|
|
| 245 |
- `parallelism_config`: None
|
| 246 |
- `deepspeed`: None
|
| 247 |
- `label_smoothing_factor`: 0.0
|
| 248 |
+
- `optim`: adamw_torch_fused
|
| 249 |
- `optim_args`: None
|
| 250 |
- `adafactor`: False
|
| 251 |
- `group_by_length`: False
|
| 252 |
- `length_column_name`: length
|
| 253 |
- `project`: huggingface
|
| 254 |
- `trackio_space_id`: trackio
|
| 255 |
+
- `ddp_find_unused_parameters`: None
|
| 256 |
- `ddp_bucket_cap_mb`: None
|
| 257 |
- `ddp_broadcast_buffers`: False
|
| 258 |
- `dataloader_pin_memory`: True
|
| 259 |
- `dataloader_persistent_workers`: False
|
| 260 |
- `skip_memory_metrics`: True
|
| 261 |
- `use_legacy_prediction_loop`: False
|
| 262 |
+
- `push_to_hub`: False
|
| 263 |
- `resume_from_checkpoint`: None
|
| 264 |
+
- `hub_model_id`: None
|
| 265 |
- `hub_strategy`: every_save
|
| 266 |
- `hub_private_repo`: None
|
| 267 |
- `hub_always_push`: False
|
|
|
|
| 288 |
- `neftune_noise_alpha`: None
|
| 289 |
- `optim_target_modules`: None
|
| 290 |
- `batch_eval_metrics`: False
|
| 291 |
+
- `eval_on_start`: False
|
| 292 |
- `use_liger_kernel`: False
|
| 293 |
- `liger_kernel_config`: None
|
| 294 |
- `eval_use_gather_object`: False
|
| 295 |
- `average_tokens_across_devices`: True
|
| 296 |
- `prompts`: None
|
| 297 |
- `batch_sampler`: batch_sampler
|
| 298 |
+
- `multi_dataset_batch_sampler`: round_robin
|
| 299 |
- `router_mapping`: {}
|
| 300 |
- `learning_rate_mapping`: {}
|
| 301 |
|
| 302 |
</details>
|
| 303 |
|
| 304 |
### Training Logs
|
| 305 |
+
| Epoch | Step | Training Loss |
|
| 306 |
+
|:------:|:----:|:-------------:|
|
| 307 |
+
| 0.3199 | 500 | 0.2284 |
|
| 308 |
+
| 0.6398 | 1000 | 0.0571 |
|
| 309 |
+
| 0.9597 | 1500 | 0.0486 |
|
| 310 |
+
| 1.2796 | 2000 | 0.0378 |
|
| 311 |
+
| 1.5995 | 2500 | 0.0367 |
|
| 312 |
+
| 1.9194 | 3000 | 0.0338 |
|
| 313 |
+
| 2.2393 | 3500 | 0.0327 |
|
| 314 |
+
| 2.5592 | 4000 | 0.0285 |
|
| 315 |
+
| 2.8791 | 4500 | 0.0285 |
|
| 316 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 317 |
|
| 318 |
### Framework Versions
|
| 319 |
- Python: 3.10.18
|
eval/Information-Retrieval_evaluation_val_results.csv
CHANGED
|
@@ -462,3 +462,147 @@ epoch,steps,cosine-Accuracy@1,cosine-Accuracy@3,cosine-Accuracy@5,cosine-Precisi
|
|
| 462 |
9.815078236130867,13800,0.763825,0.82225,0.847,0.763825,0.763825,0.2740833333333333,0.82225,0.16940000000000002,0.847,0.763825,0.7952858333333293,0.799954811507935,0.8195212182618785,0.803303660228457
|
| 463 |
9.88620199146515,13900,0.763875,0.82185,0.84695,0.763875,0.763875,0.27395,0.82185,0.16939000000000004,0.84695,0.763875,0.7952058333333295,0.7998402182539672,0.8193424066611443,0.8032217930140984
|
| 464 |
9.95732574679943,14000,0.764,0.8221,0.84635,0.764,0.764,0.27403333333333335,0.8221,0.16927000000000003,0.84635,0.764,0.7952029166666628,0.7999395039682529,0.8194586562525387,0.8032950373722874
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 462 |
9.815078236130867,13800,0.763825,0.82225,0.847,0.763825,0.763825,0.2740833333333333,0.82225,0.16940000000000002,0.847,0.763825,0.7952858333333293,0.799954811507935,0.8195212182618785,0.803303660228457
|
| 463 |
9.88620199146515,13900,0.763875,0.82185,0.84695,0.763875,0.763875,0.27395,0.82185,0.16939000000000004,0.84695,0.763875,0.7952058333333295,0.7998402182539672,0.8193424066611443,0.8032217930140984
|
| 464 |
9.95732574679943,14000,0.764,0.8221,0.84635,0.764,0.764,0.27403333333333335,0.8221,0.16927000000000003,0.84635,0.764,0.7952029166666628,0.7999395039682529,0.8194586562525387,0.8032950373722874
|
| 465 |
+
0,0,0.754525,0.807575,0.8301,0.754525,0.754525,0.26919166666666666,0.807575,0.16601999999999997,0.8301,0.754525,0.7830537499999958,0.7870734424603156,0.8044818711611815,0.790208838398216
|
| 466 |
+
0.07112375533428165,100,0.7846,0.8459,0.8711,0.7846,0.7846,0.2819666666666666,0.8459,0.17422000000000004,0.8711,0.7846,0.8175595833333287,0.8218771726190455,0.8413714264178179,0.8249413998110878
|
| 467 |
+
0.1422475106685633,200,0.79835,0.8623,0.885675,0.79835,0.79835,0.2874333333333333,0.8623,0.17713500000000004,0.885675,0.79835,0.8318512499999939,0.836128045634916,0.855655110536833,0.8388773888658537
|
| 468 |
+
0.21337126600284495,300,0.793725,0.85595,0.879675,0.793725,0.793725,0.28531666666666666,0.85595,0.175935,0.879675,0.793725,0.826637499999994,0.8306767956349161,0.8496438038402215,0.833452231279345
|
| 469 |
+
0.2844950213371266,400,0.786225,0.84655,0.8699,0.786225,0.786225,0.2821833333333333,0.84655,0.17398000000000002,0.8699,0.786225,0.8181699999999944,0.8221539781746012,0.8406979711380332,0.8248666969847387
|
| 470 |
+
0.35561877667140823,500,0.779925,0.838625,0.862225,0.779925,0.779925,0.27954166666666663,0.838625,0.172445,0.862225,0.779925,0.8113874999999955,0.8152349007936488,0.8333842785333231,0.8179856709019278
|
| 471 |
+
0.4267425320056899,600,0.77615,0.83305,0.85555,0.77615,0.77615,0.2776833333333333,0.83305,0.17111,0.85555,0.77615,0.8065837499999955,0.8106722420634908,0.8286924647605718,0.8134236724071078
|
| 472 |
+
0,0,0.757,0.811375,0.832975,0.757,0.757,0.27045833333333325,0.811375,0.16659500000000002,0.832975,0.757,0.7860745833333305,0.7898006746031736,0.8068310819573903,0.7928863872268869
|
| 473 |
+
0,0,0.7569,0.811325,0.832975,0.7569,0.7569,0.27044166666666664,0.811325,0.166595,0.832975,0.7569,0.7860187499999972,0.7897472123015863,0.8067966793307086,0.7928306424812177
|
| 474 |
+
0.2849002849002849,100,0.79975,0.86215,0.8864,0.79975,0.79975,0.2873833333333333,0.86215,0.17728,0.8864,0.79975,0.8329958333333292,0.8371555059523791,0.8563566180802749,0.8398102921473892
|
| 475 |
+
0.5698005698005698,200,0.7925,0.852,0.87525,0.7925,0.7925,0.284,0.852,0.17505,0.87525,0.7925,0.8242770833333285,0.8281965773809494,0.8465099519180141,0.830860034111173
|
| 476 |
+
0.8547008547008547,300,0.781375,0.838775,0.86035,0.781375,0.781375,0.2795916666666666,0.838775,0.17207,0.86035,0.781375,0.8116999999999953,0.815356507936506,0.8327024022306619,0.8181508397271755
|
| 477 |
+
1.1396011396011396,400,0.7759,0.8309,0.851525,0.7759,0.7759,0.27696666666666664,0.8309,0.170305,0.851525,0.7759,0.8048937499999954,0.8085814682539659,0.8254910406165096,0.8115408891716827
|
| 478 |
+
1.4245014245014245,500,0.772425,0.82605,0.84745,0.772425,0.772425,0.27535,0.82605,0.16949,0.84745,0.772425,0.801122916666663,0.8049703571428553,0.8219946152512894,0.8079477908294356
|
| 479 |
+
1.7094017094017095,600,0.7709,0.824975,0.8463,0.7709,0.7709,0.27499166666666663,0.824975,0.16926,0.8463,0.7709,0.7997583333333297,0.8036332837301585,0.8207335647024644,0.8066639744523467
|
| 480 |
+
1.9943019943019942,700,0.769975,0.823675,0.846275,0.769975,0.769975,0.2745583333333333,0.823675,0.16925500000000002,0.846275,0.769975,0.7989949999999969,0.8027658630952357,0.8199615531109039,0.8058107407966907
|
| 481 |
+
2.2792022792022792,800,0.7693,0.823875,0.846175,0.7693,0.7693,0.274625,0.823875,0.169235,0.846175,0.7693,0.7986245833333296,0.8024050099206331,0.8196545256913207,0.805510324876921
|
| 482 |
+
2.564102564102564,900,0.769575,0.82335,0.84615,0.769575,0.769575,0.27444999999999997,0.82335,0.16923000000000002,0.84615,0.769575,0.7986983333333304,0.802586666666665,0.819957118736493,0.8056878206576987
|
| 483 |
+
2.849002849002849,1000,0.7691,0.8234,0.846125,0.7691,0.7691,0.27446666666666664,0.8234,0.16922500000000001,0.846125,0.7691,0.7984204166666636,0.8023565674603175,0.8198217799919566,0.8055107499465582
|
| 484 |
+
0,0,0.756925,0.81135,0.832975,0.756925,0.756925,0.27044999999999997,0.81135,0.166595,0.832975,0.756925,0.7860320833333306,0.7897576984126973,0.8067989560864608,0.7928436652081169
|
| 485 |
+
0.2849002849002849,100,0.799775,0.862075,0.8864,0.799775,0.799775,0.2873583333333333,0.862075,0.17728,0.8864,0.799775,0.8329937499999959,0.8371539186507918,0.8563545850975712,0.8398090360749207
|
| 486 |
+
0.5698005698005698,200,0.792525,0.852,0.8752,0.792525,0.792525,0.284,0.852,0.17504000000000003,0.8752,0.792525,0.824271249999995,0.8281995238095209,0.8465115466169034,0.8308628965209384
|
| 487 |
+
0.8547008547008547,300,0.781425,0.83875,0.860325,0.781425,0.781425,0.27958333333333324,0.83875,0.172065,0.860325,0.781425,0.8117137499999953,0.8153769940476172,0.8327223716360451,0.8181691434152646
|
| 488 |
+
1.1396011396011396,400,0.776,0.8309,0.851525,0.776,0.776,0.27696666666666664,0.8309,0.170305,0.851525,0.776,0.8049424999999955,0.808629940476188,0.8255265529181479,0.8115893097540797
|
| 489 |
+
1.4245014245014245,500,0.77245,0.826075,0.847425,0.77245,0.77245,0.2753583333333333,0.826075,0.169485,0.847425,0.77245,0.8011295833333297,0.8049792857142839,0.8219959765286132,0.807958647632387
|
| 490 |
+
1.7094017094017095,600,0.770975,0.825,0.84635,0.770975,0.770975,0.275,0.825,0.16926999999999998,0.84635,0.770975,0.7998049999999963,0.8036710218253964,0.8207617608415884,0.8067018917058776
|
| 491 |
+
1.9943019943019942,700,0.770025,0.823675,0.846275,0.770025,0.770025,0.2745583333333333,0.823675,0.16925500000000002,0.846275,0.770025,0.7990199999999967,0.8027929166666641,0.8199867865324695,0.805835145066436
|
| 492 |
+
2.2792022792022792,800,0.769375,0.823875,0.846175,0.769375,0.769375,0.274625,0.823875,0.169235,0.846175,0.769375,0.7986662499999962,0.8024467063492045,0.8196855673585592,0.8055517545231466
|
| 493 |
+
2.564102564102564,900,0.7696,0.82335,0.84615,0.7696,0.7696,0.27444999999999997,0.82335,0.16923000000000002,0.84615,0.7696,0.7987095833333303,0.8025977678571411,0.8199651247639522,0.8056987038056344
|
| 494 |
+
2.849002849002849,1000,0.769175,0.8234,0.846075,0.769175,0.769175,0.27446666666666664,0.8234,0.169215,0.846075,0.769175,0.7984474999999968,0.8023916369047619,0.819847580644878,0.8055455645928971
|
| 495 |
+
0,0,0.756925,0.811325,0.832975,0.756925,0.756925,0.27044166666666664,0.811325,0.16659500000000002,0.832975,0.756925,0.7860366666666637,0.7897715277777766,0.8068261092449236,0.7928502735721635
|
| 496 |
+
0.07112375533428165,100,0.7999,0.861975,0.886425,0.7999,0.7999,0.287325,0.861975,0.177285,0.886425,0.7999,0.8329808333333287,0.8371251984126964,0.8563155772584116,0.8398022217446666
|
| 497 |
+
0.1422475106685633,200,0.794325,0.8553,0.878525,0.794325,0.794325,0.2851,0.8553,0.17570500000000003,0.878525,0.794325,0.8267374999999947,0.8305394841269801,0.8489094362604251,0.8332286266074361
|
| 498 |
+
0.21337126600284495,300,0.78485,0.842075,0.863375,0.78485,0.78485,0.2806916666666667,0.842075,0.172675,0.863375,0.78485,0.8150720833333287,0.8188160119047592,0.83619245007385,0.8215492677449724
|
| 499 |
+
0.2844950213371266,400,0.777775,0.833175,0.8542,0.777775,0.777775,0.27772499999999994,0.833175,0.17084000000000002,0.8542,0.777775,0.8070979166666624,0.8107996230158705,0.8278241316562024,0.8136652065977691
|
| 500 |
+
0.35561877667140823,500,0.77395,0.8277,0.848925,0.77395,0.77395,0.2759,0.8277,0.169785,0.848925,0.77395,0.8026024999999963,0.8064195436507905,0.8234369127110808,0.8092864273342918
|
| 501 |
+
0,0,0.75695,0.81135,0.833025,0.75695,0.75695,0.27044999999999997,0.81135,0.16660500000000003,0.833025,0.75695,0.7860599999999972,0.7897825992063481,0.8068288099196645,0.79286346457412
|
| 502 |
+
0.07112375533428165,100,0.787525,0.84765,0.872825,0.787525,0.787525,0.28255,0.84765,0.174565,0.872825,0.787525,0.8200616666666627,0.8242667361111095,0.8434021080706371,0.8272603316013856
|
| 503 |
+
0.1422475106685633,200,0.800875,0.862875,0.887625,0.800875,0.800875,0.287625,0.862875,0.177525,0.887625,0.800875,0.8341379166666631,0.8382101091269837,0.8573139451618763,0.840874494584925
|
| 504 |
+
0.21337126600284495,300,0.796675,0.857525,0.88125,0.796675,0.796675,0.2858416666666666,0.857525,0.17625,0.88125,0.796675,0.8291537499999944,0.8330430654761858,0.8516132915919113,0.8357202570608268
|
| 505 |
+
0.2844950213371266,400,0.7899,0.848,0.8712,0.7899,0.7899,0.2826666666666666,0.848,0.17424,0.8712,0.7899,0.8210562499999958,0.8248586706349186,0.8427570230120356,0.8275746227632077
|
| 506 |
+
0.35561877667140823,500,0.78435,0.84085,0.863225,0.78435,0.78435,0.28028333333333333,0.84085,0.172645,0.863225,0.78435,0.8145704166666621,0.8182268749999977,0.8355804330527246,0.8209647560975787
|
| 507 |
+
0.4267425320056899,600,0.7802,0.8361,0.8576,0.7802,0.7802,0.27869999999999995,0.8361,0.17152,0.8576,0.7802,0.8098908333333292,0.8135463492063476,0.830654625279805,0.8163927962763059
|
| 508 |
+
0.49786628733997157,700,0.777175,0.832,0.85325,0.777175,0.777175,0.2773333333333334,0.832,0.17065,0.85325,0.777175,0.80638208333333,0.810140039682537,0.8272090556062501,0.8130014044089133
|
| 509 |
+
0.5689900426742532,800,0.7758,0.829675,0.850625,0.7758,0.7758,0.27655833333333324,0.829675,0.17012500000000005,0.850625,0.7758,0.8044174999999962,0.8081950198412682,0.8250726167267949,0.8111216889217711
|
| 510 |
+
0.6401137980085349,900,0.77365,0.827625,0.84905,0.77365,0.77365,0.275875,0.827625,0.16981000000000002,0.84905,0.77365,0.8023862499999959,0.8060900198412684,0.8229627553948096,0.8090680588419786
|
| 511 |
+
0.7112375533428165,1000,0.772425,0.826775,0.84795,0.772425,0.772425,0.2755916666666667,0.826775,0.16959000000000002,0.84795,0.772425,0.8013654166666629,0.8051413888888874,0.8221576910797325,0.808084264541656
|
| 512 |
+
0.7823613086770982,1100,0.771925,0.824975,0.8465,0.771925,0.771925,0.27499166666666663,0.824975,0.1693,0.8465,0.771925,0.800552916666664,0.8043763392857131,0.8212707180610094,0.8073654553559804
|
| 513 |
+
0.8534850640113798,1200,0.771025,0.8245,0.8465,0.771025,0.771025,0.2748333333333333,0.8245,0.1693,0.8465,0.771025,0.7997141666666631,0.8034813194444425,0.8205412738249844,0.8064654982641106
|
| 514 |
+
0.9246088193456614,1300,0.770125,0.823975,0.845875,0.770125,0.770125,0.27465833333333334,0.823975,0.16917500000000005,0.845875,0.770125,0.7990629166666633,0.8028743353174591,0.819989477253076,0.8059157176756445
|
| 515 |
+
0.9957325746799431,1400,0.770025,0.823775,0.845675,0.770025,0.770025,0.2745916666666667,0.823775,0.169135,0.845675,0.770025,0.7988824999999963,0.8026089186507919,0.8195745782328137,0.8057224194613716
|
| 516 |
+
1.0668563300142249,1500,0.76935,0.823125,0.845025,0.76935,0.76935,0.274375,0.823125,0.16900500000000002,0.845025,0.76935,0.7982599999999963,0.8021238492063483,0.8192664063697898,0.8052476907087441
|
| 517 |
+
1.1379800853485065,1600,0.769675,0.824125,0.8459,0.769675,0.769675,0.27470833333333333,0.824125,0.16918000000000002,0.8459,0.769675,0.7988324999999966,0.802651726190475,0.8198127616171903,0.8057819627069863
|
| 518 |
+
1.209103840682788,1700,0.769925,0.82335,0.845475,0.769925,0.769925,0.27444999999999997,0.82335,0.169095,0.845475,0.769925,0.7985574999999965,0.8024133928571431,0.8195806889185923,0.8055723826945884
|
| 519 |
+
1.2802275960170697,1800,0.76945,0.823225,0.845675,0.76945,0.76945,0.2744083333333333,0.823225,0.16913499999999998,0.845675,0.76945,0.7984820833333308,0.8023268353174606,0.819532061273888,0.8055067107223939
|
| 520 |
+
1.3513513513513513,1900,0.7691,0.82345,0.8454,0.7691,0.7691,0.27448333333333336,0.82345,0.16908,0.8454,0.7691,0.7982649999999966,0.8022191071428562,0.8196089924598134,0.805361535028667
|
| 521 |
+
1.422475106685633,2000,0.76915,0.82355,0.84555,0.76915,0.76915,0.27451666666666663,0.82355,0.16911,0.84555,0.76915,0.7983374999999966,0.8022573412698407,0.8196429775778874,0.8054493205736073
|
| 522 |
+
1.4935988620199145,2100,0.768975,0.82265,0.846275,0.768975,0.768975,0.27421666666666666,0.82265,0.16925500000000002,0.846275,0.768975,0.7982549999999972,0.8021490277777785,0.8196726605111274,0.8052885037093309
|
| 523 |
+
1.5647226173541964,2200,0.76885,0.82275,0.845525,0.76885,0.76885,0.27425,0.82275,0.169105,0.845525,0.76885,0.7980941666666648,0.8021060019841282,0.8196099718058493,0.8052948246443948
|
| 524 |
+
1.635846372688478,2300,0.769525,0.822775,0.84605,0.769525,0.769525,0.2742583333333333,0.822775,0.16921000000000003,0.84605,0.769525,0.7986262499999974,0.8027209226190485,0.8203517201642737,0.8058530753943357
|
| 525 |
+
1.7069701280227596,2400,0.7692,0.82385,0.846825,0.7692,0.7692,0.2746166666666667,0.82385,0.16936500000000002,0.846825,0.7692,0.7988820833333301,0.8029087698412697,0.8205894446265792,0.8060693271840575
|
| 526 |
+
1.7780938833570412,2500,0.76915,0.823725,0.846075,0.76915,0.76915,0.27457499999999996,0.823725,0.169215,0.846075,0.76915,0.7985683333333297,0.8026579761904753,0.8203458536798922,0.8058257617189151
|
| 527 |
+
1.8492176386913228,2600,0.769625,0.8246,0.84655,0.769625,0.769625,0.2748666666666666,0.8246,0.16931000000000002,0.84655,0.769625,0.7989349999999956,0.8030517063492055,0.8208079574534372,0.8062010921954896
|
| 528 |
+
1.9203413940256047,2700,0.769675,0.82405,0.8468,0.769675,0.769675,0.2746833333333333,0.82405,0.16936,0.8468,0.769675,0.7990249999999968,0.8030730158730155,0.8207917590552776,0.8062321544716866
|
| 529 |
+
1.991465149359886,2800,0.769475,0.8239,0.8468,0.769475,0.769475,0.27463333333333334,0.8239,0.16936,0.8468,0.769475,0.7988862499999965,0.8029730853174593,0.8208163032289425,0.8061114050965597
|
| 530 |
+
2.062588904694168,2900,0.769525,0.824125,0.846975,0.769525,0.769525,0.2747083333333333,0.824125,0.16939500000000002,0.846975,0.769525,0.7990287499999966,0.8031705853174578,0.821090801331755,0.8063063194045639
|
| 531 |
+
2.1337126600284497,3000,0.769025,0.8238,0.846525,0.769025,0.769025,0.2746,0.8238,0.169305,0.846525,0.769025,0.7986137499999971,0.8027588888888897,0.8206359977622566,0.8059580410743369
|
| 532 |
+
2.204836415362731,3100,0.7688,0.8244,0.846275,0.7688,0.7688,0.2748,0.8244,0.169255,0.846275,0.7688,0.798440833333329,0.8026224603174588,0.820558182322503,0.8058211991919103
|
| 533 |
+
2.275960170697013,3200,0.769,0.824375,0.847525,0.769,0.769,0.27479166666666666,0.824375,0.169505,0.847525,0.769,0.7988995833333298,0.803090228174601,0.821212946340923,0.8062446723184936
|
| 534 |
+
2.3470839260312943,3300,0.768775,0.8236,0.84665,0.768775,0.768775,0.2745333333333333,0.8236,0.16933,0.84665,0.768775,0.7983658333333298,0.8026019444444423,0.8207321092229084,0.8057709182568787
|
| 535 |
+
2.418207681365576,3400,0.7693,0.824175,0.847725,0.7693,0.7693,0.27472499999999994,0.824175,0.16954500000000003,0.847725,0.7693,0.79904708333333,0.8031713095238084,0.8212219006875497,0.80638920392551
|
| 536 |
+
2.4893314366998576,3500,0.769,0.823825,0.846725,0.769,0.769,0.2746083333333333,0.823825,0.169345,0.846725,0.769,0.7985754166666633,0.8027596726190457,0.8207592889256521,0.8059864347671793
|
| 537 |
+
2.5604551920341394,3600,0.76885,0.82395,0.8472,0.76885,0.76885,0.27464999999999995,0.82395,0.16944,0.8472,0.76885,0.7986866666666633,0.8028899107142835,0.8210598982328613,0.8060860038886773
|
| 538 |
+
2.6315789473684212,3700,0.768975,0.82375,0.8473,0.768975,0.768975,0.2745833333333333,0.82375,0.16946000000000003,0.8473,0.768975,0.7987062499999968,0.8029549503968239,0.8212211242661835,0.8061437513212889
|
| 539 |
+
2.7027027027027026,3800,0.768475,0.823675,0.84715,0.768475,0.768475,0.27455833333333335,0.823675,0.16943000000000003,0.84715,0.768475,0.7984670833333309,0.8026703373015858,0.8208454989370393,0.805943199903395
|
| 540 |
+
2.7738264580369845,3900,0.76845,0.824325,0.84735,0.76845,0.76845,0.27477499999999994,0.824325,0.16946999999999998,0.84735,0.76845,0.7985499999999965,0.8027500992063477,0.8209005123089202,0.8060343590369505
|
| 541 |
+
0,0,0.754925,0.807525,0.8308,0.754925,0.754925,0.269175,0.807525,0.16616,0.8308,0.754925,0.7834312499999965,0.7873935019841247,0.8048432733807716,0.7904811958974799
|
| 542 |
+
0.07112375533428165,100,0.784875,0.8462,0.8709,0.784875,0.784875,0.28206666666666663,0.8462,0.17418000000000003,0.8709,0.784875,0.8176641666666615,0.8220572023809489,0.8415809580549388,0.8251072204917347
|
| 543 |
+
0.1422475106685633,200,0.798075,0.861975,0.885725,0.798075,0.798075,0.2873249999999999,0.861975,0.17714500000000002,0.885725,0.798075,0.8317829166666606,0.8360811706349159,0.8556872478966354,0.8388044601703539
|
| 544 |
+
0.21337126600284495,300,0.793925,0.8557,0.88025,0.793925,0.793925,0.2852333333333333,0.8557,0.17605,0.88025,0.793925,0.8268337499999948,0.8308948511904715,0.8500208698553641,0.8336162172452645
|
| 545 |
+
0.2844950213371266,400,0.786425,0.8466,0.870625,0.786425,0.786425,0.2822,0.8466,0.174125,0.870625,0.786425,0.8185116666666619,0.8224112103174577,0.8409779740110344,0.8251187590711554
|
| 546 |
+
0.35561877667140823,500,0.780775,0.838675,0.8622,0.780775,0.780775,0.2795583333333333,0.838675,0.17244,0.8622,0.780775,0.811867083333329,0.815695327380951,0.8336883141924025,0.8184760373174987
|
| 547 |
+
0.4267425320056899,600,0.77655,0.833175,0.856,0.77655,0.77655,0.277725,0.833175,0.1712,0.856,0.77655,0.807010833333329,0.8109420535714272,0.82873317460253,0.8137810123209752
|
| 548 |
+
0,0,0.75535,0.8081,0.8314,0.75535,0.75535,0.26936666666666664,0.8081,0.16628,0.8314,0.75535,0.7840524999999964,0.7878348511904746,0.8050188990447376,0.7908941824856313
|
| 549 |
+
0.07112375533428165,100,0.78485,0.845025,0.8707,0.78485,0.78485,0.28167499999999995,0.845025,0.17414000000000002,0.8707,0.78485,0.8174049999999953,0.8215923908730137,0.8408518209280792,0.824681932921971
|
| 550 |
+
0.1422475106685633,200,0.797725,0.860375,0.885575,0.797725,0.797725,0.28679166666666667,0.860375,0.17711500000000002,0.885575,0.797725,0.831151666666662,0.8353726289682505,0.85499622100629,0.8380966938179938
|
| 551 |
+
0.21337126600284495,300,0.794025,0.85545,0.88,0.794025,0.794025,0.28514999999999996,0.85545,0.17600000000000005,0.88,0.794025,0.826853749999995,0.830870674603171,0.8498411806165926,0.8336155787941057
|
| 552 |
+
0.2844950213371266,400,0.7868,0.84585,0.87015,0.7868,0.7868,0.28195,0.84585,0.17403,0.87015,0.7868,0.8186045833333289,0.8224543551587292,0.840768411489289,0.8252239450411709
|
| 553 |
+
0.35561877667140823,500,0.78115,0.838625,0.86195,0.78115,0.78115,0.27954166666666663,0.838625,0.17239000000000004,0.86195,0.78115,0.8119908333333283,0.8158102579365052,0.8337259437115861,0.8185528436776975
|
| 554 |
+
0,0,0.75485,0.807325,0.8301,0.75485,0.75485,0.26910833333333334,0.807325,0.16602000000000003,0.8301,0.75485,0.7832499999999958,0.7872831249999973,0.8046637231378345,0.7903968386415016
|
| 555 |
+
0.07112375533428165,100,0.7846,0.845875,0.870475,0.7846,0.7846,0.2819583333333333,0.845875,0.17409500000000003,0.870475,0.7846,0.8175220833333285,0.8218874404761879,0.8412730946280681,0.8249953488463487
|
| 556 |
+
0.1422475106685633,200,0.7991,0.86235,0.88605,0.7991,0.7991,0.28745,0.86235,0.17721,0.88605,0.7991,0.8323445833333273,0.8366090079365022,0.856103705751086,0.8393309442404088
|
| 557 |
+
0.21337126600284495,300,0.794425,0.85565,0.87995,0.794425,0.794425,0.2852166666666666,0.85565,0.17599,0.87995,0.794425,0.8270145833333279,0.8310784623015823,0.8500951169586342,0.8337921512636404
|
| 558 |
+
0,0,0.75485,0.807325,0.8301,0.75485,0.75485,0.26910833333333334,0.807325,0.16602000000000003,0.8301,0.75485,0.783244583333329,0.7872755357142831,0.8046476992140257,0.7903936979751335
|
| 559 |
+
0.2849002849002849,100,0.798025,0.8616,0.886625,0.798025,0.798025,0.2872,0.8616,0.177325,0.886625,0.798025,0.8318458333333274,0.8360672718253934,0.8557194444498546,0.8388601667645773
|
| 560 |
+
0.5698005698005698,200,0.798625,0.860475,0.885875,0.798625,0.798625,0.28682499999999994,0.860475,0.177175,0.885875,0.798625,0.8319029166666622,0.8359366567460295,0.8551078160914656,0.8387170552876636
|
| 561 |
+
0.8547008547008547,300,0.7974,0.8592,0.884475,0.7974,0.7974,0.28639999999999993,0.8592,0.176895,0.884475,0.7974,0.830667916666662,0.8349036011904735,0.8543353748755151,0.8377224912323109
|
| 562 |
+
1.1396011396011396,400,0.79655,0.8598,0.8852,0.79655,0.79655,0.28659999999999997,0.8598,0.17704,0.8852,0.79655,0.8303362499999941,0.8345951587301531,0.8543686412006678,0.8374535585472126
|
| 563 |
+
1.4245014245014245,500,0.797225,0.860075,0.8858,0.797225,0.797225,0.2866916666666667,0.860075,0.17716,0.8858,0.797225,0.8309320833333282,0.8352305456349177,0.8550016319959118,0.8381492588281301
|
| 564 |
+
1.7094017094017095,600,0.797725,0.86015,0.88665,0.797725,0.797725,0.2867166666666666,0.86015,0.17733000000000002,0.88665,0.797725,0.8313991666666614,0.8356998115079319,0.8555999734107165,0.8385892139129612
|
| 565 |
+
1.9943019943019942,700,0.797975,0.861325,0.887325,0.797975,0.797975,0.2871083333333333,0.861325,0.177465,0.887325,0.797975,0.8319016666666612,0.8362617460317411,0.8562739250749866,0.8391815253696681
|
| 566 |
+
2.2792022792022792,800,0.79835,0.86145,0.888175,0.79835,0.79835,0.28715,0.86145,0.17763500000000002,0.888175,0.79835,0.8324391666666625,0.8366843948412661,0.8566247944874054,0.8396564176781772
|
| 567 |
+
2.564102564102564,900,0.798375,0.862075,0.888225,0.798375,0.798375,0.2873583333333333,0.862075,0.17764500000000003,0.888225,0.798375,0.832563333333328,0.8369297420634868,0.8570087384867151,0.8398948290084045
|
| 568 |
+
2.849002849002849,1000,0.7987,0.8623,0.88935,0.7987,0.7987,0.28743333333333326,0.8623,0.17787,0.88935,0.7987,0.8329933333333287,0.8371977083333292,0.8572161454797611,0.8402122967157111
|
| 569 |
+
3.133903133903134,1100,0.7986,0.862325,0.88905,0.7986,0.7986,0.2874416666666666,0.862325,0.17781000000000002,0.88905,0.7986,0.8329541666666616,0.8373066468253925,0.8574478167950113,0.8403082974896099
|
| 570 |
+
3.4188034188034186,1200,0.799125,0.8628,0.889525,0.799125,0.799125,0.2876,0.8628,0.177905,0.889525,0.799125,0.8334449999999944,0.8377979365079314,0.8579219826325345,0.8408453839836647
|
| 571 |
+
3.7037037037037037,1300,0.7992,0.862875,0.890425,0.7992,0.7992,0.28762499999999996,0.862875,0.17808500000000005,0.890425,0.7992,0.8336524999999955,0.8379600992063455,0.8582124684478113,0.8409858197750711
|
| 572 |
+
3.9886039886039883,1400,0.7994,0.86345,0.891,0.7994,0.7994,0.2878166666666666,0.86345,0.17820000000000003,0.891,0.7994,0.8340516666666619,0.8383523908730119,0.8586412473927562,0.8413785327956702
|
| 573 |
+
4.273504273504273,1500,0.799825,0.863575,0.890875,0.799825,0.799825,0.2878583333333333,0.863575,0.178175,0.890875,0.799825,0.834187083333328,0.8385845238095195,0.8589450124547461,0.8415957957257291
|
| 574 |
+
4.5584045584045585,1600,0.7999,0.864025,0.89135,0.7999,0.7999,0.2880083333333333,0.864025,0.17827000000000004,0.89135,0.7999,0.8344108333333281,0.8388025892857096,0.8592482613458784,0.841788283131001
|
| 575 |
+
4.843304843304844,1700,0.799775,0.8643,0.891475,0.799775,0.799775,0.28809999999999997,0.8643,0.17829500000000004,0.891475,0.799775,0.8345333333333287,0.8389044543650747,0.8593056239763895,0.8419345783918155
|
| 576 |
+
5.128205128205128,1800,0.800375,0.86435,0.8917,0.800375,0.800375,0.28811666666666663,0.86435,0.17834,0.8917,0.800375,0.8348483333333283,0.8392823412698379,0.8596923791808068,0.8423084831810096
|
| 577 |
+
5.413105413105413,1900,0.800275,0.86425,0.892,0.800275,0.800275,0.2880833333333333,0.86425,0.17840000000000006,0.892,0.800275,0.8349387499999955,0.8393679861111072,0.8598629768476423,0.8423678921112896
|
| 578 |
+
5.698005698005698,2000,0.80055,0.8648,0.892175,0.80055,0.80055,0.2882666666666666,0.8648,0.178435,0.892175,0.80055,0.8351624999999946,0.8395857440476139,0.8600660936618119,0.8426103480001628
|
| 579 |
+
5.982905982905983,2100,0.80115,0.864925,0.89215,0.80115,0.80115,0.2883083333333333,0.864925,0.17843,0.89215,0.80115,0.8355258333333282,0.8399495039682501,0.8603280903031238,0.8429954225226167
|
| 580 |
+
6.267806267806268,2200,0.801225,0.865025,0.892425,0.801225,0.801225,0.28834166666666666,0.865025,0.178485,0.892425,0.801225,0.835632499999995,0.8400607440476151,0.8605022088940468,0.8430806763267096
|
| 581 |
+
6.552706552706553,2300,0.801275,0.865075,0.892575,0.801275,0.801275,0.2883583333333333,0.865075,0.178515,0.892575,0.801275,0.8357724999999949,0.8402953373015829,0.8608270909507948,0.8433080921002089
|
| 582 |
+
6.837606837606837,2400,0.801575,0.865125,0.8924,0.801575,0.801575,0.288375,0.865125,0.17848000000000003,0.8924,0.801575,0.8358745833333283,0.840366349206346,0.8607774883891669,0.8434245904693333
|
| 583 |
+
7.122507122507122,2500,0.80125,0.865,0.892525,0.80125,0.80125,0.2883333333333333,0.865,0.17850500000000002,0.892525,0.80125,0.8357508333333294,0.8403010317460295,0.8608727881432594,0.843316427466535
|
| 584 |
+
7.407407407407407,2600,0.8017,0.86565,0.893325,0.8017,0.8017,0.28855,0.86565,0.17866500000000002,0.893325,0.8017,0.8362083333333287,0.8406969047619021,0.8612783749710475,0.843708366864761
|
| 585 |
+
7.6923076923076925,2700,0.8017,0.86565,0.89265,0.8017,0.8017,0.28855,0.86565,0.17853000000000002,0.89265,0.8017,0.8360970833333286,0.8406737103174581,0.8612256926854687,0.8436828252043238
|
| 586 |
+
7.977207977207978,2800,0.801675,0.865975,0.8929,0.801675,0.801675,0.2886583333333333,0.865975,0.17858000000000002,0.8929,0.801675,0.8361433333333281,0.8407370833333302,0.8613825554424442,0.8437267421268331
|
| 587 |
+
8.262108262108262,2900,0.801675,0.866225,0.8937,0.801675,0.801675,0.2887416666666667,0.866225,0.17874,0.8937,0.801675,0.8364137499999955,0.84092972222222,0.8616041067360869,0.8439152706764873
|
| 588 |
+
8.547008547008547,3000,0.801875,0.86645,0.893725,0.801875,0.801875,0.2888166666666666,0.86645,0.17874500000000001,0.893725,0.801875,0.8365145833333281,0.8410595634920605,0.8617667237853555,0.844030886025032
|
| 589 |
+
8.831908831908832,3100,0.8023,0.86675,0.8937,0.8023,0.8023,0.2889166666666666,0.86675,0.17874,0.8937,0.8023,0.8368633333333286,0.8413887301587282,0.861958616771547,0.8443842822305276
|
| 590 |
+
9.116809116809117,3200,0.80215,0.8668,0.89375,0.80215,0.80215,0.2889333333333333,0.8668,0.17874999999999996,0.89375,0.80215,0.8368420833333284,0.8413961408730133,0.8620346988055472,0.8443798484797
|
| 591 |
+
9.401709401709402,3300,0.801675,0.8668,0.893875,0.801675,0.801675,0.2889333333333333,0.8668,0.17877500000000002,0.893875,0.801675,0.8366370833333286,0.8411731646825378,0.8618878022511639,0.8441485770363238
|
| 592 |
+
9.686609686609687,3400,0.802025,0.867075,0.89435,0.802025,0.802025,0.2890249999999999,0.867075,0.17887000000000003,0.89435,0.802025,0.8369008333333284,0.841364017857141,0.8620247212678003,0.8443709033760651
|
| 593 |
+
9.971509971509972,3500,0.802225,0.867275,0.89445,0.802225,0.802225,0.2890916666666666,0.867275,0.17889000000000002,0.89445,0.802225,0.8371345833333286,0.8416379861111096,0.8623100216286935,0.8446182487542423
|
| 594 |
+
10.256410256410255,3600,0.80225,0.8672,0.8945,0.80225,0.80225,0.28906666666666664,0.8672,0.17890000000000003,0.8945,0.80225,0.837151249999995,0.8416526091269821,0.8623240909951686,0.8446431234027594
|
| 595 |
+
10.54131054131054,3700,0.802125,0.86725,0.894475,0.802125,0.802125,0.2890833333333333,0.86725,0.17889500000000003,0.894475,0.802125,0.8371087499999949,0.8416043948412673,0.8622908336950419,0.844597619678181
|
| 596 |
+
10.826210826210826,3800,0.8023,0.867175,0.8947,0.8023,0.8023,0.2890583333333333,0.867175,0.17894000000000002,0.8947,0.8023,0.8372020833333284,0.841680109126981,0.8623660731654591,0.8446758053360729
|
| 597 |
+
11.11111111111111,3900,0.8024,0.866975,0.8947,0.8024,0.8024,0.2889916666666666,0.866975,0.17894000000000002,0.8947,0.8024,0.8372370833333284,0.8417220138888862,0.8624066595516806,0.8447202630510894
|
| 598 |
+
11.396011396011396,4000,0.802475,0.867225,0.8946,0.802475,0.802475,0.289075,0.867225,0.17892,0.8946,0.802475,0.8373216666666616,0.8418482242063464,0.8625582027110184,0.84483463585437
|
| 599 |
+
11.68091168091168,4100,0.802425,0.867075,0.894575,0.802425,0.802425,0.289025,0.867075,0.17891500000000002,0.894575,0.802425,0.837263749999995,0.8417595833333303,0.8624108110281115,0.844785572665587
|
| 600 |
+
11.965811965811966,4200,0.802525,0.86745,0.894625,0.802525,0.802525,0.28914999999999996,0.86745,0.17892500000000003,0.894625,0.802525,0.8373899999999946,0.8418941269841238,0.8625485306060944,0.8449125742356473
|
| 601 |
+
12.250712250712251,4300,0.802425,0.867425,0.8949,0.802425,0.802425,0.28914166666666663,0.867425,0.17898,0.8949,0.802425,0.8373679166666612,0.8418560416666631,0.862570562826382,0.8448589991636967
|
| 602 |
+
12.535612535612536,4400,0.802475,0.8674,0.8947,0.802475,0.802475,0.2891333333333333,0.8674,0.17894000000000004,0.8947,0.802475,0.8373841666666616,0.8419126488095208,0.8626053598219531,0.8449282197722164
|
| 603 |
+
12.820512820512821,4500,0.802325,0.86755,0.89495,0.802325,0.802325,0.2891833333333333,0.86755,0.17899,0.89495,0.802325,0.8373616666666612,0.8418324404761867,0.8625349588157627,0.8448507681783248
|
| 604 |
+
13.105413105413106,4600,0.80245,0.867575,0.8949,0.80245,0.80245,0.2891916666666666,0.867575,0.17898000000000003,0.8949,0.80245,0.8374437499999949,0.8419402380952342,0.8626571238261973,0.8449448072816027
|
| 605 |
+
13.39031339031339,4700,0.80245,0.867575,0.894825,0.80245,0.80245,0.2891916666666666,0.867575,0.178965,0.894825,0.80245,0.8374074999999948,0.841924206349203,0.8626414521760948,0.8449332980096382
|
| 606 |
+
13.675213675213675,4800,0.8024,0.867775,0.8947,0.8024,0.8024,0.2892583333333333,0.867775,0.17894000000000004,0.8947,0.8024,0.837397083333328,0.8419506051587267,0.8627007599790489,0.844947246350743
|
| 607 |
+
13.96011396011396,4900,0.802475,0.867625,0.8947,0.802475,0.802475,0.2892083333333333,0.867625,0.17894,0.8947,0.802475,0.8374054166666614,0.8419520734126953,0.8626792917708476,0.8449600866335961
|
| 608 |
+
14.245014245014245,5000,0.802275,0.8676,0.89495,0.802275,0.802275,0.28919999999999996,0.8676,0.17899,0.89495,0.802275,0.8373504166666614,0.8418555952380918,0.8626142815266079,0.8448604581747029
|
final_metrics.json
CHANGED
|
@@ -1,16 +1,16 @@
|
|
| 1 |
{
|
| 2 |
-
"val_cosine_accuracy@1": 0.
|
| 3 |
-
"val_cosine_accuracy@3": 0.
|
| 4 |
-
"val_cosine_accuracy@5": 0.
|
| 5 |
-
"val_cosine_precision@1": 0.
|
| 6 |
-
"val_cosine_precision@3": 0.
|
| 7 |
-
"val_cosine_precision@5": 0.
|
| 8 |
-
"val_cosine_recall@1": 0.
|
| 9 |
-
"val_cosine_recall@3": 0.
|
| 10 |
-
"val_cosine_recall@5": 0.
|
| 11 |
-
"val_cosine_ndcg@10": 0.
|
| 12 |
-
"val_cosine_mrr@1": 0.
|
| 13 |
-
"val_cosine_mrr@5": 0.
|
| 14 |
-
"val_cosine_mrr@10": 0.
|
| 15 |
-
"val_cosine_map@100": 0.
|
| 16 |
}
|
|
|
|
| 1 |
{
|
| 2 |
+
"val_cosine_accuracy@1": 0.76415,
|
| 3 |
+
"val_cosine_accuracy@3": 0.822175,
|
| 4 |
+
"val_cosine_accuracy@5": 0.84665,
|
| 5 |
+
"val_cosine_precision@1": 0.76415,
|
| 6 |
+
"val_cosine_precision@3": 0.2740583333333333,
|
| 7 |
+
"val_cosine_precision@5": 0.16933,
|
| 8 |
+
"val_cosine_recall@1": 0.76415,
|
| 9 |
+
"val_cosine_recall@3": 0.822175,
|
| 10 |
+
"val_cosine_recall@5": 0.84665,
|
| 11 |
+
"val_cosine_ndcg@10": 0.8195085862842543,
|
| 12 |
+
"val_cosine_mrr@1": 0.76415,
|
| 13 |
+
"val_cosine_mrr@5": 0.7953370833333295,
|
| 14 |
+
"val_cosine_mrr@10": 0.8000376587301573,
|
| 15 |
+
"val_cosine_map@100": 0.8034052122492592
|
| 16 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 114011616
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:45cdb91cc553556968b278c530594e05d926d8a2c99d01838656167e2114a507
|
| 3 |
size 114011616
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 6161
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7258c121488b2e4966f22215c04593a6fcbb89f836828ccbf066defc2eda8b93
|
| 3 |
size 6161
|