Spaces:
Running
Running
Add/update results for Llama 4 Scout 17B (version 1.33.1, guid ef91e41e46fd4351b30c5cf2e2e6f9fc)
Browse files- results.json +74 -0
results.json
CHANGED
|
@@ -450,6 +450,80 @@
|
|
| 450 |
"n_questions": 158,
|
| 451 |
"submit_timestamp": ""
|
| 452 |
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 453 |
}
|
| 454 |
}
|
| 455 |
},
|
|
|
|
| 450 |
"n_questions": 158,
|
| 451 |
"submit_timestamp": ""
|
| 452 |
}
|
| 453 |
+
},
|
| 454 |
+
"ef91e41e46fd4351b30c5cf2e2e6f9fc": {
|
| 455 |
+
"model_name": "Llama 4 Scout 17B",
|
| 456 |
+
"timestamp": "2025-05-22T14:47:39",
|
| 457 |
+
"config": {
|
| 458 |
+
"embedding_model": "sentence-transformers/distiluse-base-multilingual-cased",
|
| 459 |
+
"retriever_type": "mmr",
|
| 460 |
+
"retrieval_config": {
|
| 461 |
+
"top_k": 4,
|
| 462 |
+
"chunk_size": 1000,
|
| 463 |
+
"chunk_overlap": 100
|
| 464 |
+
}
|
| 465 |
+
},
|
| 466 |
+
"metrics": {
|
| 467 |
+
"mh": {
|
| 468 |
+
"retrieval": {
|
| 469 |
+
"hit_rate": 0.4791666666666667,
|
| 470 |
+
"mrr": 0.328125,
|
| 471 |
+
"precision": 0.13541666666666666
|
| 472 |
+
},
|
| 473 |
+
"generation": {
|
| 474 |
+
"rouge1": 0.04071969696969697,
|
| 475 |
+
"rougeL": 0.04071969696969697
|
| 476 |
+
}
|
| 477 |
+
},
|
| 478 |
+
"cond": {
|
| 479 |
+
"retrieval": {
|
| 480 |
+
"hit_rate": 0.5686274509803921,
|
| 481 |
+
"mrr": 0.43137254901960786,
|
| 482 |
+
"precision": 0.18627450980392157
|
| 483 |
+
},
|
| 484 |
+
"generation": {
|
| 485 |
+
"rouge1": 0.012118736383442266,
|
| 486 |
+
"rougeL": 0.012118736383442266
|
| 487 |
+
}
|
| 488 |
+
},
|
| 489 |
+
"simple": {
|
| 490 |
+
"retrieval": {
|
| 491 |
+
"hit_rate": 0.5555555555555556,
|
| 492 |
+
"mrr": 0.39351851851851843,
|
| 493 |
+
"precision": 0.1527777777777778
|
| 494 |
+
},
|
| 495 |
+
"generation": {
|
| 496 |
+
"rouge1": 0.04487118921642731,
|
| 497 |
+
"rougeL": 0.04487118921642731
|
| 498 |
+
}
|
| 499 |
+
},
|
| 500 |
+
"set": {
|
| 501 |
+
"retrieval": {
|
| 502 |
+
"hit_rate": 0.6,
|
| 503 |
+
"mrr": 0.25,
|
| 504 |
+
"precision": 0.2
|
| 505 |
+
},
|
| 506 |
+
"generation": {
|
| 507 |
+
"rouge1": 0.0,
|
| 508 |
+
"rougeL": 0.0
|
| 509 |
+
}
|
| 510 |
+
},
|
| 511 |
+
"overall": {
|
| 512 |
+
"retrieval": {
|
| 513 |
+
"hit_rate": 0.5379746835443038,
|
| 514 |
+
"mrr": 0.38132911392405067,
|
| 515 |
+
"precision": 0.15981012658227847
|
| 516 |
+
},
|
| 517 |
+
"generation": {
|
| 518 |
+
"rouge1": 0.03161800777081066,
|
| 519 |
+
"rougeL": 0.03161800777081066
|
| 520 |
+
}
|
| 521 |
+
}
|
| 522 |
+
},
|
| 523 |
+
"metadata": {
|
| 524 |
+
"n_questions": 158,
|
| 525 |
+
"submit_timestamp": ""
|
| 526 |
+
}
|
| 527 |
}
|
| 528 |
}
|
| 529 |
},
|