| eval/beir-arguana_ndcg@10 = 0.36665 | |
| eval/beir-arguana_recall@100 = 0.96728 | |
| eval/beir-avg_ndcg@10 = 0.30767 | |
| eval/beir-avg_recall@10 = 0.33533500000000005 | |
| eval/beir-avg_recall@100 = 0.523215 | |
| eval/beir-avg_recall@20 = 0.383875 | |
| eval/beir-climate-fever_ndcg@10 = 0.12105 | |
| eval/beir-climate-fever_recall@10 = 0.1548 | |
| eval/beir-climate-fever_recall@100 = 0.37817 | |
| eval/beir-climate-fever_recall@20 = 0.20165 | |
| eval/beir-cqadupstack_ndcg@10 = 0.3031158333333333 | |
| eval/beir-cqadupstack_recall@100 = 0.6389233333333334 | |
| eval/beir-dbpedia-entity_ndcg@10 = 0.3116 | |
| eval/beir-dbpedia-entity_recall@100 = 0.4652 | |
| eval/beir-fever_ndcg@10 = 0.63923 | |
| eval/beir-fever_recall@100 = 0.93145 | |
| eval/beir-fiqa_ndcg@10 = 0.28062 | |
| eval/beir-fiqa_recall@100 = 0.61765 | |
| eval/beir-hotpotqa_ndcg@10 = 0.49429 | |
| eval/beir-hotpotqa_recall@10 = 0.51587 | |
| eval/beir-hotpotqa_recall@100 = 0.66826 | |
| eval/beir-hotpotqa_recall@20 = 0.5661 | |
| eval/beir-msmarco_ndcg@10 = 0.20651 | |
| eval/beir-msmarco_recall@100 = 0.67481 | |
| eval/beir-nfcorpus_ndcg@10 = 0.34134 | |
| eval/beir-nfcorpus_recall@100 = 0.31962 | |
| eval/beir-nq_ndcg@10 = 0.27574 | |
| eval/beir-nq_recall@100 = 0.7927 | |
| eval/beir-quora_ndcg@10 = 0.78549 | |
| eval/beir-quora_recall@100 = 0.9752 | |
| eval/beir-scidocs_ndcg@10 = 0.15962 | |
| eval/beir-scidocs_recall@100 = 0.39283 | |
| eval/beir-scifact_ndcg@10 = 0.618 | |
| eval/beir-scifact_recall@100 = 0.911 | |
| eval/beir-trec-covid_ndcg@10 = 0.58244 | |
| eval/beir-trec-covid_recall@100 = 0.10597 | |
| eval/beir-webis-touche2020_ndcg@10 = 0.18565 | |
| eval/beir-webis-touche2020_recall@100 = 0.43843 | |
| eval/qa-curatedtrec-test-acc@100 = 0.9265129682997119 | |
| eval/qa-curatedtrec-test-acc@20 = 0.8371757925072046 | |
| eval/qa-curatedtrec-test-acc@5 = 0.6354466858789626 | |
| eval/qa-entityqs-macro-acc@100 = 0.7567939708496935 | |
| eval/qa-entityqs-macro-acc@20 = 0.6403777676502388 | |
| eval/qa-entityqs-macro-acc@5 = 0.49428236812619747 | |
| eval/qa-nq-test-acc@100 = 0.8024930747922437 | |
| eval/qa-nq-test-acc@20 = 0.6501385041551246 | |
| eval/qa-nq-test-acc@5 = 0.44155124653739614 | |
| eval/qa-squad1-test-acc@100 = 0.7376537369914853 | |
| eval/qa-squad1-test-acc@20 = 0.5704824976348155 | |
| eval/qa-squad1-test-acc@5 = 0.3709555345316935 | |
| eval/qa-trivia-test-acc@100 = 0.8229470520639972 | |
| eval/qa-trivia-test-acc@20 = 0.7227083885795103 | |
| eval/qa-trivia-test-acc@5 = 0.5689030319101919 | |
| eval/qa-webq-test-acc@100 = 0.8061023622047244 | |
| eval/qa-webq-test-acc@20 = 0.6520669291338582 | |
| eval/qa-webq-test-acc@5 = 0.40994094488188976 | |
| eval/senteval-CR = 87.02 | |
| eval/senteval-MPQA = 88.67 | |
| eval/senteval-MR = 80.75 | |
| eval/senteval-MRPC = 70.17 | |
| eval/senteval-SICKRelatedness = 0.7122486034563766 | |
| eval/senteval-SST2 = 82.8 | |
| eval/senteval-STS12 = 0.6520942930121344 | |
| eval/senteval-STS13 = 0.7489808796663362 | |
| eval/senteval-STS14 = 0.6214541243526541 | |
| eval/senteval-STS15 = 0.7497940183329284 | |
| eval/senteval-STS16 = 0.7490972681198437 | |
| eval/senteval-STSBenchmark = 0.708944371429387 | |
| eval/senteval-SUBJ = 95.02 | |
| eval/senteval-TREC = 78.63 | |
| eval/senteval-avg_sts_7 = 0.7060876511956659 | |
| eval/senteval-avg_transfer = 83.2942857142857 | |
| train/global_step = 100000 | |