| eval/beir-arguana_ndcg@10 = 0.37881 | |
| eval/beir-arguana_recall@100 = 0.9175 | |
| eval/beir-avg_ndcg@10 = 0.36036 | |
| eval/beir-avg_recall@10 = 0.42374800000000007 | |
| eval/beir-avg_recall@100 = 0.651238 | |
| eval/beir-avg_recall@20 = 0.49372999999999995 | |
| eval/beir-climate-fever_ndcg@10 = 0.17148 | |
| eval/beir-climate-fever_recall@10 = 0.21584 | |
| eval/beir-climate-fever_recall@100 = 0.47004 | |
| eval/beir-climate-fever_recall@20 = 0.28686 | |
| eval/beir-cqadupstack_ndcg@10 = 0.2803316666666667 | |
| eval/beir-cqadupstack_recall@100 = 0.604765 | |
| eval/beir-dbpedia-entity_ndcg@10 = 0.30952 | |
| eval/beir-dbpedia-entity_recall@10 = 0.21 | |
| eval/beir-dbpedia-entity_recall@100 = 0.47187 | |
| eval/beir-dbpedia-entity_recall@20 = 0.2695 | |
| eval/beir-fever_ndcg@10 = 0.59083 | |
| eval/beir-fever_recall@10 = 0.77263 | |
| eval/beir-fever_recall@100 = 0.90212 | |
| eval/beir-fever_recall@20 = 0.82815 | |
| eval/beir-fiqa_ndcg@10 = 0.25271 | |
| eval/beir-fiqa_recall@100 = 0.57736 | |
| eval/beir-hotpotqa_ndcg@10 = 0.51242 | |
| eval/beir-hotpotqa_recall@10 = 0.54774 | |
| eval/beir-hotpotqa_recall@100 = 0.71269 | |
| eval/beir-hotpotqa_recall@20 = 0.60459 | |
| eval/beir-msmarco_ndcg@10 = 0.21755 | |
| eval/beir-msmarco_recall@10 = 0.37253 | |
| eval/beir-msmarco_recall@100 = 0.69947 | |
| eval/beir-msmarco_recall@20 = 0.47955 | |
| eval/beir-nfcorpus_ndcg@10 = 0.29721 | |
| eval/beir-nfcorpus_recall@100 = 0.27314 | |
| eval/beir-nq_ndcg@10 = 0.28932 | |
| eval/beir-nq_recall@100 = 0.80159 | |
| eval/beir-quora_ndcg@10 = 0.81279 | |
| eval/beir-quora_recall@100 = 0.98333 | |
| eval/beir-scidocs_ndcg@10 = 0.1582 | |
| eval/beir-scidocs_recall@100 = 0.37162 | |
| eval/beir-scifact_ndcg@10 = 0.63462 | |
| eval/beir-scifact_recall@100 = 0.90322 | |
| eval/beir-trec-covid_ndcg@10 = 0.5527 | |
| eval/beir-trec-covid_recall@100 = 0.10295 | |
| eval/beir-webis-touche2020_ndcg@10 = 0.18394 | |
| eval/beir-webis-touche2020_recall@100 = 0.44207 | |
| eval/qa-curatedtrec-test-acc@100 = 0.9351585014409222 | |
| eval/qa-curatedtrec-test-acc@20 = 0.8515850144092219 | |
| eval/qa-curatedtrec-test-acc@5 = 0.7046109510086456 | |
| eval/qa-entityqs-macro-acc@100 = 0.7602876877587357 | |
| eval/qa-entityqs-macro-acc@20 = 0.6447794141198262 | |
| eval/qa-entityqs-macro-acc@5 = 0.5100471129421901 | |
| eval/qa-nq-test-acc@100 = 0.8013850415512466 | |
| eval/qa-nq-test-acc@20 = 0.6750692520775623 | |
| eval/qa-nq-test-acc@5 = 0.4692520775623269 | |
| eval/qa-squad1-test-acc@100 = 0.7788079470198676 | |
| eval/qa-squad1-test-acc@20 = 0.6403973509933775 | |
| eval/qa-squad1-test-acc@5 = 0.4489120151371807 | |
| eval/qa-trivia-test-acc@100 = 0.8371784672500663 | |
| eval/qa-trivia-test-acc@20 = 0.7591266684345443 | |
| eval/qa-trivia-test-acc@5 = 0.6233536639264563 | |
| eval/qa-webq-test-acc@100 = 0.8188976377952756 | |
| eval/qa-webq-test-acc@20 = 0.7052165354330708 | |
| eval/qa-webq-test-acc@5 = 0.5068897637795275 | |
| eval/senteval-CR = 87.01 | |
| eval/senteval-MPQA = 88.89 | |
| eval/senteval-MR = 81.09 | |
| eval/senteval-MRPC = 70.8 | |
| eval/senteval-SICKRelatedness = 0.7025121676824595 | |
| eval/senteval-SST2 = 84.06 | |
| eval/senteval-STS12 = 0.6198695741812278 | |
| eval/senteval-STS13 = 0.7458274512781141 | |
| eval/senteval-STS14 = 0.6864467990669687 | |
| eval/senteval-STS15 = 0.7986117512115241 | |
| eval/senteval-STS16 = 0.7823935956761426 | |
| eval/senteval-STSBenchmark = 0.7668425138024548 | |
| eval/senteval-SUBJ = 95.34 | |
| eval/senteval-TREC = 82.7 | |
| eval/senteval-avg_sts_7 = 0.7289291218426988 | |
| eval/senteval-avg_transfer = 84.27 | |
| train/global_step = 100000 | |