AugTriever-Hybrid-All / eval_results.txt
memray's picture
Upload 130 files
07c3c61
eval/beir-arguana_ndcg@10 = 0.37955
eval/beir-arguana_recall@100 = 0.93457
eval/beir-avg_ndcg@10 = 0.376576
eval/beir-avg_recall@10 = 0.440174
eval/beir-avg_recall@100 = 0.6589700000000001
eval/beir-avg_recall@20 = 0.506552
eval/beir-climate-fever_ndcg@10 = 0.17477
eval/beir-climate-fever_recall@10 = 0.21757
eval/beir-climate-fever_recall@100 = 0.47127
eval/beir-climate-fever_recall@20 = 0.27878
eval/beir-cqadupstack_ndcg@10 = 0.264705
eval/beir-cqadupstack_recall@100 = 0.6014158333333334
eval/beir-dbpedia-entity_ndcg@10 = 0.32317
eval/beir-dbpedia-entity_recall@10 = 0.21027
eval/beir-dbpedia-entity_recall@100 = 0.46921
eval/beir-dbpedia-entity_recall@20 = 0.27414
eval/beir-fever_ndcg@10 = 0.6139
eval/beir-fever_recall@10 = 0.79606
eval/beir-fever_recall@100 = 0.91454
eval/beir-fever_recall@20 = 0.85168
eval/beir-fiqa_ndcg@10 = 0.25245
eval/beir-fiqa_recall@100 = 0.57848
eval/beir-hotpotqa_ndcg@10 = 0.53618
eval/beir-hotpotqa_recall@10 = 0.57016
eval/beir-hotpotqa_recall@100 = 0.72073
eval/beir-hotpotqa_recall@20 = 0.61931
eval/beir-msmarco_ndcg@10 = 0.23486
eval/beir-msmarco_recall@10 = 0.40681
eval/beir-msmarco_recall@100 = 0.7191
eval/beir-msmarco_recall@20 = 0.50885
eval/beir-nfcorpus_ndcg@10 = 0.30262
eval/beir-nfcorpus_recall@100 = 0.28685
eval/beir-nq_ndcg@10 = 0.28918
eval/beir-nq_recall@100 = 0.81301
eval/beir-quora_ndcg@10 = 0.78744
eval/beir-quora_recall@100 = 0.97907
eval/beir-scidocs_ndcg@10 = 0.15287
eval/beir-scidocs_recall@100 = 0.37422
eval/beir-scifact_ndcg@10 = 0.65164
eval/beir-scifact_recall@100 = 0.93156
eval/beir-trec-covid_ndcg@10 = 0.58524
eval/beir-trec-covid_recall@100 = 0.11007
eval/beir-webis-touche2020_ndcg@10 = 0.19644
eval/beir-webis-touche2020_recall@100 = 0.43738
eval/qa-curatedtrec-test-acc@100 = 0.9322766570605188
eval/qa-curatedtrec-test-acc@20 = 0.8515850144092219
eval/qa-curatedtrec-test-acc@5 = 0.6801152737752162
eval/qa-entityqs-macro-acc@100 = 0.7733104099252927
eval/qa-entityqs-macro-acc@20 = 0.6646029778857252
eval/qa-entityqs-macro-acc@5 = 0.5346457057615871
eval/qa-nq-test-acc@100 = 0.8066481994459834
eval/qa-nq-test-acc@20 = 0.6642659279778393
eval/qa-nq-test-acc@5 = 0.4601108033240997
eval/qa-squad1-test-acc@100 = 0.7769157994323557
eval/qa-squad1-test-acc@20 = 0.6240302743614002
eval/qa-squad1-test-acc@5 = 0.43122043519394515
eval/qa-trivia-test-acc@100 = 0.8374436488994962
eval/qa-trivia-test-acc@20 = 0.7558560947582428
eval/qa-trivia-test-acc@5 = 0.629452841863343
eval/qa-webq-test-acc@100 = 0.8223425196850394
eval/qa-webq-test-acc@20 = 0.6938976377952756
eval/qa-webq-test-acc@5 = 0.4867125984251969
eval/senteval-CR = 87.16
eval/senteval-MPQA = 88.77
eval/senteval-MR = 81.02
eval/senteval-MRPC = 71.25
eval/senteval-SICKRelatedness = 0.6882116860582081
eval/senteval-SST2 = 83.6
eval/senteval-STS12 = 0.6088736294798399
eval/senteval-STS13 = 0.7464064251943163
eval/senteval-STS14 = 0.6678913458222988
eval/senteval-STS15 = 0.7877087174315147
eval/senteval-STS16 = 0.7885672061699577
eval/senteval-STSBenchmark = 0.7504771603185239
eval/senteval-SUBJ = 95.33
eval/senteval-TREC = 80.56
eval/senteval-avg_sts_7 = 0.7197337386392372
eval/senteval-avg_transfer = 83.9557142857143
train/global_step = 100000