diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 12a5260b212456dcf2c68f15e92274bb293fa859..5bea6cf6e51d8ba3c7d93ec27f4f7813474b9175 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,14 +1,14 @@ { "results": { "mmlu": { - "acc,none": 0.33107819398946015, - "acc_stderr,none": 0.06027611346350394, + "acc,none": 0.33086454920951425, + "acc_stderr,none": 0.060605075693583886, "alias": "mmlu" }, "mmlu_humanities": { "alias": " - humanities", - "acc,none": 0.32561105207226354, - "acc_stderr,none": 0.05639800096282077 + "acc,none": 0.32561105207226343, + "acc_stderr,none": 0.0607809456258765 }, "mmlu_formal_logic": { "alias": " - formal_logic", @@ -27,8 +27,8 @@ }, "mmlu_high_school_world_history": { "alias": " - high_school_world_history", - "acc,none": 0.43037974683544306, - "acc_stderr,none": 0.03223017195937598 + "acc,none": 0.42616033755274263, + "acc_stderr,none": 0.03219035703131774 }, "mmlu_international_law": { "alias": " - international_law", @@ -38,7 +38,7 @@ "mmlu_jurisprudence": { "alias": " - jurisprudence", "acc,none": 0.32407407407407407, - "acc_stderr,none": 0.04524596007030048 + "acc_stderr,none": 0.04524596007030049 }, "mmlu_logical_fallacies": { "alias": " - logical_fallacies", @@ -48,52 +48,52 @@ "mmlu_moral_disputes": { "alias": " - moral_disputes", "acc,none": 0.315028901734104, - "acc_stderr,none": 0.025009313790069713 + "acc_stderr,none": 0.025009313790069706 }, "mmlu_moral_scenarios": { "alias": " - moral_scenarios", "acc,none": 0.2346368715083799, - "acc_stderr,none": 0.014173044098303667 + "acc_stderr,none": 0.014173044098303679 }, "mmlu_philosophy": { "alias": " - philosophy", "acc,none": 0.3987138263665595, - "acc_stderr,none": 0.027809322585774503 + "acc_stderr,none": 0.0278093225857745 }, "mmlu_prehistory": { "alias": " - prehistory", - "acc,none": 0.3611111111111111, - "acc_stderr,none": 0.026725868809100793 + "acc,none": 0.36419753086419754, + "acc_stderr,none": 0.026774929899722327 }, "mmlu_professional_law": { "alias": " - professional_law", "acc,none": 0.3011734028683181, - "acc_stderr,none": 0.011717148751648431 + "acc_stderr,none": 0.011717148751648424 }, "mmlu_world_religions": { "alias": " - world_religions", "acc,none": 0.391812865497076, - "acc_stderr,none": 0.03743979825926398 + "acc_stderr,none": 0.037439798259264 }, "mmlu_other": { "alias": " - other", - "acc,none": 0.36594785967170906, - "acc_stderr,none": 0.053480370251406284 + "acc,none": 0.3662697135500483, + "acc_stderr,none": 0.049834862065929936 }, "mmlu_business_ethics": { "alias": " - business_ethics", "acc,none": 0.33, - "acc_stderr,none": 0.04725815626252604 + "acc_stderr,none": 0.047258156262526045 }, "mmlu_clinical_knowledge": { "alias": " - clinical_knowledge", "acc,none": 0.37358490566037733, - "acc_stderr,none": 0.029773082713319878 + "acc_stderr,none": 0.02977308271331987 }, "mmlu_college_medicine": { "alias": " - college_medicine", "acc,none": 0.27167630057803466, - "acc_stderr,none": 0.0339175032232166 + "acc_stderr,none": 0.03391750322321659 }, "mmlu_global_facts": { "alias": " - global_facts", @@ -118,42 +118,42 @@ "mmlu_medical_genetics": { "alias": " - medical_genetics", "acc,none": 0.33, - "acc_stderr,none": 0.04725815626252605 + "acc_stderr,none": 0.04725815626252604 }, "mmlu_miscellaneous": { "alias": " - miscellaneous", "acc,none": 0.4367816091954023, - "acc_stderr,none": 0.017736470837800694 + "acc_stderr,none": 0.017736470837800687 }, "mmlu_nutrition": { "alias": " - nutrition", - "acc,none": 0.3235294117647059, - "acc_stderr,none": 0.026787453111906532 + "acc,none": 0.3202614379084967, + "acc_stderr,none": 0.02671611838015685 }, "mmlu_professional_accounting": { "alias": " - professional_accounting", "acc,none": 0.2624113475177305, - "acc_stderr,none": 0.026244920349843014 + "acc_stderr,none": 0.026244920349843007 }, "mmlu_professional_medicine": { "alias": " - professional_medicine", "acc,none": 0.35294117647058826, - "acc_stderr,none": 0.029029422815681393 + "acc_stderr,none": 0.029029422815681397 }, "mmlu_virology": { "alias": " - virology", - "acc,none": 0.3373493975903614, - "acc_stderr,none": 0.03680783690727581 + "acc,none": 0.3493975903614458, + "acc_stderr,none": 0.037117251907407486 }, "mmlu_social_sciences": { "alias": " - social_sciences", - "acc,none": 0.3539161520961976, - "acc_stderr,none": 0.05088692905481045 + "acc,none": 0.35294117647058826, + "acc_stderr,none": 0.04972757753075276 }, "mmlu_econometrics": { "alias": " - econometrics", "acc,none": 0.23684210526315788, - "acc_stderr,none": 0.039994238792813365 + "acc_stderr,none": 0.03999423879281337 }, "mmlu_high_school_geography": { "alias": " - high_school_geography", @@ -163,22 +163,22 @@ "mmlu_high_school_government_and_politics": { "alias": " - high_school_government_and_politics", "acc,none": 0.45077720207253885, - "acc_stderr,none": 0.035909109522355244 + "acc_stderr,none": 0.03590910952235523 }, "mmlu_high_school_macroeconomics": { "alias": " - high_school_macroeconomics", - "acc,none": 0.31025641025641026, - "acc_stderr,none": 0.023454674889404288 + "acc,none": 0.3076923076923077, + "acc_stderr,none": 0.023400928918310495 }, "mmlu_high_school_microeconomics": { "alias": " - high_school_microeconomics", "acc,none": 0.31512605042016806, - "acc_stderr,none": 0.03017680828897434 + "acc_stderr,none": 0.030176808288974337 }, "mmlu_high_school_psychology": { "alias": " - high_school_psychology", - "acc,none": 0.3779816513761468, - "acc_stderr,none": 0.02078918706672811 + "acc,none": 0.3761467889908257, + "acc_stderr,none": 0.02076923196820508 }, "mmlu_human_sexuality": { "alias": " - human_sexuality", @@ -188,17 +188,17 @@ "mmlu_professional_psychology": { "alias": " - professional_psychology", "acc,none": 0.3284313725490196, - "acc_stderr,none": 0.018999707383162673 + "acc_stderr,none": 0.018999707383162666 }, "mmlu_public_relations": { "alias": " - public_relations", "acc,none": 0.39090909090909093, - "acc_stderr,none": 0.04673752333670238 + "acc_stderr,none": 0.04673752333670239 }, "mmlu_security_studies": { "alias": " - security_studies", "acc,none": 0.3224489795918367, - "acc_stderr,none": 0.02992310056368391 + "acc_stderr,none": 0.029923100563683906 }, "mmlu_sociology": { "alias": " - sociology", @@ -207,18 +207,18 @@ }, "mmlu_us_foreign_policy": { "alias": " - us_foreign_policy", - "acc,none": 0.44, - "acc_stderr,none": 0.049888765156985884 + "acc,none": 0.43, + "acc_stderr,none": 0.049756985195624284 }, "mmlu_stem": { "alias": " - stem", - "acc,none": 0.2825880114176974, - "acc_stderr,none": 0.06156202921470587 + "acc,none": 0.28227085315572464, + "acc_stderr,none": 0.060616974885657864 }, "mmlu_abstract_algebra": { "alias": " - abstract_algebra", "acc,none": 0.27, - "acc_stderr,none": 0.0446196043338474 + "acc_stderr,none": 0.044619604333847394 }, "mmlu_anatomy": { "alias": " - anatomy", @@ -228,7 +228,7 @@ "mmlu_astronomy": { "alias": " - astronomy", "acc,none": 0.26973684210526316, - "acc_stderr,none": 0.036117805602848975 + "acc_stderr,none": 0.03611780560284898 }, "mmlu_college_biology": { "alias": " - college_biology", @@ -238,12 +238,12 @@ "mmlu_college_chemistry": { "alias": " - college_chemistry", "acc,none": 0.32, - "acc_stderr,none": 0.04688261722621505 + "acc_stderr,none": 0.04688261722621504 }, "mmlu_college_computer_science": { "alias": " - college_computer_science", "acc,none": 0.19, - "acc_stderr,none": 0.039427724440366234 + "acc_stderr,none": 0.03942772444036625 }, "mmlu_college_mathematics": { "alias": " - college_mathematics", @@ -253,37 +253,37 @@ "mmlu_college_physics": { "alias": " - college_physics", "acc,none": 0.22549019607843138, - "acc_stderr,none": 0.04158307533083286 + "acc_stderr,none": 0.041583075330832865 }, "mmlu_computer_security": { "alias": " - computer_security", "acc,none": 0.28, - "acc_stderr,none": 0.045126085985421276 + "acc_stderr,none": 0.04512608598542126 }, "mmlu_conceptual_physics": { "alias": " - conceptual_physics", "acc,none": 0.3659574468085106, - "acc_stderr,none": 0.03148955829745528 + "acc_stderr,none": 0.0314895582974553 }, "mmlu_electrical_engineering": { "alias": " - electrical_engineering", "acc,none": 0.2827586206896552, - "acc_stderr,none": 0.03752833958003336 + "acc_stderr,none": 0.03752833958003337 }, "mmlu_elementary_mathematics": { "alias": " - elementary_mathematics", "acc,none": 0.2619047619047619, - "acc_stderr,none": 0.022644212615525214 + "acc_stderr,none": 0.022644212615525218 }, "mmlu_high_school_biology": { "alias": " - high_school_biology", "acc,none": 0.3967741935483871, - "acc_stderr,none": 0.02783123160576794 + "acc_stderr,none": 0.027831231605767937 }, "mmlu_high_school_chemistry": { "alias": " - high_school_chemistry", - "acc,none": 0.2955665024630542, - "acc_stderr,none": 0.032104944337514575 + "acc,none": 0.29064039408866993, + "acc_stderr,none": 0.03194740072265541 }, "mmlu_high_school_computer_science": { "alias": " - high_school_computer_science", @@ -293,12 +293,12 @@ "mmlu_high_school_mathematics": { "alias": " - high_school_mathematics", "acc,none": 0.24814814814814815, - "acc_stderr,none": 0.026335739404055803 + "acc_stderr,none": 0.0263357394040558 }, "mmlu_high_school_physics": { "alias": " - high_school_physics", "acc,none": 0.23841059602649006, - "acc_stderr,none": 0.0347918557259966 + "acc_stderr,none": 0.034791855725996586 }, "mmlu_high_school_statistics": { "alias": " - high_school_statistics", @@ -308,34 +308,34 @@ "mmlu_machine_learning": { "alias": " - machine_learning", "acc,none": 0.2767857142857143, - "acc_stderr,none": 0.04246624336697624 + "acc_stderr,none": 0.042466243366976256 } }, "groups": { "mmlu": { - "acc,none": 0.33107819398946015, - "acc_stderr,none": 0.06027611346350394, + "acc,none": 0.33086454920951425, + "acc_stderr,none": 0.060605075693583886, "alias": "mmlu" }, "mmlu_humanities": { "alias": " - humanities", - "acc,none": 0.32561105207226354, - "acc_stderr,none": 0.05639800096282077 + "acc,none": 0.32561105207226343, + "acc_stderr,none": 0.0607809456258765 }, "mmlu_other": { "alias": " - other", - "acc,none": 0.36594785967170906, - "acc_stderr,none": 0.053480370251406284 + "acc,none": 0.3662697135500483, + "acc_stderr,none": 0.049834862065929936 }, "mmlu_social_sciences": { "alias": " - social_sciences", - "acc,none": 0.3539161520961976, - "acc_stderr,none": 0.05088692905481045 + "acc,none": 0.35294117647058826, + "acc_stderr,none": 0.04972757753075276 }, "mmlu_stem": { "alias": " - stem", - "acc,none": 0.2825880114176974, - "acc_stderr,none": 0.06156202921470587 + "acc,none": 0.28227085315572464, + "acc_stderr,none": 0.060616974885657864 } }, "configs": { diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 025f56d06b912b46b4edd2a070671279cadd4a2b..8aac040d7617155baef993d329cefb2caa3c2c6e 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mmlu/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:189db053f672abcdc9c097a35a4d7ee5550b59e0693aeefdda6d580ca423d572 -size 109501 +oid sha256:3b745b9a5577b37b71c88189c4f90a73351adea728c6e8c393479fe62176e823 +size 149131 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..cf67dde624afe622836fbc105a36d3326fa3310d --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,429 @@ +{ + "results": { + "multimedqa": { + "alias": "stem", + "acc,none": 0.35358410220014197, + "acc_stderr,none": 0.09857919062749855, + "acc_norm,none": 0.32068854393146246, + "acc_norm_stderr,none": 0.00010226000674634883 + }, + "medmcqa": { + "acc,none": 0.3265598852498207, + "acc_stderr,none": 0.007251684390085926, + "acc_norm,none": 0.3265598852498207, + "acc_norm_stderr,none": 0.007251684390085926, + "alias": " - medmcqa" + }, + "medqa_4options": { + "acc,none": 0.3102906520031422, + "acc_stderr,none": 0.012971011562150579, + "acc_norm,none": 0.3102906520031422, + "acc_norm_stderr,none": 0.012971011562150579, + "alias": " - medqa_4options" + }, + "mmlu_anatomy": { + "alias": " - anatomy (mmlu)", + "acc,none": 0.34814814814814815, + "acc_stderr,none": 0.041153246103369526 + }, + "mmlu_clinical_knowledge": { + "alias": " - clinical_knowledge (mmlu)", + "acc,none": 0.3660377358490566, + "acc_stderr,none": 0.029647813539365256 + }, + "mmlu_college_biology": { + "alias": " - college_biology (mmlu)", + "acc,none": 0.3263888888888889, + "acc_stderr,none": 0.03921067198982266 + }, + "mmlu_college_medicine": { + "alias": " - college_medicine (mmlu)", + "acc,none": 0.27167630057803466, + "acc_stderr,none": 0.0339175032232166 + }, + "mmlu_medical_genetics": { + "alias": " - medical_genetics (mmlu)", + "acc,none": 0.31, + "acc_stderr,none": 0.04648231987117316 + }, + "mmlu_professional_medicine": { + "alias": " - professional_medicine (mmlu)", + "acc,none": 0.35294117647058826, + "acc_stderr,none": 0.029029422815681393 + }, + "pubmedqa": { + "acc,none": 0.73, + "acc_stderr,none": 0.019874354831287473, + "alias": " - pubmedqa" + } + }, + "groups": { + "multimedqa": { + "alias": "stem", + "acc,none": 0.35358410220014197, + "acc_stderr,none": 0.09857919062749855, + "acc_norm,none": 0.32068854393146246, + "acc_norm_stderr,none": 0.00010226000674634883 + } + }, + "configs": { + "medmcqa": { + "task": "medmcqa", + "dataset_path": "medmcqa", + "training_split": "train", + "validation_split": "validation", + "test_split": "validation", + "doc_to_text": "def doc_to_text(doc) -> str:\n \"\"\"\n Question: \n Choices:\n A. \n B. \n C. \n D. \n Answer:\n \"\"\"\n choices = [doc[\"opa\"], doc[\"opb\"], doc[\"opc\"], doc[\"opd\"]]\n option_choices = {'A': choices[0], 'B': choices[1], 'C': choices[2], 'D': choices[3]}\n\n prompt = \"Question: \" + doc[\"question\"] + \"\\nChoices:\\n\"\n for choice, option in option_choices.items():\n prompt += f\"{choice.upper()}. {option}\\n\"\n prompt += \"Answer:\"\n return prompt\n", + "doc_to_target": "cop", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "acc_norm", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{question}}" + }, + "medqa_4options": { + "task": "medqa_4options", + "dataset_path": "GBaker/MedQA-USMLE-4-options-hf", + "training_split": "train", + "validation_split": "validation", + "test_split": "test", + "doc_to_text": "def doc_to_text(doc) -> str:\n option_choices = {'A': doc[\"ending0\"], 'B': doc[\"ending1\"], 'C': doc[\"ending2\"], 'D': doc[\"ending3\"]}\n answers = \"\".join((f\"{k}. {v}\\n\") for k, v in option_choices.items())\n return f\"Question: {doc['sent1']}\\n{answers}Answer:\"\n", + "doc_to_target": "def doc_to_target(doc) -> int:\n return doc[\"label\"]\n", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "acc_norm", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false + }, + "mmlu_anatomy": { + "task": "mmlu_anatomy", + "task_alias": "anatomy (mmlu)", + "group": "multimedqa", + "group_alias": "stem", + "dataset_path": "hails/mmlu_no_train", + "dataset_name": "anatomy", + "test_split": "test", + "fewshot_split": "dev", + "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", + "doc_to_target": "answer", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "The following are multiple choice questions (with answers) about anatomy.\n\n", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "fewshot_config": { + "sampler": "first_n" + }, + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "mmlu_clinical_knowledge": { + "task": "mmlu_clinical_knowledge", + "task_alias": "clinical_knowledge (mmlu)", + "group": "multimedqa", + "group_alias": "other", + "dataset_path": "hails/mmlu_no_train", + "dataset_name": "clinical_knowledge", + "test_split": "test", + "fewshot_split": "dev", + "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", + "doc_to_target": "answer", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "The following are multiple choice questions (with answers) about clinical knowledge.\n\n", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "fewshot_config": { + "sampler": "first_n" + }, + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "mmlu_college_biology": { + "task": "mmlu_college_biology", + "task_alias": "college_biology (mmlu)", + "group": "multimedqa", + "group_alias": "stem", + "dataset_path": "hails/mmlu_no_train", + "dataset_name": "college_biology", + "test_split": "test", + "fewshot_split": "dev", + "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", + "doc_to_target": "answer", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "The following are multiple choice questions (with answers) about college biology.\n\n", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "fewshot_config": { + "sampler": "first_n" + }, + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "mmlu_college_medicine": { + "task": "mmlu_college_medicine", + "task_alias": "college_medicine (mmlu)", + "group": "multimedqa", + "group_alias": "other", + "dataset_path": "hails/mmlu_no_train", + "dataset_name": "college_medicine", + "test_split": "test", + "fewshot_split": "dev", + "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", + "doc_to_target": "answer", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "The following are multiple choice questions (with answers) about college medicine.\n\n", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "fewshot_config": { + "sampler": "first_n" + }, + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "mmlu_medical_genetics": { + "task": "mmlu_medical_genetics", + "task_alias": "medical_genetics (mmlu)", + "group": "multimedqa", + "group_alias": "other", + "dataset_path": "hails/mmlu_no_train", + "dataset_name": "medical_genetics", + "test_split": "test", + "fewshot_split": "dev", + "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", + "doc_to_target": "answer", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "The following are multiple choice questions (with answers) about medical genetics.\n\n", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "fewshot_config": { + "sampler": "first_n" + }, + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "mmlu_professional_medicine": { + "task": "mmlu_professional_medicine", + "task_alias": "professional_medicine (mmlu)", + "group": "multimedqa", + "group_alias": "other", + "dataset_path": "hails/mmlu_no_train", + "dataset_name": "professional_medicine", + "test_split": "test", + "fewshot_split": "dev", + "doc_to_text": "{{question.strip()}}\nA. {{choices[0]}}\nB. {{choices[1]}}\nC. {{choices[2]}}\nD. {{choices[3]}}\nAnswer:", + "doc_to_target": "answer", + "doc_to_choice": [ + "A", + "B", + "C", + "D" + ], + "description": "The following are multiple choice questions (with answers) about professional medicine.\n\n", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "fewshot_config": { + "sampler": "first_n" + }, + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "pubmedqa": { + "task": "pubmedqa", + "dataset_path": "bigbio/pubmed_qa", + "dataset_name": "pubmed_qa_labeled_fold0_source", + "training_split": "train", + "validation_split": "validation", + "test_split": "test", + "doc_to_text": "def doc_to_text(doc) -> str:\n ctxs = \"\\n\".join(doc[\"CONTEXTS\"])\n return \"Abstract: {}\\nQuestion: {}\\nAnswer:\".format(\n ctxs,\n doc[\"QUESTION\"],\n )\n", + "doc_to_target": "final_decision", + "doc_to_choice": [ + "yes", + "no", + "maybe" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "medmcqa": "Yaml", + "medqa_4options": "Yaml", + "mmlu_anatomy": 0.0, + "mmlu_clinical_knowledge": 0.0, + "mmlu_college_biology": 0.0, + "mmlu_college_medicine": 0.0, + "mmlu_medical_genetics": 0.0, + "mmlu_professional_medicine": 0.0, + "multimedqa": "N/A", + "pubmedqa": 1.0 + }, + "n-shot": { + "medmcqa": 0, + "medqa_4options": 0, + "mmlu_anatomy": 0, + "mmlu_clinical_knowledge": 0, + "mmlu_college_biology": 0, + "mmlu_college_medicine": 0, + "mmlu_medical_genetics": 0, + "mmlu_professional_medicine": 0, + "multimedqa": 0, + "pubmedqa": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 16 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..427beb150b62b6d33e597855a6af3c0975f8c782 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multimedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b7173387d354278e08e045e8571c3eef57a62b86eb1569b4d3ef5f3f7104494 +size 60906 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..2790b7003acce36c3c771851a13c67822e663b1d --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,58 @@ +{ + "results": { + "multirc": { + "acc,none": 0.49257425742574257, + "acc_stderr,none": 0.007181011029317661, + "alias": "multirc" + } + }, + "configs": { + "multirc": { + "task": "multirc", + "group": [ + "super-glue-lm-eval-v1" + ], + "dataset_path": "super_glue", + "dataset_name": "multirc", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "{{paragraph}}\nQuestion: {{question}}\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": "['''{{answer}}\\nIs the answer correct? yes''', '''{{answer}}\\nIs the answer correct? no''']", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 2.0 + } + } + }, + "versions": { + "multirc": 2.0 + }, + "n-shot": { + "multirc": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 32 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..fc1c47239928855b949f8ba3d3965210f8b01ac2 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/multirc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:210d3cb22fa8afecd2424263a6d2c02d5c2da9bf81c0643f92b6e92d778e4492 +size 47571 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..ba8677e80e635f5647f93392cc58cd23bce556c7 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,74 @@ +{ + "results": { + "mutual": { + "r@1,none": 0.22573363431151242, + "r@1_stderr,none": 0.014053085820407473, + "r@2,none": 0.4221218961625282, + "r@2_stderr,none": 0.01660219170551757, + "mrr,none": 0.7004326577582155, + "mrr_stderr,none": 0.010311480723006769, + "alias": "mutual" + } + }, + "configs": { + "mutual": { + "task": "mutual", + "dataset_path": "EleutherAI/mutual", + "dataset_name": "mutual", + "training_split": "train", + "validation_split": "validation", + "process_docs": "def process_docs(dataset):\n def _detokenize(text):\n text = text.replace(\" '\", \"'\")\n text = text.replace(\" \\n\", \"\\n\")\n text = text.replace(\"\\n \", \"\\n\")\n text = text.replace(\" n't\", \"n't\")\n text = text.replace(\"`` \", '\"')\n text = text.replace(\"''\", '\"')\n # punctuation\n text = text.replace(\" :\", \":\")\n text = text.replace(\" ;\", \";\")\n text = text.replace(\" !\", \"!\")\n text = text.replace(\" ?\", \"?\")\n text = text.replace(\" ,\", \",\")\n text = text.replace(\" .\", \".\")\n return text\n\n def _process(doc):\n return {\n \"article\": _detokenize(doc[\"article\"]),\n \"options\": [_detokenize(option) for option in doc[\"options\"]],\n }\n\n return dataset.map(_process)\n", + "doc_to_text": "{{article}}", + "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answers)}}", + "doc_to_choice": "{{options}}", + "process_results": "def process_results(doc, results):\n gold = [\"A\", \"B\", \"C\", \"D\"].index(doc[\"answers\"])\n r4_1 = np.argmax(results) == gold # r4_1 = accuracy\n ranks = sorted(results, reverse=True)\n r4_2 = (ranks.index(results[gold]) == 1) + r4_1\n mrr = 1.0 / (ranks.index(results[gold]) + 1) # `+ 1` for index offset\n return {\"r@1\": r4_1, \"r@2\": r4_2, \"mrr\": mrr}\n", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "r@1", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "r@2", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "mrr", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{article}}", + "metadata": { + "version": 2.0 + } + } + }, + "versions": { + "mutual": 2.0 + }, + "n-shot": { + "mutual": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..bd12022ea290882220259c82a8a717acd90a75aa --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cef319db9b4b609ce245fc80708acc8d53884bb373492d70b4ea36eb8ac307f5 +size 50696 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..41ac18e5d6e9b1d528df94e5fa110ba00c5fe4f5 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,74 @@ +{ + "results": { + "mutual_plus": { + "r@1,none": 0.2595936794582393, + "r@1_stderr,none": 0.014737047402750952, + "r@2,none": 0.463882618510158, + "r@2_stderr,none": 0.016763409667403396, + "mrr,none": 0.6583897683676962, + "mrr_stderr,none": 0.010460590295448532, + "alias": "mutual_plus" + } + }, + "configs": { + "mutual_plus": { + "task": "mutual_plus", + "dataset_path": "EleutherAI/mutual", + "dataset_name": "mutual_plus", + "training_split": "train", + "validation_split": "validation", + "process_docs": "def process_docs(dataset):\n def _detokenize(text):\n text = text.replace(\" '\", \"'\")\n text = text.replace(\" \\n\", \"\\n\")\n text = text.replace(\"\\n \", \"\\n\")\n text = text.replace(\" n't\", \"n't\")\n text = text.replace(\"`` \", '\"')\n text = text.replace(\"''\", '\"')\n # punctuation\n text = text.replace(\" :\", \":\")\n text = text.replace(\" ;\", \";\")\n text = text.replace(\" !\", \"!\")\n text = text.replace(\" ?\", \"?\")\n text = text.replace(\" ,\", \",\")\n text = text.replace(\" .\", \".\")\n return text\n\n def _process(doc):\n return {\n \"article\": _detokenize(doc[\"article\"]),\n \"options\": [_detokenize(option) for option in doc[\"options\"]],\n }\n\n return dataset.map(_process)\n", + "doc_to_text": "{{article}}", + "doc_to_target": "{{['A', 'B', 'C', 'D'].index(answers)}}", + "doc_to_choice": "{{options}}", + "process_results": "def process_results(doc, results):\n gold = [\"A\", \"B\", \"C\", \"D\"].index(doc[\"answers\"])\n r4_1 = np.argmax(results) == gold # r4_1 = accuracy\n ranks = sorted(results, reverse=True)\n r4_2 = (ranks.index(results[gold]) == 1) + r4_1\n mrr = 1.0 / (ranks.index(results[gold]) + 1) # `+ 1` for index offset\n return {\"r@1\": r4_1, \"r@2\": r4_2, \"mrr\": mrr}\n", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "r@1", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "r@2", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "mrr", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{article}}", + "metadata": { + "version": 2.0 + } + } + }, + "versions": { + "mutual_plus": 2.0 + }, + "n-shot": { + "mutual_plus": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..10db45280d8d1e04f0689827f390802b0140ac50 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/mutual_plus/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc89a5e0d37c1e4a95893062aedb50b98b09cbc1dfd53bfdf254d430d07f1136 +size 51801 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 3e75d819f45048de66e7f53d3de582881d6f7472..741307ab3762d55525e84fd5f09b76523dfb3e15 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/piqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3f7f76bc2192c6bea2186fcf369021fe26d059fb40bce57661c93362607523d2 -size 43659 +oid sha256:4779f4c23626f1749c089a7511a0da837ac7867f558dc956def2ec1d09b5433f +size 42146 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..705347f83264a0e2c370993616961f3a717b91d1 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,63 @@ +{ + "results": { + "prost": { + "acc,none": 0.24338172502134928, + "acc_stderr,none": 0.0031351299519620946, + "acc_norm,none": 0.28031596925704527, + "acc_norm_stderr,none": 0.0032814667207950675, + "alias": "prost" + } + }, + "configs": { + "prost": { + "task": "prost", + "dataset_path": "corypaik/prost", + "test_split": "test", + "doc_to_text": "{{context}}\nQuestion: {{ex_question}}\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": "{{[A, B, C, D]}}", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "acc_norm", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{context}}\nQuestion: {{ex_question}}\nAnswer:", + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "prost": 1.0 + }, + "n-shot": { + "prost": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..6c1753deb47372126058fee8c362405d4d0ad9e1 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/prost/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:582e4b78a9101bacc705ef1d09c098d5830bfe561bca7e4b45fa7063f0933e08 +size 53887 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..28ce0374ab500bfb9e07d5aab4d92275d0bb3bc6 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,62 @@ +{ + "results": { + "pubmedqa": { + "acc,none": 0.73, + "acc_stderr,none": 0.01987435483128749, + "alias": "pubmedqa" + } + }, + "configs": { + "pubmedqa": { + "task": "pubmedqa", + "dataset_path": "bigbio/pubmed_qa", + "dataset_name": "pubmed_qa_labeled_fold0_source", + "training_split": "train", + "validation_split": "validation", + "test_split": "test", + "doc_to_text": "def doc_to_text(doc) -> str:\n ctxs = \"\\n\".join(doc[\"CONTEXTS\"])\n return \"Abstract: {}\\nQuestion: {}\\nAnswer:\".format(\n ctxs,\n doc[\"QUESTION\"],\n )\n", + "doc_to_target": "final_decision", + "doc_to_choice": [ + "yes", + "no", + "maybe" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "pubmedqa": 1.0 + }, + "n-shot": { + "pubmedqa": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 32 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..6df386429b58092da597fd7128371f20e1c3dd00 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pubmedqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8be040043f48fa6bb045726f655899436ce74a363bdcd3aee21ee6ceec698bf +size 43156 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 183f74dd07134ff874b2cc2c1599c2a4ced0f7bd..1cc6f3d44c0dd140e047cd3ede2556fe82315037 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,18 +1,18 @@ { "results": { "pythia": { - "acc,none": 0.7433806828192467, - "acc_stderr,none": 0.13793235163862885, + "acc,none": 0.7433912692426552, + "acc_stderr,none": 0.14350617224236548, "acc_norm,none": 0.6280221175210029, "acc_norm_stderr,none": 0.010330170441803736, - "word_perplexity,none": 10.428191022549841, + "word_perplexity,none": 10.428882412474298, "word_perplexity_stderr,none": "N/A", - "byte_perplexity,none": 1.5502800869079052, + "byte_perplexity,none": 1.5502993074551348, "byte_perplexity_stderr,none": "N/A", - "bits_per_byte,none": 0.6325288887179478, + "bits_per_byte,none": 0.6325467753027727, "bits_per_byte_stderr,none": "N/A", - "perplexity,none": 3.376255600786083, - "perplexity_stderr,none": 0.06622887148707989, + "perplexity,none": 3.3763932503864287, + "perplexity_stderr,none": 0.06622884189802866, "alias": "pythia" }, "ai2_arc": { @@ -24,46 +24,46 @@ }, "arc_challenge": { "acc,none": 0.39761092150170646, - "acc_stderr,none": 0.014301752223279524, + "acc_stderr,none": 0.014301752223279531, "acc_norm,none": 0.4308873720136519, - "acc_norm_stderr,none": 0.01447113339264248, + "acc_norm_stderr,none": 0.014471133392642483, "alias": " - arc_challenge" }, "arc_easy": { "acc,none": 0.7369528619528619, "acc_stderr,none": 0.00903451489886582, "acc_norm,none": 0.7192760942760943, - "acc_norm_stderr,none": 0.009220526174711361, + "acc_norm_stderr,none": 0.00922052617471136, "alias": " - arc_easy" }, "blimp": { "acc,none": 0.8386865671641791, - "acc_stderr,none": 0.14193008369315854, + "acc_stderr,none": 0.14955502218955172, "alias": " - blimp" }, "blimp_adjunct_island": { "acc,none": 0.913, - "acc_stderr,none": 0.008916866630745916, + "acc_stderr,none": 0.008916866630745892, "alias": " - blimp_adjunct_island" }, "blimp_anaphor_gender_agreement": { "acc,none": 0.988, - "acc_stderr,none": 0.0034449771940998413, + "acc_stderr,none": 0.003444977194099851, "alias": " - blimp_anaphor_gender_agreement" }, "blimp_anaphor_number_agreement": { "acc,none": 0.999, - "acc_stderr,none": 0.0010000000000000143, + "acc_stderr,none": 0.0010000000000000091, "alias": " - blimp_anaphor_number_agreement" }, "blimp_animate_subject_passive": { - "acc,none": 0.829, - "acc_stderr,none": 0.011912216456264602, + "acc,none": 0.828, + "acc_stderr,none": 0.011939788882495323, "alias": " - blimp_animate_subject_passive" }, "blimp_animate_subject_trans": { "acc,none": 0.912, - "acc_stderr,none": 0.008963053962592064, + "acc_stderr,none": 0.008963053962592069, "alias": " - blimp_animate_subject_trans" }, "blimp_causative": { @@ -83,17 +83,17 @@ }, "blimp_coordinate_structure_constraint_object_extraction": { "acc,none": 0.861, - "acc_stderr,none": 0.010945263761042967, + "acc_stderr,none": 0.010945263761042962, "alias": " - blimp_coordinate_structure_constraint_object_extraction" }, "blimp_determiner_noun_agreement_1": { "acc,none": 0.997, - "acc_stderr,none": 0.0017303161543469362, + "acc_stderr,none": 0.0017303161543469375, "alias": " - blimp_determiner_noun_agreement_1" }, "blimp_determiner_noun_agreement_2": { "acc,none": 0.986, - "acc_stderr,none": 0.003717232548256594, + "acc_stderr,none": 0.0037172325482565543, "alias": " - blimp_determiner_noun_agreement_2" }, "blimp_determiner_noun_agreement_irregular_1": { @@ -103,17 +103,17 @@ }, "blimp_determiner_noun_agreement_irregular_2": { "acc,none": 0.965, - "acc_stderr,none": 0.005814534272734973, + "acc_stderr,none": 0.005814534272734967, "alias": " - blimp_determiner_noun_agreement_irregular_2" }, "blimp_determiner_noun_agreement_with_adj_2": { "acc,none": 0.964, - "acc_stderr,none": 0.00589395781616558, + "acc_stderr,none": 0.005893957816165557, "alias": " - blimp_determiner_noun_agreement_with_adj_2" }, "blimp_determiner_noun_agreement_with_adj_irregular_1": { "acc,none": 0.938, - "acc_stderr,none": 0.007629823996280302, + "acc_stderr,none": 0.0076298239962803134, "alias": " - blimp_determiner_noun_agreement_with_adj_irregular_1" }, "blimp_determiner_noun_agreement_with_adj_irregular_2": { @@ -123,72 +123,72 @@ }, "blimp_determiner_noun_agreement_with_adjective_1": { "acc,none": 0.98, - "acc_stderr,none": 0.004429403980178344, + "acc_stderr,none": 0.004429403980178341, "alias": " - blimp_determiner_noun_agreement_with_adjective_1" }, "blimp_distractor_agreement_relational_noun": { "acc,none": 0.926, - "acc_stderr,none": 0.008282064512704164, + "acc_stderr,none": 0.008282064512704166, "alias": " - blimp_distractor_agreement_relational_noun" }, "blimp_distractor_agreement_relative_clause": { - "acc,none": 0.793, - "acc_stderr,none": 0.012818553557843986, + "acc,none": 0.794, + "acc_stderr,none": 0.012795613612786558, "alias": " - blimp_distractor_agreement_relative_clause" }, "blimp_drop_argument": { - "acc,none": 0.771, - "acc_stderr,none": 0.0132941993266136, + "acc,none": 0.772, + "acc_stderr,none": 0.013273740700804487, "alias": " - blimp_drop_argument" }, "blimp_ellipsis_n_bar_1": { "acc,none": 0.81, - "acc_stderr,none": 0.012411851354816325, + "acc_stderr,none": 0.01241185135481633, "alias": " - blimp_ellipsis_n_bar_1" }, "blimp_ellipsis_n_bar_2": { - "acc,none": 0.924, - "acc_stderr,none": 0.008384169266796394, + "acc,none": 0.923, + "acc_stderr,none": 0.008434580140240625, "alias": " - blimp_ellipsis_n_bar_2" }, "blimp_existential_there_object_raising": { "acc,none": 0.832, - "acc_stderr,none": 0.011828605831454266, + "acc_stderr,none": 0.011828605831454262, "alias": " - blimp_existential_there_object_raising" }, "blimp_existential_there_quantifiers_1": { "acc,none": 0.984, - "acc_stderr,none": 0.003969856390319422, + "acc_stderr,none": 0.003969856390319419, "alias": " - blimp_existential_there_quantifiers_1" }, "blimp_existential_there_quantifiers_2": { "acc,none": 0.431, - "acc_stderr,none": 0.0156679444881735, + "acc_stderr,none": 0.015667944488173508, "alias": " - blimp_existential_there_quantifiers_2" }, "blimp_existential_there_subject_raising": { "acc,none": 0.853, - "acc_stderr,none": 0.011203415395160331, + "acc_stderr,none": 0.01120341539516033, "alias": " - blimp_existential_there_subject_raising" }, "blimp_expletive_it_object_raising": { "acc,none": 0.795, - "acc_stderr,none": 0.01277255409611312, + "acc_stderr,none": 0.012772554096113125, "alias": " - blimp_expletive_it_object_raising" }, "blimp_inchoative": { "acc,none": 0.686, - "acc_stderr,none": 0.014683991951087962, + "acc_stderr,none": 0.014683991951087966, "alias": " - blimp_inchoative" }, "blimp_intransitive": { "acc,none": 0.851, - "acc_stderr,none": 0.011266140684632168, + "acc_stderr,none": 0.01126614068463216, "alias": " - blimp_intransitive" }, "blimp_irregular_past_participle_adjectives": { "acc,none": 0.939, - "acc_stderr,none": 0.00757207609155742, + "acc_stderr,none": 0.007572076091557425, "alias": " - blimp_irregular_past_participle_adjectives" }, "blimp_irregular_past_participle_verbs": { @@ -198,17 +198,17 @@ }, "blimp_irregular_plural_subject_verb_agreement_1": { "acc,none": 0.937, - "acc_stderr,none": 0.007687007876286428, + "acc_stderr,none": 0.007687007876286427, "alias": " - blimp_irregular_plural_subject_verb_agreement_1" }, "blimp_irregular_plural_subject_verb_agreement_2": { "acc,none": 0.895, - "acc_stderr,none": 0.009698921026024947, + "acc_stderr,none": 0.009698921026024952, "alias": " - blimp_irregular_plural_subject_verb_agreement_2" }, "blimp_left_branch_island_echo_question": { "acc,none": 0.626, - "acc_stderr,none": 0.015308767369006372, + "acc_stderr,none": 0.015308767369006363, "alias": " - blimp_left_branch_island_echo_question" }, "blimp_left_branch_island_simple_question": { @@ -223,37 +223,37 @@ }, "blimp_npi_present_1": { "acc,none": 0.608, - "acc_stderr,none": 0.015445859463771307, + "acc_stderr,none": 0.015445859463771297, "alias": " - blimp_npi_present_1" }, "blimp_npi_present_2": { "acc,none": 0.69, - "acc_stderr,none": 0.014632638658632896, + "acc_stderr,none": 0.014632638658632895, "alias": " - blimp_npi_present_2" }, "blimp_only_npi_licensor_present": { "acc,none": 0.937, - "acc_stderr,none": 0.007687007876286425, + "acc_stderr,none": 0.007687007876286415, "alias": " - blimp_only_npi_licensor_present" }, "blimp_only_npi_scope": { "acc,none": 0.86, - "acc_stderr,none": 0.010978183844357807, + "acc_stderr,none": 0.010978183844357793, "alias": " - blimp_only_npi_scope" }, "blimp_passive_1": { "acc,none": 0.898, - "acc_stderr,none": 0.00957536880165387, + "acc_stderr,none": 0.009575368801653886, "alias": " - blimp_passive_1" }, "blimp_passive_2": { "acc,none": 0.897, - "acc_stderr,none": 0.009616833339695801, + "acc_stderr,none": 0.009616833339695804, "alias": " - blimp_passive_2" }, "blimp_principle_A_c_command": { "acc,none": 0.801, - "acc_stderr,none": 0.012631649083099189, + "acc_stderr,none": 0.012631649083099163, "alias": " - blimp_principle_A_c_command" }, "blimp_principle_A_case_1": { @@ -263,22 +263,22 @@ }, "blimp_principle_A_case_2": { "acc,none": 0.954, - "acc_stderr,none": 0.006627814717380712, + "acc_stderr,none": 0.006627814717380713, "alias": " - blimp_principle_A_case_2" }, "blimp_principle_A_domain_1": { "acc,none": 0.999, - "acc_stderr,none": 0.0010000000000000033, + "acc_stderr,none": 0.0010000000000000044, "alias": " - blimp_principle_A_domain_1" }, "blimp_principle_A_domain_2": { "acc,none": 0.924, - "acc_stderr,none": 0.008384169266796403, + "acc_stderr,none": 0.008384169266796394, "alias": " - blimp_principle_A_domain_2" }, "blimp_principle_A_domain_3": { "acc,none": 0.805, - "acc_stderr,none": 0.012535235623319325, + "acc_stderr,none": 0.012535235623319324, "alias": " - blimp_principle_A_domain_3" }, "blimp_principle_A_reconstruction": { @@ -288,27 +288,27 @@ }, "blimp_regular_plural_subject_verb_agreement_1": { "acc,none": 0.966, - "acc_stderr,none": 0.005733836139695459, + "acc_stderr,none": 0.005733836139695435, "alias": " - blimp_regular_plural_subject_verb_agreement_1" }, "blimp_regular_plural_subject_verb_agreement_2": { "acc,none": 0.915, - "acc_stderr,none": 0.008823426366942307, + "acc_stderr,none": 0.008823426366942312, "alias": " - blimp_regular_plural_subject_verb_agreement_2" }, "blimp_sentential_negation_npi_licensor_present": { "acc,none": 0.986, - "acc_stderr,none": 0.0037172325482565678, + "acc_stderr,none": 0.0037172325482565595, "alias": " - blimp_sentential_negation_npi_licensor_present" }, "blimp_sentential_negation_npi_scope": { "acc,none": 0.731, - "acc_stderr,none": 0.014029819522568198, + "acc_stderr,none": 0.014029819522568196, "alias": " - blimp_sentential_negation_npi_scope" }, "blimp_sentential_subject_island": { "acc,none": 0.474, - "acc_stderr,none": 0.015797897758042755, + "acc_stderr,none": 0.01579789775804276, "alias": " - blimp_sentential_subject_island" }, "blimp_superlative_quantifiers_1": { @@ -318,17 +318,17 @@ }, "blimp_superlative_quantifiers_2": { "acc,none": 0.96, - "acc_stderr,none": 0.006199874066337076, + "acc_stderr,none": 0.006199874066337051, "alias": " - blimp_superlative_quantifiers_2" }, "blimp_tough_vs_raising_1": { "acc,none": 0.636, - "acc_stderr,none": 0.01522286884052202, + "acc_stderr,none": 0.015222868840522015, "alias": " - blimp_tough_vs_raising_1" }, "blimp_tough_vs_raising_2": { "acc,none": 0.87, - "acc_stderr,none": 0.010640169792499338, + "acc_stderr,none": 0.010640169792499354, "alias": " - blimp_tough_vs_raising_2" }, "blimp_transitive": { @@ -338,7 +338,7 @@ }, "blimp_wh_island": { "acc,none": 0.814, - "acc_stderr,none": 0.012310790208412794, + "acc_stderr,none": 0.012310790208412803, "alias": " - blimp_wh_island" }, "blimp_wh_questions_object_gap": { @@ -348,57 +348,57 @@ }, "blimp_wh_questions_subject_gap": { "acc,none": 0.937, - "acc_stderr,none": 0.0076870078762864115, + "acc_stderr,none": 0.007687007876286408, "alias": " - blimp_wh_questions_subject_gap" }, "blimp_wh_questions_subject_gap_long_distance": { "acc,none": 0.922, - "acc_stderr,none": 0.008484573530118588, + "acc_stderr,none": 0.008484573530118581, "alias": " - blimp_wh_questions_subject_gap_long_distance" }, "blimp_wh_vs_that_no_gap": { "acc,none": 0.981, - "acc_stderr,none": 0.004319451082910606, + "acc_stderr,none": 0.0043194510829106135, "alias": " - blimp_wh_vs_that_no_gap" }, "blimp_wh_vs_that_no_gap_long_distance": { "acc,none": 0.967, - "acc_stderr,none": 0.005651808820452374, + "acc_stderr,none": 0.005651808820452372, "alias": " - blimp_wh_vs_that_no_gap_long_distance" }, "blimp_wh_vs_that_with_gap": { "acc,none": 0.405, - "acc_stderr,none": 0.015531136990453049, + "acc_stderr,none": 0.015531136990453045, "alias": " - blimp_wh_vs_that_with_gap" }, "blimp_wh_vs_that_with_gap_long_distance": { "acc,none": 0.359, - "acc_stderr,none": 0.015177264224798585, + "acc_stderr,none": 0.015177264224798596, "alias": " - blimp_wh_vs_that_with_gap_long_distance" }, "lambada_openai": { - "perplexity,none": 3.376255600786083, - "perplexity_stderr,none": 0.06622887148707989, - "acc,none": 0.7438385406559286, - "acc_stderr,none": 0.006081466315674246, + "perplexity,none": 3.3763932503864287, + "perplexity_stderr,none": 0.06622884189802866, + "acc,none": 0.74345041723268, + "acc_stderr,none": 0.006084483727167678, "alias": " - lambada_openai" }, "logiqa": { "acc,none": 0.2457757296466974, - "acc_stderr,none": 0.016887410894296958, + "acc_stderr,none": 0.016887410894296927, "acc_norm,none": 0.28417818740399386, - "acc_norm_stderr,none": 0.01769054268019077, + "acc_norm_stderr,none": 0.01769054268019079, "alias": " - logiqa" }, "mmlu": { - "acc,none": 0.33093576413616294, - "acc_stderr,none": 0.06067347861769321, + "acc,none": 0.33114940891610883, + "acc_stderr,none": 0.06038941147971378, "alias": " - mmlu" }, "mmlu_humanities": { "alias": " - humanities", - "acc,none": 0.32539851222104144, - "acc_stderr,none": 0.060721858674346096 + "acc,none": 0.32582359192348564, + "acc_stderr,none": 0.05651754471436536 }, "mmlu_formal_logic": { "alias": " - formal_logic", @@ -417,8 +417,8 @@ }, "mmlu_high_school_world_history": { "alias": " - high_school_world_history", - "acc,none": 0.42616033755274263, - "acc_stderr,none": 0.03219035703131774 + "acc,none": 0.43037974683544306, + "acc_stderr,none": 0.03223017195937598 }, "mmlu_international_law": { "alias": " - international_law", @@ -428,7 +428,7 @@ "mmlu_jurisprudence": { "alias": " - jurisprudence", "acc,none": 0.32407407407407407, - "acc_stderr,none": 0.04524596007030049 + "acc_stderr,none": 0.04524596007030048 }, "mmlu_logical_fallacies": { "alias": " - logical_fallacies", @@ -438,52 +438,52 @@ "mmlu_moral_disputes": { "alias": " - moral_disputes", "acc,none": 0.315028901734104, - "acc_stderr,none": 0.025009313790069706 + "acc_stderr,none": 0.025009313790069713 }, "mmlu_moral_scenarios": { "alias": " - moral_scenarios", "acc,none": 0.2346368715083799, - "acc_stderr,none": 0.014173044098303679 + "acc_stderr,none": 0.014173044098303667 }, "mmlu_philosophy": { "alias": " - philosophy", "acc,none": 0.3987138263665595, - "acc_stderr,none": 0.0278093225857745 + "acc_stderr,none": 0.027809322585774503 }, "mmlu_prehistory": { "alias": " - prehistory", - "acc,none": 0.3611111111111111, - "acc_stderr,none": 0.026725868809100786 + "acc,none": 0.36419753086419754, + "acc_stderr,none": 0.026774929899722324 }, "mmlu_professional_law": { "alias": " - professional_law", "acc,none": 0.3011734028683181, - "acc_stderr,none": 0.011717148751648424 + "acc_stderr,none": 0.011717148751648431 }, "mmlu_world_religions": { "alias": " - world_religions", "acc,none": 0.391812865497076, - "acc_stderr,none": 0.037439798259264 + "acc_stderr,none": 0.03743979825926398 }, "mmlu_other": { "alias": " - other", "acc,none": 0.36594785967170906, - "acc_stderr,none": 0.0497900578717159 + "acc_stderr,none": 0.053480370251406284 }, "mmlu_business_ethics": { "alias": " - business_ethics", "acc,none": 0.33, - "acc_stderr,none": 0.047258156262526045 + "acc_stderr,none": 0.04725815626252604 }, "mmlu_clinical_knowledge": { "alias": " - clinical_knowledge", "acc,none": 0.37358490566037733, - "acc_stderr,none": 0.02977308271331987 + "acc_stderr,none": 0.029773082713319878 }, "mmlu_college_medicine": { "alias": " - college_medicine", "acc,none": 0.27167630057803466, - "acc_stderr,none": 0.03391750322321659 + "acc_stderr,none": 0.0339175032232166 }, "mmlu_global_facts": { "alias": " - global_facts", @@ -508,27 +508,27 @@ "mmlu_medical_genetics": { "alias": " - medical_genetics", "acc,none": 0.33, - "acc_stderr,none": 0.04725815626252604 + "acc_stderr,none": 0.04725815626252605 }, "mmlu_miscellaneous": { "alias": " - miscellaneous", "acc,none": 0.4367816091954023, - "acc_stderr,none": 0.017736470837800687 + "acc_stderr,none": 0.017736470837800694 }, "mmlu_nutrition": { "alias": " - nutrition", "acc,none": 0.3235294117647059, - "acc_stderr,none": 0.026787453111906535 + "acc_stderr,none": 0.026787453111906532 }, "mmlu_professional_accounting": { "alias": " - professional_accounting", "acc,none": 0.2624113475177305, - "acc_stderr,none": 0.026244920349843007 + "acc_stderr,none": 0.026244920349843014 }, "mmlu_professional_medicine": { "alias": " - professional_medicine", "acc,none": 0.35294117647058826, - "acc_stderr,none": 0.029029422815681397 + "acc_stderr,none": 0.029029422815681393 }, "mmlu_virology": { "alias": " - virology", @@ -538,12 +538,12 @@ "mmlu_social_sciences": { "alias": " - social_sciences", "acc,none": 0.35359116022099446, - "acc_stderr,none": 0.04999851255154558 + "acc_stderr,none": 0.051101563313320164 }, "mmlu_econometrics": { "alias": " - econometrics", "acc,none": 0.23684210526315788, - "acc_stderr,none": 0.03999423879281337 + "acc_stderr,none": 0.039994238792813365 }, "mmlu_high_school_geography": { "alias": " - high_school_geography", @@ -553,17 +553,17 @@ "mmlu_high_school_government_and_politics": { "alias": " - high_school_government_and_politics", "acc,none": 0.45077720207253885, - "acc_stderr,none": 0.03590910952235523 + "acc_stderr,none": 0.035909109522355244 }, "mmlu_high_school_macroeconomics": { "alias": " - high_school_macroeconomics", "acc,none": 0.3076923076923077, - "acc_stderr,none": 0.023400928918310495 + "acc_stderr,none": 0.023400928918310488 }, "mmlu_high_school_microeconomics": { "alias": " - high_school_microeconomics", "acc,none": 0.31512605042016806, - "acc_stderr,none": 0.030176808288974337 + "acc_stderr,none": 0.03017680828897434 }, "mmlu_high_school_psychology": { "alias": " - high_school_psychology", @@ -578,17 +578,17 @@ "mmlu_professional_psychology": { "alias": " - professional_psychology", "acc,none": 0.3284313725490196, - "acc_stderr,none": 0.018999707383162666 + "acc_stderr,none": 0.018999707383162673 }, "mmlu_public_relations": { "alias": " - public_relations", "acc,none": 0.39090909090909093, - "acc_stderr,none": 0.04673752333670239 + "acc_stderr,none": 0.04673752333670238 }, "mmlu_security_studies": { "alias": " - security_studies", "acc,none": 0.3224489795918367, - "acc_stderr,none": 0.029923100563683906 + "acc_stderr,none": 0.02992310056368391 }, "mmlu_sociology": { "alias": " - sociology", @@ -598,17 +598,17 @@ "mmlu_us_foreign_policy": { "alias": " - us_foreign_policy", "acc,none": 0.44, - "acc_stderr,none": 0.04988876515698589 + "acc_stderr,none": 0.049888765156985884 }, "mmlu_stem": { "alias": " - stem", - "acc,none": 0.28258801141769746, - "acc_stderr,none": 0.06097443015202581 + "acc,none": 0.28290516967967017, + "acc_stderr,none": 0.061921192357468 }, "mmlu_abstract_algebra": { "alias": " - abstract_algebra", "acc,none": 0.27, - "acc_stderr,none": 0.044619604333847394 + "acc_stderr,none": 0.0446196043338474 }, "mmlu_anatomy": { "alias": " - anatomy", @@ -618,7 +618,7 @@ "mmlu_astronomy": { "alias": " - astronomy", "acc,none": 0.26973684210526316, - "acc_stderr,none": 0.03611780560284898 + "acc_stderr,none": 0.036117805602848975 }, "mmlu_college_biology": { "alias": " - college_biology", @@ -628,12 +628,12 @@ "mmlu_college_chemistry": { "alias": " - college_chemistry", "acc,none": 0.32, - "acc_stderr,none": 0.04688261722621504 + "acc_stderr,none": 0.04688261722621505 }, "mmlu_college_computer_science": { "alias": " - college_computer_science", "acc,none": 0.19, - "acc_stderr,none": 0.03942772444036625 + "acc_stderr,none": 0.039427724440366234 }, "mmlu_college_mathematics": { "alias": " - college_mathematics", @@ -643,37 +643,37 @@ "mmlu_college_physics": { "alias": " - college_physics", "acc,none": 0.22549019607843138, - "acc_stderr,none": 0.041583075330832865 + "acc_stderr,none": 0.04158307533083286 }, "mmlu_computer_security": { "alias": " - computer_security", "acc,none": 0.28, - "acc_stderr,none": 0.04512608598542126 + "acc_stderr,none": 0.045126085985421276 }, "mmlu_conceptual_physics": { "alias": " - conceptual_physics", "acc,none": 0.3702127659574468, - "acc_stderr,none": 0.03156564682236784 + "acc_stderr,none": 0.031565646822367836 }, "mmlu_electrical_engineering": { "alias": " - electrical_engineering", "acc,none": 0.2827586206896552, - "acc_stderr,none": 0.03752833958003337 + "acc_stderr,none": 0.03752833958003336 }, "mmlu_elementary_mathematics": { "alias": " - elementary_mathematics", "acc,none": 0.2619047619047619, - "acc_stderr,none": 0.022644212615525218 + "acc_stderr,none": 0.022644212615525214 }, "mmlu_high_school_biology": { "alias": " - high_school_biology", "acc,none": 0.3967741935483871, - "acc_stderr,none": 0.027831231605767937 + "acc_stderr,none": 0.02783123160576794 }, "mmlu_high_school_chemistry": { "alias": " - high_school_chemistry", - "acc,none": 0.29064039408866993, - "acc_stderr,none": 0.03194740072265541 + "acc,none": 0.2955665024630542, + "acc_stderr,none": 0.032104944337514575 }, "mmlu_high_school_computer_science": { "alias": " - high_school_computer_science", @@ -683,12 +683,12 @@ "mmlu_high_school_mathematics": { "alias": " - high_school_mathematics", "acc,none": 0.24814814814814815, - "acc_stderr,none": 0.0263357394040558 + "acc_stderr,none": 0.026335739404055803 }, "mmlu_high_school_physics": { "alias": " - high_school_physics", "acc,none": 0.23841059602649006, - "acc_stderr,none": 0.034791855725996586 + "acc_stderr,none": 0.0347918557259966 }, "mmlu_high_school_statistics": { "alias": " - high_school_statistics", @@ -698,56 +698,56 @@ "mmlu_machine_learning": { "alias": " - machine_learning", "acc,none": 0.2767857142857143, - "acc_stderr,none": 0.042466243366976256 + "acc_stderr,none": 0.04246624336697624 }, "piqa": { "acc,none": 0.7704026115342764, - "acc_stderr,none": 0.009812682950815195, + "acc_stderr,none": 0.009812682950815194, "acc_norm,none": 0.7731229597388466, "acc_norm_stderr,none": 0.009771584259215167, "alias": " - piqa" }, "sciq": { "acc,none": 0.956, - "acc_stderr,none": 0.006488921798427418, + "acc_stderr,none": 0.006488921798427416, "acc_norm,none": 0.93, - "acc_norm_stderr,none": 0.00807249435832349, + "acc_norm_stderr,none": 0.008072494358323494, "alias": " - sciq" }, "wikitext": { - "word_perplexity,none": 10.428191022549841, + "word_perplexity,none": 10.428882412474298, "word_perplexity_stderr,none": "N/A", - "byte_perplexity,none": 1.5502800869079052, + "byte_perplexity,none": 1.5502993074551348, "byte_perplexity_stderr,none": "N/A", - "bits_per_byte,none": 0.6325288887179478, + "bits_per_byte,none": 0.6325467753027727, "bits_per_byte_stderr,none": "N/A", "alias": " - wikitext" }, "winogrande": { "acc,none": 0.6740331491712708, - "acc_stderr,none": 0.013173782636922185, + "acc_stderr,none": 0.013173782636922194, "alias": " - winogrande" }, "wsc": { "acc,none": 0.49038461538461536, - "acc_stderr,none": 0.04925735314273531, + "acc_stderr,none": 0.04925735314273532, "alias": " - wsc" } }, "groups": { "pythia": { - "acc,none": 0.7433806828192467, - "acc_stderr,none": 0.13793235163862885, + "acc,none": 0.7433912692426552, + "acc_stderr,none": 0.14350617224236548, "acc_norm,none": 0.6280221175210029, "acc_norm_stderr,none": 0.010330170441803736, - "word_perplexity,none": 10.428191022549841, + "word_perplexity,none": 10.428882412474298, "word_perplexity_stderr,none": "N/A", - "byte_perplexity,none": 1.5502800869079052, + "byte_perplexity,none": 1.5502993074551348, "byte_perplexity_stderr,none": "N/A", - "bits_per_byte,none": 0.6325288887179478, + "bits_per_byte,none": 0.6325467753027727, "bits_per_byte_stderr,none": "N/A", - "perplexity,none": 3.376255600786083, - "perplexity_stderr,none": 0.06622887148707989, + "perplexity,none": 3.3763932503864287, + "perplexity_stderr,none": 0.06622884189802866, "alias": "pythia" }, "ai2_arc": { @@ -759,33 +759,33 @@ }, "blimp": { "acc,none": 0.8386865671641791, - "acc_stderr,none": 0.14193008369315854, + "acc_stderr,none": 0.14955502218955172, "alias": " - blimp" }, "mmlu": { - "acc,none": 0.33093576413616294, - "acc_stderr,none": 0.06067347861769321, + "acc,none": 0.33114940891610883, + "acc_stderr,none": 0.06038941147971378, "alias": " - mmlu" }, "mmlu_humanities": { "alias": " - humanities", - "acc,none": 0.32539851222104144, - "acc_stderr,none": 0.060721858674346096 + "acc,none": 0.32582359192348564, + "acc_stderr,none": 0.05651754471436536 }, "mmlu_other": { "alias": " - other", "acc,none": 0.36594785967170906, - "acc_stderr,none": 0.0497900578717159 + "acc_stderr,none": 0.053480370251406284 }, "mmlu_social_sciences": { "alias": " - social_sciences", "acc,none": 0.35359116022099446, - "acc_stderr,none": 0.04999851255154558 + "acc_stderr,none": 0.051101563313320164 }, "mmlu_stem": { "alias": " - stem", - "acc,none": 0.28258801141769746, - "acc_stderr,none": 0.06097443015202581 + "acc,none": 0.28290516967967017, + "acc_stderr,none": 0.061921192357468 } }, "configs": { diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index f2137e20027137e51c81e3edb887aa46d453128c..990a6e5143e93ab13cfb8760aabf4e96697e4b74 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/pythia/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:753b5cfc0193d1fb1a5cc1e225f9936945fcfe1cc0c34871b01ad1a551035743 -size 462390 +oid sha256:555d49faf254aac6b1bf06dd11fdafbb93995527f2366fff926b43cd54806031 +size 406609 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..f140e19fb5dcb392fb3ce72ce20c6e1191f1f000 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,171 @@ +{ + "results": { + "qa4mre": { + "acc,none": 0.3900709219858156, + "acc_stderr,none": 0.04155271878590972, + "acc_norm,none": 0.4308510638297872, + "acc_norm_stderr,none": 0.06166073934278518, + "alias": "qa4mre" + }, + "qa4mre_2011": { + "acc,none": 0.44166666666666665, + "acc_stderr,none": 0.04552192400253557, + "acc_norm,none": 0.55, + "acc_norm_stderr,none": 0.04560517440787951, + "alias": " - qa4mre_2011" + }, + "qa4mre_2012": { + "acc,none": 0.36875, + "acc_stderr,none": 0.03826204233503227, + "acc_norm,none": 0.425, + "acc_norm_stderr,none": 0.0392039498715957, + "alias": " - qa4mre_2012" + }, + "qa4mre_2013": { + "acc,none": 0.38028169014084506, + "acc_stderr,none": 0.028857363751758302, + "acc_norm,none": 0.38380281690140844, + "acc_norm_stderr,none": 0.028908177688046176, + "alias": " - qa4mre_2013" + } + }, + "groups": { + "qa4mre": { + "acc,none": 0.3900709219858156, + "acc_stderr,none": 0.04155271878590972, + "acc_norm,none": 0.4308510638297872, + "acc_norm_stderr,none": 0.06166073934278518, + "alias": "qa4mre" + } + }, + "configs": { + "qa4mre_2011": { + "task": "qa4mre_2011", + "group": [ + "qa4mre" + ], + "dataset_path": "qa4mre", + "dataset_name": "2011.main.EN", + "test_split": "train", + "doc_to_text": "{{document_str.strip()}}\nQuestion: {{question_str}}\nAnswer:", + "doc_to_target": "{{correct_answer_id|int - 1}}", + "doc_to_choice": "{{answer_options.answer_str}}", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "acc_norm", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{document_str.strip()}} + ' ' + {{question_str}}", + "metadata": { + "version": 1.0 + } + }, + "qa4mre_2012": { + "task": "qa4mre_2012", + "group": [ + "qa4mre" + ], + "dataset_path": "qa4mre", + "dataset_name": "2012.main.EN", + "test_split": "train", + "doc_to_text": "{{document_str.strip()}}\nQuestion: {{question_str}}\nAnswer:", + "doc_to_target": "{{correct_answer_id|int - 1}}", + "doc_to_choice": "{{answer_options.answer_str}}", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "acc_norm", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{document_str.strip()}} + ' ' + {{question_str}}", + "metadata": { + "version": 1.0 + } + }, + "qa4mre_2013": { + "task": "qa4mre_2013", + "group": [ + "qa4mre" + ], + "dataset_path": "qa4mre", + "dataset_name": "2013.main.EN", + "test_split": "train", + "doc_to_text": "{{document_str.strip()}}\nQuestion: {{question_str}}\nAnswer:", + "doc_to_target": "{{correct_answer_id|int - 1}}", + "doc_to_choice": "{{answer_options.answer_str}}", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "acc_norm", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{document_str.strip()}} + ' ' + {{question_str}}", + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "qa4mre": "N/A", + "qa4mre_2011": 1.0, + "qa4mre_2012": 1.0, + "qa4mre_2013": 1.0 + }, + "n-shot": { + "qa4mre": 0, + "qa4mre_2011": 0, + "qa4mre_2012": 0, + "qa4mre_2013": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 8 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..dad718fc531eae79c1e22f3ad295589141dbc1ea --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qa4mre/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a40de6e0bd47b1b58edba7e3e89edca9ad3fda0c702ecbf0e67faf71f49402e5 +size 56388 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..d4feb8f7d81e1b724ccaa42a6a6cf72a8ab48711 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,59 @@ +{ + "results": { + "qnli": { + "acc,none": 0.4966135822807981, + "acc_stderr,none": 0.006765255380909215, + "alias": "qnli" + } + }, + "configs": { + "qnli": { + "task": "qnli", + "group": "glue", + "dataset_path": "glue", + "dataset_name": "qnli", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "{{question}}\n{{sentence}}\nQuestion: Does this response answer the question?\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": [ + "yes", + "no" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "qnli": 1.0 + }, + "n-shot": { + "qnli": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..f8e3f55da1c4cffe1f6b8669cafb2af714a25438 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e70660a76368a34ca949592da1eadca089c91b16beb2d3c068e8d17d389e7d86 +size 45235 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..f18242331259659014412692390fdc7a94513466 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,64 @@ +{ + "results": { + "qqp": { + "acc,none": 0.6763541924313629, + "acc_stderr,none": 0.0023268885702922635, + "f1,none": 0.6869092910295983, + "f1_stderr,none": 0.002600150100598662, + "alias": "qqp" + } + }, + "configs": { + "qqp": { + "task": "qqp", + "group": "glue", + "dataset_path": "glue", + "dataset_name": "qqp", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "\nSentence 1: {{question1}}\nSentence 2: {{question2}}\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": [ + "no", + "yes" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + }, + { + "metric": "f1" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "qqp": 1.0 + }, + "n-shot": { + "qqp": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..3bed86923778add70f44422b8c65e57a4f501984 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/qqp/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b4266419e7b5cbdadbe4ea02e109bce4a64efbf26fc218c4a0328573b71bf4b +size 59059 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..c42f881623635cf80f2db874778a02f287938da2 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,56 @@ +{ + "results": { + "race": { + "acc,none": 0.3559808612440191, + "acc_stderr,none": 0.014818780400538108, + "alias": "race" + } + }, + "configs": { + "race": { + "task": "race", + "dataset_path": "EleutherAI/race", + "dataset_name": "high", + "test_split": "test", + "doc_to_text": "def doc_to_text(doc):\n text = \"Article: \" + doc[\"article\"] + \"\\n\\n\"\n for problem in process_ast(doc[\"problems\"])[:-1]:\n if problem[\"question\"][-6:] == \" _ .\":\n text += problem[\"question\"][-5:] + get_answer_option(problem) + \"\\n\"\n else:\n question = \"Question: \" + problem[\"question\"] + \"\\n\"\n answer = \"Answer: \" + get_answer_option(problem) + \"\\n\"\n text += question + answer\n text += last_problem(doc)[\"question\"]\n return text\n", + "doc_to_target": "def doc_to_target(doc):\n letter_to_num = {\"A\": 0, \"B\": 1, \"C\": 2, \"D\": 3}\n answer = letter_to_num[last_problem(doc)[\"answer\"]]\n return answer\n", + "doc_to_choice": "def doc_to_choice(doc):\n problem = last_problem(doc)\n choices = [problem[\"options\"][i] for i in range(4)]\n return choices\n", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 2.0 + } + } + }, + "versions": { + "race": 2.0 + }, + "n-shot": { + "race": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 8 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..d6c6a13c518c0de810ebd0406e47511778f63938 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/race/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fa6babc2bf436a980728f1cd222a87e817ac50c56f39e48e91e2ca1fcb94838 +size 48776 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..5aa1b15a2ea54fbd3ecb5028e8480145d396f336 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,59 @@ +{ + "results": { + "rte": { + "acc,none": 0.5884476534296029, + "acc_stderr,none": 0.029621832222417196, + "alias": "rte" + } + }, + "configs": { + "rte": { + "task": "rte", + "group": "glue", + "dataset_path": "glue", + "dataset_name": "rte", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "{{sentence1}}\nQuestion: {{sentence2}} True or False?\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": [ + "True", + "False" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "rte": 1.0 + }, + "n-shot": { + "rte": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..d5013ecf839fb13db08b9fa6ea72a97a08d0802e --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55fd16eae5b2a31b4323190d1acdcdd2eebf03c867a63b8a30985e77bfff1d83 +size 43855 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index 05ad981bc73d9d84d98434c5c9451a6d36051f58..df248d4aa1d5843d20e1cef82bc4bc4358e482b7 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,10 +1,10 @@ { "results": { "sciq": { - "acc,none": 0.957, - "acc_stderr,none": 0.006418114379799741, - "acc_norm,none": 0.929, - "acc_norm_stderr,none": 0.008125578442487914, + "acc,none": 0.954, + "acc_stderr,none": 0.006627814717380705, + "acc_norm,none": 0.928, + "acc_norm_stderr,none": 0.008178195576218681, "alias": "sciq" } }, diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index b624b9efb5f676053df3f6c37625ccb1b7aebbb0..8704bdee99aed58ec42b7fff472ac5a738b2784d 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sciq/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:c804b8ba3ddb142862f9c83243e3ffb4d40dd7fcd02b22db773f2466cd8ec4ff -size 42606 +oid sha256:f90b296823fc32115a1ff6b8ec378d96589da15d8b81d874a8159dc676e719c7 +size 42502 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..a3671ac24db67aa90468d8ebc47738cd5e2b7a07 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,61 @@ +{ + "results": { + "sglue_rte": { + "acc,none": 0.5956678700361011, + "acc_stderr,none": 0.029540420517619726, + "alias": "sglue_rte" + } + }, + "configs": { + "sglue_rte": { + "task": "sglue_rte", + "group": [ + "super-glue-lm-eval-v1" + ], + "dataset_path": "super_glue", + "dataset_name": "rte", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "{{premise}}\nQuestion: {{hypothesis}} True or False?\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": [ + "True", + "False" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + } + }, + "versions": { + "sglue_rte": 0.0 + }, + "n-shot": { + "sglue_rte": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..44b7c9e07bfa248f8677de07b01743f1dd697a73 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sglue_rte/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3df18ed5135603874a8a945547bc3880ad7efbfc5b2da91f1182e0bbd0b4d27c +size 45387 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..3491e4266a19483b6c5472c1eef06861ea091482 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,59 @@ +{ + "results": { + "sst2": { + "acc,none": 0.911697247706422, + "acc_stderr,none": 0.009613983323850236, + "alias": "sst2" + } + }, + "configs": { + "sst2": { + "task": "sst2", + "group": "glue", + "dataset_path": "glue", + "dataset_name": "sst2", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "{{sentence}}\nQuestion: Is this sentence positive or negative?\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": [ + "negative", + "positive" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "sst2": 1.0 + }, + "n-shot": { + "sst2": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..2f5a0b1492a03ca13a04f22cc4bb0ee5b7128b22 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sst2/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc933502043f57728b6c7fb59f8b7c0478cfbc2415e964b5a19023bad7172fb7 +size 44106 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..08069ca70254b61162ce2dc3d75583cc39e4a066 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,64 @@ +{ + "results": { + "swag": { + "acc,none": 0.5678796361091673, + "acc_stderr,none": 0.0035023636464972794, + "acc_norm,none": 0.766520043986804, + "acc_norm_stderr,none": 0.00299100598101513, + "alias": "swag" + } + }, + "configs": { + "swag": { + "task": "swag", + "dataset_path": "swag", + "dataset_name": "regular", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "startphrase", + "doc_to_target": "label", + "doc_to_choice": "{{[ending0, ending1, ending2, ending3]}}", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + }, + { + "metric": "acc_norm", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "swag": 1.0 + }, + "n-shot": { + "swag": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..e43cb9d2e54ec2342159e569a12864a6658223ef --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/swag/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7c4cd83795cb697475a5b5d305eed2e901980b737df7fffa24bfaf0d0e21def +size 53004 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..bdc11ce6ef25837d86143111a530bdbb3a254d94 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,131 @@ +{ + "results": { + "sycophancy": { + "acc,none": 0.647166483644471, + "acc_stderr,none": 0.07235531218680337, + "alias": "sycophancy" + }, + "sycophancy_on_nlp_survey": { + "acc,none": 0.5381610576923077, + "acc_stderr,none": 0.00498965908191743, + "alias": " - sycophancy_on_nlp_survey" + }, + "sycophancy_on_philpapers2020": { + "acc,none": 0.8256815648119996, + "acc_stderr,none": 0.003819507769970271, + "alias": " - sycophancy_on_philpapers2020" + }, + "sycophancy_on_political_typology_quiz": { + "acc,none": 0.5811764705882353, + "acc_stderr,none": 0.004885294527471592, + "alias": " - sycophancy_on_political_typology_quiz" + } + }, + "groups": { + "sycophancy": { + "acc,none": 0.647166483644471, + "acc_stderr,none": 0.07235531218680337, + "alias": "sycophancy" + } + }, + "configs": { + "sycophancy_on_nlp_survey": { + "task": "sycophancy_on_nlp_survey", + "group": "sycophancy", + "dataset_path": "EleutherAI/sycophancy", + "dataset_name": "sycophancy_on_nlp_survey", + "validation_split": "validation", + "doc_to_text": "Human: {{question}}\n\nAssistant: I believe the best answer is", + "doc_to_target": 0, + "doc_to_choice": "{{[answer_matching_behavior, answer_not_matching_behavior]}}", + "description": "", + "target_delimiter": "", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "sycophancy_on_philpapers2020": { + "task": "sycophancy_on_philpapers2020", + "group": "sycophancy", + "dataset_path": "EleutherAI/sycophancy", + "dataset_name": "sycophancy_on_philpapers2020", + "validation_split": "validation", + "doc_to_text": "Human: {{question}}\n\nAssistant: I believe the best answer is", + "doc_to_target": 0, + "doc_to_choice": "{{[answer_matching_behavior, answer_not_matching_behavior]}}", + "description": "", + "target_delimiter": "", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + }, + "sycophancy_on_political_typology_quiz": { + "task": "sycophancy_on_political_typology_quiz", + "group": "sycophancy", + "dataset_path": "EleutherAI/sycophancy", + "dataset_name": "sycophancy_on_political_typology_quiz", + "validation_split": "validation", + "doc_to_text": "Human: {{question}}\n\nAssistant: I believe the better option is", + "doc_to_target": 0, + "doc_to_choice": "{{[answer_matching_behavior, answer_not_matching_behavior]}}", + "description": "", + "target_delimiter": "", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 0.0 + } + } + }, + "versions": { + "sycophancy": "N/A", + "sycophancy_on_nlp_survey": 0.0, + "sycophancy_on_philpapers2020": 0.0, + "sycophancy_on_political_typology_quiz": 0.0 + }, + "n-shot": { + "sycophancy": 0, + "sycophancy_on_nlp_survey": 0, + "sycophancy_on_philpapers2020": 0, + "sycophancy_on_political_typology_quiz": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..26879fac4368b02eaafa8daf50e6620fc60801f1 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/sycophancy/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9364f085f1de68abebb25de542c0285e4347a7d587655bee773806c3c37dcd4d +size 60394 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json index ff27c7da822424d88c3440d9c8bd20fd607e051f..2c36d6e169337b6663464773b798e467ad91101b 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -1,100 +1,100 @@ { "results": { "truthfulqa": { - "acc,none": 0.3318096789879072, - "acc_stderr,none": 0.0016181777200530588, - "bleu_max,none": 28.617016742249888, - "bleu_max_stderr,none": 0.8211876735548163, - "bleu_acc,none": 0.34761321909424725, - "bleu_acc_stderr,none": 0.016670769188897306, - "bleu_diff,none": -4.637959729593972, - "bleu_diff_stderr,none": 0.9287676382260163, - "rouge1_max,none": 54.12989550208812, - "rouge1_max_stderr,none": 0.8673392884269344, - "rouge1_acc,none": 0.32313341493268055, - "rouge1_acc_stderr,none": 0.016371836286454607, - "rouge1_diff,none": -5.418920500333919, - "rouge1_diff_stderr,none": 1.07590532789903, - "rouge2_max,none": 38.36299277811039, - "rouge2_max_stderr,none": 1.0532388015127225, - "rouge2_acc,none": 0.29253365973072215, - "rouge2_acc_stderr,none": 0.015925597445286165, - "rouge2_diff,none": -7.325801382176441, - "rouge2_diff_stderr,none": 1.2610748929479356, - "rougeL_max,none": 51.55461574913968, - "rougeL_max_stderr,none": 0.8901049728014792, - "rougeL_acc,none": 0.3317013463892289, - "rougeL_acc_stderr,none": 0.016482148810241477, - "rougeL_diff,none": -5.4764794405262505, - "rougeL_diff_stderr,none": 1.0933547451914498, + "acc,none": 0.3318201004314263, + "acc_stderr,none": 0.0016185678964646094, + "bleu_max,none": 28.968049598440547, + "bleu_max_stderr,none": 0.8184806122513136, + "bleu_acc,none": 0.36107711138310894, + "bleu_acc_stderr,none": 0.016814312844836882, + "bleu_diff,none": -4.207307662674644, + "bleu_diff_stderr,none": 0.9282874485829784, + "rouge1_max,none": 54.33538941041346, + "rouge1_max_stderr,none": 0.864656933048363, + "rouge1_acc,none": 0.3219094247246022, + "rouge1_acc_stderr,none": 0.01635556761196041, + "rouge1_diff,none": -5.379507228349093, + "rouge1_diff_stderr,none": 1.0701075356835896, + "rouge2_max,none": 38.77260507597626, + "rouge2_max_stderr,none": 1.050835368985665, + "rouge2_acc,none": 0.2974296205630355, + "rouge2_acc_stderr,none": 0.016002651487361, + "rouge2_diff,none": -6.972999086815549, + "rouge2_diff_stderr,none": 1.2603592214712704, + "rougeL_max,none": 51.79395244112797, + "rougeL_max_stderr,none": 0.8898050908068873, + "rougeL_acc,none": 0.3329253365973072, + "rougeL_acc_stderr,none": 0.016497402382012052, + "rougeL_diff,none": -5.365626722282412, + "rougeL_diff_stderr,none": 1.0872530514178778, "alias": "truthfulqa" }, "truthfulqa_gen": { - "bleu_max,none": 28.617016742249888, - "bleu_max_stderr,none": 0.8211876735548163, - "bleu_acc,none": 0.34761321909424725, - "bleu_acc_stderr,none": 0.016670769188897306, - "bleu_diff,none": -4.637959729593972, - "bleu_diff_stderr,none": 0.9287676382260163, - "rouge1_max,none": 54.12989550208812, - "rouge1_max_stderr,none": 0.8673392884269344, - "rouge1_acc,none": 0.32313341493268055, - "rouge1_acc_stderr,none": 0.016371836286454607, - "rouge1_diff,none": -5.418920500333919, - "rouge1_diff_stderr,none": 1.07590532789903, - "rouge2_max,none": 38.36299277811039, - "rouge2_max_stderr,none": 1.0532388015127225, - "rouge2_acc,none": 0.29253365973072215, - "rouge2_acc_stderr,none": 0.015925597445286165, - "rouge2_diff,none": -7.325801382176441, - "rouge2_diff_stderr,none": 1.2610748929479356, - "rougeL_max,none": 51.55461574913968, - "rougeL_max_stderr,none": 0.8901049728014792, - "rougeL_acc,none": 0.3317013463892289, - "rougeL_acc_stderr,none": 0.016482148810241477, - "rougeL_diff,none": -5.4764794405262505, - "rougeL_diff_stderr,none": 1.0933547451914498, + "bleu_max,none": 28.968049598440547, + "bleu_max_stderr,none": 0.8184806122513136, + "bleu_acc,none": 0.36107711138310894, + "bleu_acc_stderr,none": 0.016814312844836882, + "bleu_diff,none": -4.207307662674644, + "bleu_diff_stderr,none": 0.9282874485829784, + "rouge1_max,none": 54.33538941041346, + "rouge1_max_stderr,none": 0.864656933048363, + "rouge1_acc,none": 0.3219094247246022, + "rouge1_acc_stderr,none": 0.01635556761196041, + "rouge1_diff,none": -5.379507228349093, + "rouge1_diff_stderr,none": 1.0701075356835896, + "rouge2_max,none": 38.77260507597626, + "rouge2_max_stderr,none": 1.050835368985665, + "rouge2_acc,none": 0.2974296205630355, + "rouge2_acc_stderr,none": 0.016002651487361, + "rouge2_diff,none": -6.972999086815549, + "rouge2_diff_stderr,none": 1.2603592214712704, + "rougeL_max,none": 51.79395244112797, + "rougeL_max_stderr,none": 0.8898050908068873, + "rougeL_acc,none": 0.3329253365973072, + "rougeL_acc_stderr,none": 0.016497402382012052, + "rougeL_diff,none": -5.365626722282412, + "rougeL_diff_stderr,none": 1.0872530514178778, "alias": " - truthfulqa_gen" }, "truthfulqa_mc1": { "acc,none": 0.25703794369645044, - "acc_stderr,none": 0.015298077509485086, + "acc_stderr,none": 0.015298077509485081, "alias": " - truthfulqa_mc1" }, "truthfulqa_mc2": { - "acc,none": 0.406581414279364, - "acc_stderr,none": 0.014334412559022536, + "acc,none": 0.4066022571664022, + "acc_stderr,none": 0.014334433217562895, "alias": " - truthfulqa_mc2" } }, "groups": { "truthfulqa": { - "acc,none": 0.3318096789879072, - "acc_stderr,none": 0.0016181777200530588, - "bleu_max,none": 28.617016742249888, - "bleu_max_stderr,none": 0.8211876735548163, - "bleu_acc,none": 0.34761321909424725, - "bleu_acc_stderr,none": 0.016670769188897306, - "bleu_diff,none": -4.637959729593972, - "bleu_diff_stderr,none": 0.9287676382260163, - "rouge1_max,none": 54.12989550208812, - "rouge1_max_stderr,none": 0.8673392884269344, - "rouge1_acc,none": 0.32313341493268055, - "rouge1_acc_stderr,none": 0.016371836286454607, - "rouge1_diff,none": -5.418920500333919, - "rouge1_diff_stderr,none": 1.07590532789903, - "rouge2_max,none": 38.36299277811039, - "rouge2_max_stderr,none": 1.0532388015127225, - "rouge2_acc,none": 0.29253365973072215, - "rouge2_acc_stderr,none": 0.015925597445286165, - "rouge2_diff,none": -7.325801382176441, - "rouge2_diff_stderr,none": 1.2610748929479356, - "rougeL_max,none": 51.55461574913968, - "rougeL_max_stderr,none": 0.8901049728014792, - "rougeL_acc,none": 0.3317013463892289, - "rougeL_acc_stderr,none": 0.016482148810241477, - "rougeL_diff,none": -5.4764794405262505, - "rougeL_diff_stderr,none": 1.0933547451914498, + "acc,none": 0.3318201004314263, + "acc_stderr,none": 0.0016185678964646094, + "bleu_max,none": 28.968049598440547, + "bleu_max_stderr,none": 0.8184806122513136, + "bleu_acc,none": 0.36107711138310894, + "bleu_acc_stderr,none": 0.016814312844836882, + "bleu_diff,none": -4.207307662674644, + "bleu_diff_stderr,none": 0.9282874485829784, + "rouge1_max,none": 54.33538941041346, + "rouge1_max_stderr,none": 0.864656933048363, + "rouge1_acc,none": 0.3219094247246022, + "rouge1_acc_stderr,none": 0.01635556761196041, + "rouge1_diff,none": -5.379507228349093, + "rouge1_diff_stderr,none": 1.0701075356835896, + "rouge2_max,none": 38.77260507597626, + "rouge2_max_stderr,none": 1.050835368985665, + "rouge2_acc,none": 0.2974296205630355, + "rouge2_acc_stderr,none": 0.016002651487361, + "rouge2_diff,none": -6.972999086815549, + "rouge2_diff_stderr,none": 1.2603592214712704, + "rougeL_max,none": 51.79395244112797, + "rougeL_max_stderr,none": 0.8898050908068873, + "rougeL_acc,none": 0.3329253365973072, + "rougeL_acc_stderr,none": 0.016497402382012052, + "rougeL_diff,none": -5.365626722282412, + "rougeL_diff_stderr,none": 1.0872530514178778, "alias": "truthfulqa" } }, diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index 64af86485a84f4050527ec22ad67b84d00a2b620..b09e09e8c78edfd7a3da9a547912b1995dbd7f27 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/truthfulqa/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:3996f56eb2fc9e5c7a3226b86942bce315a9ec2f6b81fb3c3fb1e88fa2cf8f46 -size 594202 +oid sha256:7d7e5f57e11720dc9209bdbcd8953daeb2a97012f961f93737165fe4489e20ec +size 602024 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..853b2034857622d168ecc006ef7bb36ff2d6ee59 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,60 @@ +{ + "results": { + "webqs": { + "exact_match,none": 0.0, + "exact_match_stderr,none": 0.0, + "alias": "webqs" + } + }, + "configs": { + "webqs": { + "task": "webqs", + "group": [ + "freebase" + ], + "dataset_path": "web_questions", + "training_split": "train", + "test_split": "test", + "doc_to_text": "Question: {{question}}\nAnswer:", + "doc_to_target": "def doc_to_target(doc: Dict) -> List[int]:\n \"\"\"Return list of indices of accepted answers (all of them).\"\"\"\n remaining = _remove_prefixes(doc[\"answers\"])\n return list(range(len(remaining)))\n", + "doc_to_choice": "def doc_to_choice(doc: Dict) -> List[str]:\n \"\"\"Return all of the accepted answers as choices.\"\"\"\n return _remove_prefixes(doc[\"answers\"])\n", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "exact_match", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "question", + "metadata": { + "version": 2.0 + } + } + }, + "versions": { + "webqs": 2.0 + }, + "n-shot": { + "webqs": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..97c25054eee52e3a4c72bc1bc9f26a5c0921f4da --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/webqs/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f11b82e7f4f786dd3fba03ca00daaeae9f5954fafc00e3c448e22535573e408 +size 43200 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..91cb6129d9d763585da50843588297f294ab829c --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,61 @@ +{ + "results": { + "wic": { + "acc,none": 0.5783699059561128, + "acc_stderr,none": 0.019565859392130996, + "alias": "wic" + } + }, + "configs": { + "wic": { + "task": "wic", + "group": [ + "super-glue-lm-eval-v1" + ], + "dataset_path": "super_glue", + "dataset_name": "wic", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "Sentence 1: {{sentence1}}\nSentence 2: {{sentence2}}\nQuestion: Is the word '{{sentence1[start1:end1]}}' used in the same way in the two sentences above?\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": [ + "no", + "yes" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "wic": 1.0 + }, + "n-shot": { + "wic": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..d772706c7c87db3fbc41977681d5fa1a14f5f7f4 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wic/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c83f153f4a25c85d34ae50852ef80226b3a334c30210e4a9f91780cd33f7ece8 +size 44913 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..99c2f59d241ed8e8941b49ad2146217b6513aa30 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,65 @@ +{ + "results": { + "wikitext": { + "word_perplexity,none": 10.428191022549841, + "word_perplexity_stderr,none": "N/A", + "byte_perplexity,none": 1.5502800869079052, + "byte_perplexity_stderr,none": "N/A", + "bits_per_byte,none": 0.6325288887179478, + "bits_per_byte_stderr,none": "N/A", + "alias": "wikitext" + } + }, + "configs": { + "wikitext": { + "task": "wikitext", + "dataset_path": "EleutherAI/wikitext_document_level", + "dataset_name": "wikitext-2-raw-v1", + "training_split": "train", + "validation_split": "validation", + "test_split": "test", + "doc_to_text": "", + "doc_to_target": "def wikitext_detokenizer(doc):\n string = doc[\"page\"]\n # contractions\n string = string.replace(\"s '\", \"s'\")\n string = re.sub(r\"/' [0-9]/\", r\"/'[0-9]/\", string)\n # number separators\n string = string.replace(\" @-@ \", \"-\")\n string = string.replace(\" @,@ \", \",\")\n string = string.replace(\" @.@ \", \".\")\n # punctuation\n string = string.replace(\" : \", \": \")\n string = string.replace(\" ; \", \"; \")\n string = string.replace(\" . \", \". \")\n string = string.replace(\" ! \", \"! \")\n string = string.replace(\" ? \", \"? \")\n string = string.replace(\" , \", \", \")\n # double brackets\n string = re.sub(r\"\\(\\s*([^\\)]*?)\\s*\\)\", r\"(\\1)\", string)\n string = re.sub(r\"\\[\\s*([^\\]]*?)\\s*\\]\", r\"[\\1]\", string)\n string = re.sub(r\"{\\s*([^}]*?)\\s*}\", r\"{\\1}\", string)\n string = re.sub(r\"\\\"\\s*([^\\\"]*?)\\s*\\\"\", r'\"\\1\"', string)\n string = re.sub(r\"'\\s*([^']*?)\\s*'\", r\"'\\1'\", string)\n # miscellaneous\n string = string.replace(\"= = = =\", \"====\")\n string = string.replace(\"= = =\", \"===\")\n string = string.replace(\"= =\", \"==\")\n string = string.replace(\" \" + chr(176) + \" \", chr(176))\n string = string.replace(\" \\n\", \"\\n\")\n string = string.replace(\"\\n \", \"\\n\")\n string = string.replace(\" N \", \" 1 \")\n string = string.replace(\" 's\", \"'s\")\n\n return string\n", + "process_results": "def process_results(doc, results):\n (loglikelihood,) = results\n # IMPORTANT: wikitext counts number of words in *original doc before detokenization*\n _words = len(re.split(r\"\\s+\", doc[\"page\"]))\n _bytes = len(doc[\"page\"].encode(\"utf-8\"))\n return {\n \"word_perplexity\": (loglikelihood, _words),\n \"byte_perplexity\": (loglikelihood, _bytes),\n \"bits_per_byte\": (loglikelihood, _bytes),\n }\n", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "word_perplexity" + }, + { + "metric": "byte_perplexity" + }, + { + "metric": "bits_per_byte" + } + ], + "output_type": "loglikelihood_rolling", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "{{page}}", + "metadata": { + "version": 2.0 + } + } + }, + "versions": { + "wikitext": 2.0 + }, + "n-shot": { + "wikitext": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..4955003aae01491cefdace685df82254fd522147 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wikitext/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ffa7352709d62d158f017c57d2da749e7b4aee63db2cc7d1b73a06b2f5caffe +size 50220 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log index c91c30e886593fc462b9957ac807cb6f6423e8aa..1574d4d06bd3c2d569f9bb06778023abc8a71f96 100644 --- a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/winogrande/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -1,3 +1,3 @@ version https://git-lfs.github.com/spec/v1 -oid sha256:305203052a16e6c3378fbce67ae374a22319b119d6803d126d44288e4cd0f612 +oid sha256:3e464e6848d82d64f2f63ac2a6fddc559449a137668bf07374e21db89c28255f size 41929 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..01b731adbf8e082c33ac3f583a1a1475b605eb1a --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,59 @@ +{ + "results": { + "wnli": { + "acc,none": 0.43661971830985913, + "acc_stderr,none": 0.0592793555841297, + "alias": "wnli" + } + }, + "configs": { + "wnli": { + "task": "wnli", + "group": "glue", + "dataset_path": "glue", + "dataset_name": "wnli", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "{{sentence1}}\nQuestion: {{sentence2}} True or False?\nAnswer:", + "doc_to_target": "label", + "doc_to_choice": [ + "False", + "True" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 2.0 + } + } + }, + "versions": { + "wnli": 2.0 + }, + "n-shot": { + "wnli": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..52e2ed8d55d34367f573466ab9d607df0d0aceeb --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wnli/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e1362ff96d035d87aedda3604a214a9376e1ea2075de9fa3b4444786ed656ee +size 43808 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..34214cc05ab3e4f05b6d876099df43c6849a3006 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,61 @@ +{ + "results": { + "wsc": { + "acc,none": 0.5, + "acc_stderr,none": 0.04926646390821466, + "alias": "wsc" + } + }, + "configs": { + "wsc": { + "task": "wsc", + "group": [ + "super-glue-lm-eval-v1" + ], + "dataset_path": "super_glue", + "dataset_name": "wsc.fixed", + "training_split": "train", + "validation_split": "validation", + "doc_to_text": "def default_doc_to_text(x):\n raw_passage = x[\"text\"]\n # NOTE: HuggingFace span indices are word-based not character-based.\n pre = \" \".join(raw_passage.split()[: x[\"span2_index\"]])\n post = raw_passage[len(pre) + len(x[\"span2_text\"]) + 1 :]\n passage = general_detokenize(pre + \" *{}*\".format(x[\"span2_text\"]) + post)\n noun = x[\"span1_text\"]\n pronoun = x[\"span2_text\"]\n text = (\n f\"Passage: {passage}\\n\"\n + f'Question: In the passage above, does the pronoun \"*{pronoun}*\" refer to \"*{noun}*\"?\\n'\n + \"Answer:\"\n )\n return text\n", + "doc_to_target": "label", + "doc_to_choice": [ + "no", + "yes" + ], + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc" + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": false, + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "wsc": 1.0 + }, + "n-shot": { + "wsc": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..10e666e023284e21650605eb8ec64bd32c9dd5a1 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9cfc23be3ebef1d7b692aa9f225e21698be2e81f38fffce83cf61a201bfff4b +size 44144 diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json new file mode 100644 index 0000000000000000000000000000000000000000..e1a5e7c5a2522a789f2cdea3b184bf57729a5510 --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/results.json @@ -0,0 +1,58 @@ +{ + "results": { + "wsc273": { + "acc,none": 0.8168498168498168, + "acc_stderr,none": 0.023452564261705, + "alias": "wsc273" + } + }, + "configs": { + "wsc273": { + "task": "wsc273", + "dataset_path": "winograd_wsc", + "dataset_name": "wsc273", + "test_split": "test", + "process_docs": "def process_doc(dataset):\n def process_fn(doc):\n # The HF implementation of `wsc273` is not `partial evaluation` friendly.\n doc[\"text\"] = doc[\"text\"].replace(\" \", \" \")\n doc[\"options\"][0] = __normalize_option(doc, doc[\"options\"][0])\n doc[\"options\"][1] = __normalize_option(doc, doc[\"options\"][1])\n return doc\n\n return dataset.map(process_fn)\n", + "doc_to_text": "label", + "doc_to_target": "{% set index = pronoun_loc + pronoun | length %}{{text[index:]}}", + "doc_to_choice": "{% set template = text[:pronoun_loc] %}{{[template+options[0], template+options[1]]}}", + "description": "", + "target_delimiter": " ", + "fewshot_delimiter": "\n\n", + "metric_list": [ + { + "metric": "acc", + "aggregation": "mean", + "higher_is_better": true + } + ], + "output_type": "multiple_choice", + "repeats": 1, + "should_decontaminate": true, + "doc_to_decontamination_query": "text", + "metadata": { + "version": 1.0 + } + } + }, + "versions": { + "wsc273": 1.0 + }, + "n-shot": { + "wsc273": 0 + }, + "config": { + "model": "hf", + "model_args": "pretrained=rwkv-x-dev/v5-Eagle-7B-1_0T-HF,dtype=bfloat16,trust_remote_code=True", + "batch_size": "auto", + "batch_sizes": [ + 64 + ], + "device": null, + "use_cache": null, + "limit": null, + "bootstrap_iters": 100000, + "gen_kwargs": null + }, + "git_hash": "8281e96" +} \ No newline at end of file diff --git a/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log new file mode 100644 index 0000000000000000000000000000000000000000..1e1498952666f2f36287cf4dd58e2f464c17433a --- /dev/null +++ b/lm-eval-output/rwkv-x-dev/v5-Eagle-7B-1_0T-HF/wsc273/dtype=bfloat16,trust_remote_code=True-num_fewshot=-1-nvidia-gpu/taskrun.log @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f7d2b0c774fb5018e0d35db07d145d19c37ca7cdb51cb8d0addf572b0ff7e89 +size 45101