model large_stringclasses 19
values | step float64 1k 950k ⌀ | task large_stringclasses 305
values | metric large_stringclasses 3
values | metric_filter large_stringclasses 1
value | score float64 0 5.6 | score_stderr float64 0 0.25 ⌀ | task_type large_stringclasses 3
values | parent_task large_stringclasses 12
values | task_display_name large_stringclasses 305
values | num_fewshot float64 0 5 ⌀ | task_formulation large_stringclasses 2
values | higher_is_better bool 2
classes | language large_stringclasses 2
values | n_samples float64 100 87k | subtask_tree large_stringclasses 20
values | batch_size float64 2.1M 2.1M ⌀ | tokens_trained int64 2.1B 36,000B | eval_date large_stringdate 2026-02-16 08:28:42 2026-02-18 13:45:55 | result_source large_stringlengths 120 166 ⌀ | eval_config dict | task_config dict | model_path large_stringclasses 338
values | is_checkpoint bool 2
classes | model_display_name large_stringclasses 19
values | train_batch_size int64 2.1M 4.19M ⌀ |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
HPLT2c_deu | null | deu_qa_rc | acc_norm | none | 0.561757 | null | task_group | deu_base_easy | deu_qa_rc | null | null | true | deu | 30,715 | {"deu_qa_rc": ["deu_arc_challenge_rc", "deu_arc_easy_rc", "deu_csqa_rc", "deu_csqa_easy_rc", "deu_siqa_rc", "deu_siqa_easy_rc", "deu_piqa_rc", "deu_piqa_easy_rc", "deu_csprp_rc", "deu_csnli_rc", "deu_csnli_easy_rc", "deu_mmmlu_filtered_rc"], "deu_mmmlu_filtered_rc": ["deu_mmmlu_abstract_algebra_rc", "deu_mmmlu_college_... | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | null | null | null | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | deu_siqa_bpb | bits_per_byte | none | 0.689631 | 0.00463 | benchmark | null | deu_siqa_bpb | 5 | bits_per_byte | false | deu | 1,954 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "ellamind/siqa-german-preview",
"doc_to_choice": null,
"doc_to_target": " {{correct_answer}}",
"doc_to_text": "Frage: {{context}} {{question}}\nAntwort:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | deu_siqa_easy_rc | acc | none | 0.514841 | 0.011309 | benchmark | null | deu_siqa_easy_rc | 5 | rank_choice | true | deu | 1,954 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "ellamind/siqa-german-preview",
"doc_to_choice": "{{choices}}",
"doc_to_target": "{{answer_idx}}",
"doc_to_text": "Frage: {{context}} {{question}}\nAntwort:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": fals... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | deu_siqa_easy_rc | acc_norm | none | 0.639713 | 0.010863 | benchmark | null | deu_siqa_easy_rc | 5 | rank_choice | true | deu | 1,954 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "ellamind/siqa-german-preview",
"doc_to_choice": "{{choices}}",
"doc_to_target": "{{answer_idx}}",
"doc_to_text": "Frage: {{context}} {{question}}\nAntwort:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": fals... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | deu_siqa_rc | acc | none | 0.574207 | 0.011189 | benchmark | null | deu_siqa_rc | 5 | rank_choice | true | deu | 1,954 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "ellamind/siqa-german-preview",
"doc_to_choice": "{{choices}}",
"doc_to_target": "{{answer_idx}}",
"doc_to_text": "Frage: {{context}} {{question}}\nAntwort:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": fals... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | deu_siqa_rc | acc_norm | none | 0.548618 | 0.01126 | benchmark | null | deu_siqa_rc | 5 | rank_choice | true | deu | 1,954 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "ellamind/siqa-german-preview",
"doc_to_choice": "{{choices}}",
"doc_to_target": "{{answer_idx}}",
"doc_to_text": "Frage: {{context}} {{question}}\nAntwort:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": fals... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | drop_bpb | bits_per_byte | none | 1.958088 | 0.010973 | benchmark | null | drop_bpb | 5 | bits_per_byte | false | eng | 9,536 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "EleutherAI/drop",
"doc_to_choice": null,
"doc_to_target": " {% if answer.spans %}{{answer.spans[0]}}{% elif answer.number %}{{answer.number}}{% else %}{{answer.date.day}} {{answer.date.month}} {{answer.date.year}}{% endif %}",
"doc_to_text": "Passage: {{passage}}\nQues... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | drop_rc | acc | none | 0.281265 | 0.004912 | benchmark | null | drop_rc | 5 | rank_choice | true | eng | 8,380 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "allenai/drop-gen2mc",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}",
"doc_to_text": "Passage: {{passage_original}}\nQuestion: {{question_original}}\nAnswer:",
"generation_kwargs": null,
"output_type": "mult... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | drop_rc | acc_norm | none | 0.263126 | 0.00481 | benchmark | null | drop_rc | 5 | rank_choice | true | eng | 8,380 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "allenai/drop-gen2mc",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}",
"doc_to_text": "Passage: {{passage_original}}\nQuestion: {{question_original}}\nAnswer:",
"generation_kwargs": null,
"output_type": "mult... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | eng_base_easy | acc | none | 0.411789 | null | eval_suite | null | eng_base_easy | null | null | true | eng | 86,236 | {"eng_base_easy": ["qa_rc"], "qa_rc": ["arc_easy_rc", "arc_challenge_rc", "mmlu_rc", "csqa_rc", "hellaswag_rc", "winogrande_rc", "socialiqa_rc", "piqa_rc", "coqa_rc", "drop_rc", "jeopardy_rc", "naturalqs_rc", "squad_rc", "sciq_rc", "qasper_yesno_rc", "lambada_rc", "medmcqa_rc", "medqa_rc"], "mmlu_rc": ["mmlu_abstract_a... | null | 100,000,000,000 | 2026-02-17T11:43:01.575599+00:00 | null | null | null | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | eng_base_easy | acc_norm | none | 0.416695 | null | eval_suite | null | eng_base_easy | null | null | true | eng | 81,083 | {"eng_base_easy": ["qa_rc"], "qa_rc": ["arc_easy_rc", "arc_challenge_rc", "mmlu_rc", "csqa_rc", "hellaswag_rc", "winogrande_rc", "socialiqa_rc", "piqa_rc", "coqa_rc", "drop_rc", "jeopardy_rc", "naturalqs_rc", "squad_rc", "sciq_rc", "qasper_yesno_rc", "medmcqa_rc", "medqa_rc"], "mmlu_rc": ["mmlu_abstract_algebra_rc", "m... | null | 100,000,000,000 | 2026-02-17T11:43:01.575599+00:00 | null | null | null | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | eng_base_easy | bits_per_byte | none | 1.378868 | null | eval_suite | null | eng_base_easy | null | null | false | eng | 86,964 | {"eng_base_easy": ["math_bpb", "code_bpb", "qa_bpb"], "math_bpb": ["minerva_math_algebra_bpb", "minerva_math_counting_and_probability_bpb", "minerva_math_geometry_bpb", "minerva_math_intermediate_algebra_bpb", "minerva_math_number_theory_bpb", "minerva_math_prealgebra_bpb", "minerva_math_precalculus_bpb"], "code_bpb": ... | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | null | null | null | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | hellaswag_bpb | bits_per_byte | none | 1.047292 | 0.002084 | benchmark | null | hellaswag_bpb | 5 | bits_per_byte | false | eng | 10,042 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "allenai/hellaswag",
"doc_to_choice": null,
"doc_to_target": " {{endings[label | int]}}",
"doc_to_text": "{{activity_label}}: {{ctx_a}} {{ctx_b}}",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_ha... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | hellaswag_rc | acc | none | 0.2525 | 0.004345 | benchmark | null | hellaswag_rc | 5 | rank_choice | true | eng | 10,000 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "allenai/hellaswag",
"doc_to_choice": "{{endings}}",
"doc_to_target": "{{label | int}}",
"doc_to_text": "{{activity_label}}: {{ctx_a}} {{ctx_b}}",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"task_h... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | hellaswag_rc | acc_norm | none | 0.2478 | 0.004318 | benchmark | null | hellaswag_rc | 5 | rank_choice | true | eng | 10,000 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "allenai/hellaswag",
"doc_to_choice": "{{endings}}",
"doc_to_target": "{{label | int}}",
"doc_to_text": "{{activity_label}}: {{ctx_a}} {{ctx_b}}",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"task_h... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | jeopardy_bpb | bits_per_byte | none | 1.360302 | 0.018296 | benchmark | null | jeopardy_bpb | 5 | bits_per_byte | false | eng | 2,116 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "mosaicml_gauntlet",
"dataset_path": "soldni/jeopardy",
"doc_to_choice": null,
"doc_to_target": " {{continuation}}",
"doc_to_text": "Category: {{context.split(': ', 1)[0]}}\nClue: {{context.split(': ', 1)[1]}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repea... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | jeopardy_rc | acc | none | 0.462125 | 0.010885 | benchmark | null | jeopardy_rc | 5 | rank_choice | true | eng | 2,099 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "allenai/jeopardy-gen2mc",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}",
"doc_to_text": "Category: {{context_original.split(': ', 1)[0]}}\nQuestion: {{context_original.split(': ', 1)[1]}}\nAnswer:",
"generati... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | jeopardy_rc | acc_norm | none | 0.396379 | 0.010679 | benchmark | null | jeopardy_rc | 5 | rank_choice | true | eng | 2,099 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "allenai/jeopardy-gen2mc",
"doc_to_choice": "{{choices.text}}",
"doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}",
"doc_to_text": "Category: {{context_original.split(': ', 1)[0]}}\nQuestion: {{context_original.split(': ', 1)[1]}}\nAnswer:",
"generati... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | lab_bench_dbqa_bpb | bits_per_byte | none | 4.576707 | 0.033691 | benchmark | null | lab_bench_dbqa_bpb | 3 | bits_per_byte | false | eng | 520 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "DbQA",
"dataset_path": "futurehouse/lab-bench",
"doc_to_choice": null,
"doc_to_target": " {{ideal}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash": null,
"tas... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | lab_bench_protocolqa_bpb | bits_per_byte | none | 1.630068 | 0.038825 | benchmark | null | lab_bench_protocolqa_bpb | 3 | bits_per_byte | false | eng | 108 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "ProtocolQA",
"dataset_path": "futurehouse/lab-bench",
"doc_to_choice": null,
"doc_to_target": " {{ideal}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash": null,
... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | lambada_bpb | bits_per_byte | none | 1.437608 | null | benchmark | null | lambada_bpb | 0 | bits_per_byte | false | eng | 5,153 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "default",
"dataset_path": "EleutherAI/lambada_openai",
"doc_to_choice": null,
"doc_to_target": "{{text}}",
"doc_to_text": "",
"generation_kwargs": null,
"output_type": "loglikelihood_rolling",
"repeats": 1,
"should_decontaminate": false,
"task_hash": null,
"task_version": "1.0"
... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | lambada_rc | acc | none | 0.363672 | 0.006702 | benchmark | null | lambada_rc | 0 | rank_choice | true | eng | 5,153 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "default",
"dataset_path": "EleutherAI/lambada_openai",
"doc_to_choice": null,
"doc_to_target": " {{text.rsplit(' ', 1)[1]}}",
"doc_to_text": "{{text.rsplit(' ', 1)[0]}}",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | math_bpb | bits_per_byte | none | 1.327082 | null | task_group | eng_base_easy | math_bpb | null | null | false | eng | 5,000 | {"math_bpb": ["minerva_math_algebra_bpb", "minerva_math_counting_and_probability_bpb", "minerva_math_geometry_bpb", "minerva_math_intermediate_algebra_bpb", "minerva_math_number_theory_bpb", "minerva_math_prealgebra_bpb", "minerva_math_precalculus_bpb"]} | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | null | null | null | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mbpp_bpb | bits_per_byte | none | 1.982307 | 0.033944 | benchmark | null | mbpp_bpb | 3 | bits_per_byte | false | eng | 500 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "full",
"dataset_path": "google-research-datasets/mbpp",
"doc_to_choice": null,
"doc_to_target": "def doc_to_target_bpb(doc):\n \"\"\"Code body after function signature (no overlap with context).\n\n Example:\n \\\\n for i in range(len(s)):\\\\n ...\n \"\"\"\n code ... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | medmcqa_bpb | bits_per_byte | none | 1.823879 | 0.014063 | benchmark | null | medmcqa_bpb | 5 | bits_per_byte | false | eng | 4,183 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "openlifescienceai/medmcqa",
"doc_to_choice": null,
"doc_to_target": " {{[opa, opb, opc, opd][cop]}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"ta... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | medmcqa_rc | acc | none | 0.220894 | 0.006415 | benchmark | null | medmcqa_rc | 5 | rank_choice | true | eng | 4,183 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "openlifescienceai/medmcqa",
"doc_to_choice": "{{[opa, opb, opc, opd]}}",
"doc_to_target": "{{cop}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"t... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | medmcqa_rc | acc_norm | none | 0.263447 | 0.006812 | benchmark | null | medmcqa_rc | 5 | rank_choice | true | eng | 4,183 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "openlifescienceai/medmcqa",
"doc_to_choice": "{{[opa, opb, opc, opd]}}",
"doc_to_target": "{{cop}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"t... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | medqa_bpb | bits_per_byte | none | 1.359605 | 0.018932 | benchmark | null | medqa_bpb | 5 | bits_per_byte | false | eng | 1,273 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "davidheineman/medqa-en",
"doc_to_choice": null,
"doc_to_target": " {{answer}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash": null,
"tas... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | medqa_rc | acc | none | 0.210526 | 0.011431 | benchmark | null | medqa_rc | 5 | rank_choice | true | eng | 1,273 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "davidheineman/medqa-en",
"doc_to_choice": "{{choices}}",
"doc_to_target": "{{answer_idx}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"task_hash"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | medqa_rc | acc_norm | none | 0.209741 | 0.011415 | benchmark | null | medqa_rc | 5 | rank_choice | true | eng | 1,273 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": null,
"dataset_path": "davidheineman/medqa-en",
"doc_to_choice": "{{choices}}",
"doc_to_target": "{{answer_idx}}",
"doc_to_text": "Question: {{question}}\nAnswer:",
"generation_kwargs": null,
"output_type": "multiple_choice",
"repeats": 1,
"should_decontaminate": false,
"task_hash"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | minerva_math_algebra_bpb | bits_per_byte | none | 1.434884 | 0.008872 | benchmark | null | minerva_math_algebra_bpb | 4 | bits_per_byte | false | eng | 1,187 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "algebra",
"dataset_path": "EleutherAI/hendrycks_math",
"doc_to_choice": null,
"doc_to_target": " {{solution}}",
"doc_to_text": "Problem:\n{{problem}}\n\nSolution:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash":... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | minerva_math_counting_and_probability_bpb | bits_per_byte | none | 1.170161 | 0.014208 | benchmark | null | minerva_math_counting_and_probability_bpb | 4 | bits_per_byte | false | eng | 474 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "counting_and_probability",
"dataset_path": "EleutherAI/hendrycks_math",
"doc_to_choice": null,
"doc_to_target": " {{solution}}",
"doc_to_text": "Problem:\n{{problem}}\n\nSolution:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": fals... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | minerva_math_geometry_bpb | bits_per_byte | none | 1.36389 | 0.015558 | benchmark | null | minerva_math_geometry_bpb | 4 | bits_per_byte | false | eng | 479 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "geometry",
"dataset_path": "EleutherAI/hendrycks_math",
"doc_to_choice": null,
"doc_to_target": " {{solution}}",
"doc_to_text": "Problem:\n{{problem}}\n\nSolution:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | minerva_math_intermediate_algebra_bpb | bits_per_byte | none | 1.469797 | 0.00995 | benchmark | null | minerva_math_intermediate_algebra_bpb | 4 | bits_per_byte | false | eng | 903 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "intermediate_algebra",
"dataset_path": "EleutherAI/hendrycks_math",
"doc_to_choice": null,
"doc_to_target": " {{solution}}",
"doc_to_text": "Problem:\n{{problem}}\n\nSolution:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | minerva_math_number_theory_bpb | bits_per_byte | none | 1.336563 | 0.011864 | benchmark | null | minerva_math_number_theory_bpb | 4 | bits_per_byte | false | eng | 540 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "number_theory",
"dataset_path": "EleutherAI/hendrycks_math",
"doc_to_choice": null,
"doc_to_target": " {{solution}}",
"doc_to_text": "Problem:\n{{problem}}\n\nSolution:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | minerva_math_prealgebra_bpb | bits_per_byte | none | 1.23569 | 0.010422 | benchmark | null | minerva_math_prealgebra_bpb | 4 | bits_per_byte | false | eng | 871 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "prealgebra",
"dataset_path": "EleutherAI/hendrycks_math",
"doc_to_choice": null,
"doc_to_target": " {{solution}}",
"doc_to_text": "Problem:\n{{problem}}\n\nSolution:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_has... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | minerva_math_precalculus_bpb | bits_per_byte | none | 1.278591 | 0.015545 | benchmark | null | minerva_math_precalculus_bpb | 4 | bits_per_byte | false | eng | 546 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "precalculus",
"dataset_path": "EleutherAI/hendrycks_math",
"doc_to_choice": null,
"doc_to_target": " {{solution}}",
"doc_to_text": "Problem:\n{{problem}}\n\nSolution:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_ha... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_abstract_algebra_bpb | bits_per_byte | none | 1.480533 | 0.148554 | benchmark | null | mmlu_abstract_algebra_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "abstract_algebra",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_h... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_abstract_algebra_rc | acc | none | 0.16 | 0.036845 | benchmark | null | mmlu_abstract_algebra_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "abstract_algebra",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_abstract_algebra_rc | acc_norm | none | 0.17 | 0.037753 | benchmark | null | mmlu_abstract_algebra_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "abstract_algebra",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_anatomy_bpb | bits_per_byte | none | 1.292732 | 0.039576 | benchmark | null | mmlu_anatomy_bpb | 5 | bits_per_byte | false | eng | 135 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "anatomy",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash": nul... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_anatomy_rc | acc | none | 0.296296 | 0.039446 | benchmark | null | mmlu_anatomy_rc | 5 | rank_choice | true | eng | 135 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "anatomy",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: 'Qu... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_anatomy_rc | acc_norm | none | 0.318519 | 0.040248 | benchmark | null | mmlu_anatomy_rc | 5 | rank_choice | true | eng | 135 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "anatomy",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: 'Qu... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_astronomy_bpb | bits_per_byte | none | 1.361239 | 0.062982 | benchmark | null | mmlu_astronomy_bpb | 5 | bits_per_byte | false | eng | 152 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "astronomy",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash": n... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_astronomy_rc | acc | none | 0.210526 | 0.033177 | benchmark | null | mmlu_astronomy_rc | 5 | rank_choice | true | eng | 152 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "astronomy",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: '... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_astronomy_rc | acc_norm | none | 0.256579 | 0.035542 | benchmark | null | mmlu_astronomy_rc | 5 | rank_choice | true | eng | 152 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "astronomy",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: '... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_bpb | bits_per_byte | none | 1.433623 | null | task_group | qa_bpb | mmlu_bpb | null | null | false | eng | 14,042 | {"mmlu_bpb": ["mmlu_abstract_algebra_bpb", "mmlu_anatomy_bpb", "mmlu_astronomy_bpb", "mmlu_business_ethics_bpb", "mmlu_clinical_knowledge_bpb", "mmlu_college_biology_bpb", "mmlu_college_chemistry_bpb", "mmlu_college_computer_science_bpb", "mmlu_college_mathematics_bpb", "mmlu_college_medicine_bpb", "mmlu_college_physic... | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | null | null | null | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_business_ethics_bpb | bits_per_byte | none | 1.532048 | 0.126497 | benchmark | null | mmlu_business_ethics_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "business_ethics",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_ha... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_business_ethics_rc | acc | none | 0.42 | 0.049604 | benchmark | null | mmlu_business_ethics_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "business_ethics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_business_ethics_rc | acc_norm | none | 0.33 | 0.047258 | benchmark | null | mmlu_business_ethics_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "business_ethics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_clinical_knowledge_bpb | bits_per_byte | none | 1.336823 | 0.035128 | benchmark | null | mmlu_clinical_knowledge_bpb | 5 | bits_per_byte | false | eng | 265 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "clinical_knowledge",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_clinical_knowledge_rc | acc | none | 0.218868 | 0.025448 | benchmark | null | mmlu_clinical_knowledge_rc | 5 | rank_choice | true | eng | 265 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "clinical_knowledge",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_clinical_knowledge_rc | acc_norm | none | 0.267925 | 0.027257 | benchmark | null | mmlu_clinical_knowledge_rc | 5 | rank_choice | true | eng | 265 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "clinical_knowledge",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_biology_bpb | bits_per_byte | none | 1.310637 | 0.043286 | benchmark | null | mmlu_college_biology_bpb | 5 | bits_per_byte | false | eng | 144 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_biology",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_ha... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_biology_rc | acc | none | 0.236111 | 0.035514 | benchmark | null | mmlu_college_biology_rc | 5 | rank_choice | true | eng | 144 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_biology",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_biology_rc | acc_norm | none | 0.25 | 0.03621 | benchmark | null | mmlu_college_biology_rc | 5 | rank_choice | true | eng | 144 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_biology",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_chemistry_bpb | bits_per_byte | none | 2.226741 | 0.110052 | benchmark | null | mmlu_college_chemistry_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_chemistry",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_chemistry_rc | acc | none | 0.33 | 0.047258 | benchmark | null | mmlu_college_chemistry_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_chemistry",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_chemistry_rc | acc_norm | none | 0.31 | 0.046482 | benchmark | null | mmlu_college_chemistry_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_chemistry",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_computer_science_bpb | bits_per_byte | none | 1.980096 | 0.109393 | benchmark | null | mmlu_college_computer_science_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_computer_science",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_computer_science_rc | acc | none | 0.32 | 0.046883 | benchmark | null | mmlu_college_computer_science_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_computer_science",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_computer_science_rc | acc_norm | none | 0.24 | 0.042923 | benchmark | null | mmlu_college_computer_science_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_computer_science",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_mathematics_bpb | bits_per_byte | none | 2.365635 | 0.106684 | benchmark | null | mmlu_college_mathematics_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_mathematics",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"tas... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_mathematics_rc | acc | none | 0.14 | 0.034874 | benchmark | null | mmlu_college_mathematics_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_mathematics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_mathematics_rc | acc_norm | none | 0.22 | 0.041633 | benchmark | null | mmlu_college_mathematics_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_mathematics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_medicine_bpb | bits_per_byte | none | 1.361238 | 0.047204 | benchmark | null | mmlu_college_medicine_bpb | 5 | bits_per_byte | false | eng | 173 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_medicine",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_h... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_medicine_rc | acc | none | 0.236994 | 0.032424 | benchmark | null | mmlu_college_medicine_rc | 5 | rank_choice | true | eng | 173 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_medicine",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_medicine_rc | acc_norm | none | 0.225434 | 0.031862 | benchmark | null | mmlu_college_medicine_rc | 5 | rank_choice | true | eng | 173 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_medicine",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_physics_bpb | bits_per_byte | none | 2.34413 | 0.120026 | benchmark | null | mmlu_college_physics_bpb | 5 | bits_per_byte | false | eng | 102 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_physics",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_ha... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_physics_rc | acc | none | 0.147059 | 0.035241 | benchmark | null | mmlu_college_physics_rc | 5 | rank_choice | true | eng | 102 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_physics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_college_physics_rc | acc_norm | none | 0.156863 | 0.036187 | benchmark | null | mmlu_college_physics_rc | 5 | rank_choice | true | eng | 102 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "college_physics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_computer_security_bpb | bits_per_byte | none | 1.530627 | 0.072114 | benchmark | null | mmlu_computer_security_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "computer_security",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_computer_security_rc | acc | none | 0.24 | 0.042923 | benchmark | null | mmlu_computer_security_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "computer_security",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_computer_security_rc | acc_norm | none | 0.3 | 0.046057 | benchmark | null | mmlu_computer_security_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "computer_security",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_conceptual_physics_bpb | bits_per_byte | none | 1.415904 | 0.04384 | benchmark | null | mmlu_conceptual_physics_bpb | 5 | bits_per_byte | false | eng | 235 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "conceptual_physics",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_conceptual_physics_rc | acc | none | 0.323404 | 0.030579 | benchmark | null | mmlu_conceptual_physics_rc | 5 | rank_choice | true | eng | 235 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "conceptual_physics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_conceptual_physics_rc | acc_norm | none | 0.297872 | 0.029896 | benchmark | null | mmlu_conceptual_physics_rc | 5 | rank_choice | true | eng | 235 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "conceptual_physics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_econometrics_bpb | bits_per_byte | none | 1.143167 | 0.073116 | benchmark | null | mmlu_econometrics_bpb | 5 | bits_per_byte | false | eng | 114 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "econometrics",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_econometrics_rc | acc | none | 0.254386 | 0.04097 | benchmark | null | mmlu_econometrics_rc | 5 | rank_choice | true | eng | 114 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "econometrics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_econometrics_rc | acc_norm | none | 0.245614 | 0.040493 | benchmark | null | mmlu_econometrics_rc | 5 | rank_choice | true | eng | 114 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "econometrics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_electrical_engineering_bpb | bits_per_byte | none | 1.917137 | 0.063081 | benchmark | null | mmlu_electrical_engineering_bpb | 5 | bits_per_byte | false | eng | 145 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "electrical_engineering",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_electrical_engineering_rc | acc | none | 0.262069 | 0.036647 | benchmark | null | mmlu_electrical_engineering_rc | 5 | rank_choice | true | eng | 145 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "electrical_engineering",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_electrical_engineering_rc | acc_norm | none | 0.268966 | 0.036952 | benchmark | null | mmlu_electrical_engineering_rc | 5 | rank_choice | true | eng | 145 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "electrical_engineering",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_elementary_mathematics_bpb | bits_per_byte | none | 2.559193 | 0.053839 | benchmark | null | mmlu_elementary_mathematics_bpb | 5 | bits_per_byte | false | eng | 378 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "elementary_mathematics",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_elementary_mathematics_rc | acc | none | 0.214286 | 0.021133 | benchmark | null | mmlu_elementary_mathematics_rc | 5 | rank_choice | true | eng | 378 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "elementary_mathematics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_elementary_mathematics_rc | acc_norm | none | 0.23545 | 0.021852 | benchmark | null | mmlu_elementary_mathematics_rc | 5 | rank_choice | true | eng | 378 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "elementary_mathematics",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_formal_logic_bpb | bits_per_byte | none | 1.988142 | 0.116928 | benchmark | null | mmlu_formal_logic_bpb | 5 | bits_per_byte | false | eng | 126 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "formal_logic",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_formal_logic_rc | acc | none | 0.31746 | 0.041635 | benchmark | null | mmlu_formal_logic_rc | 5 | rank_choice | true | eng | 126 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "formal_logic",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_formal_logic_rc | acc_norm | none | 0.261905 | 0.039325 | benchmark | null | mmlu_formal_logic_rc | 5 | rank_choice | true | eng | 126 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "formal_logic",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_global_facts_bpb | bits_per_byte | none | 1.821181 | 0.064829 | benchmark | null | mmlu_global_facts_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "global_facts",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"task_hash"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_global_facts_rc | acc | none | 0.22 | 0.041633 | benchmark | null | mmlu_global_facts_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "global_facts",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_global_facts_rc | acc_norm | none | 0.21 | 0.040936 | benchmark | null | mmlu_global_facts_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "global_facts",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_biology_bpb | bits_per_byte | none | 1.258708 | 0.039439 | benchmark | null | mmlu_high_school_biology_bpb | 5 | bits_per_byte | false | eng | 310 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_biology",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"tas... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_biology_rc | acc | none | 0.245161 | 0.024472 | benchmark | null | mmlu_high_school_biology_rc | 5 | rank_choice | true | eng | 310 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_biology",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_biology_rc | acc_norm | none | 0.312903 | 0.026378 | benchmark | null | mmlu_high_school_biology_rc | 5 | rank_choice | true | eng | 310 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_biology",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_chemistry_bpb | bits_per_byte | none | 1.797248 | 0.066453 | benchmark | null | mmlu_high_school_chemistry_bpb | 5 | bits_per_byte | false | eng | 203 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_chemistry",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": false,
"t... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_chemistry_rc | acc | none | 0.197044 | 0.027987 | benchmark | null | mmlu_high_school_chemistry_rc | 5 | rank_choice | true | eng | 203 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_chemistry",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Clo... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_chemistry_rc | acc_norm | none | 0.221675 | 0.029226 | benchmark | null | mmlu_high_school_chemistry_rc | 5 | rank_choice | true | eng | 203 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_chemistry",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Clo... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_computer_science_bpb | bits_per_byte | none | 1.65856 | 0.134434 | benchmark | null | mmlu_high_school_computer_science_bpb | 5 | bits_per_byte | false | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_computer_science",
"dataset_path": "cais/mmlu",
"doc_to_choice": null,
"doc_to_target": " {{choices[answer]}}",
"doc_to_text": "Question: {{question.strip()}}\nAnswer:",
"generation_kwargs": null,
"output_type": "loglikelihood",
"repeats": 1,
"should_decontaminate": fals... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
HPLT2c_deu | null | mmlu_high_school_computer_science_rc | acc | none | 0.22 | 0.041633 | benchmark | null | mmlu_high_school_computer_science_rc | 5 | rank_choice | true | eng | 100 | null | null | 100,000,000,000 | 2026-02-17T08:57:35.964148+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json | {
"chat_template": null,
"chat_template_sha": null,
"eot_token_id": 1,
"git_hash": "7463f94",
"lm_eval_version": "0.4.12.dev0",
"max_length": 2047,
"model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main",
"model_source": "local-completions",
"system_instruction": null,
"... | {
"dataset_name": "high_school_computer_science",
"dataset_path": "cais/mmlu",
"doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n",
"doc_to_target": "{{answer}}",
"doc_to_text": "def doc_to_text_rc(doc):\n \"... | /leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main | false | HPLT2c deu | null |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.