model
large_stringclasses
19 values
step
float64
1k
950k
task
large_stringclasses
305 values
metric
large_stringclasses
3 values
metric_filter
large_stringclasses
1 value
score
float64
0
5.6
score_stderr
float64
0
0.25
task_type
large_stringclasses
3 values
parent_task
large_stringclasses
12 values
task_display_name
large_stringclasses
305 values
num_fewshot
float64
0
5
task_formulation
large_stringclasses
2 values
higher_is_better
bool
2 classes
language
large_stringclasses
2 values
n_samples
float64
100
87k
subtask_tree
large_stringclasses
20 values
batch_size
float64
2.1M
2.1M
tokens_trained
int64
2.1B
36,000B
eval_date
large_stringdate
2026-02-16 08:28:42
2026-02-18 13:45:55
result_source
large_stringlengths
120
166
eval_config
dict
task_config
dict
model_path
large_stringclasses
338 values
is_checkpoint
bool
2 classes
model_display_name
large_stringclasses
19 values
train_batch_size
int64
2.1M
4.19M
HPLT2c_deu
null
deu_qa_rc
acc_norm
none
0.561757
null
task_group
deu_base_easy
deu_qa_rc
null
null
true
deu
30,715
{"deu_qa_rc": ["deu_arc_challenge_rc", "deu_arc_easy_rc", "deu_csqa_rc", "deu_csqa_easy_rc", "deu_siqa_rc", "deu_siqa_easy_rc", "deu_piqa_rc", "deu_piqa_easy_rc", "deu_csprp_rc", "deu_csnli_rc", "deu_csnli_easy_rc", "deu_mmmlu_filtered_rc"], "deu_mmmlu_filtered_rc": ["deu_mmmlu_abstract_algebra_rc", "deu_mmmlu_college_...
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
null
null
null
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
deu_siqa_bpb
bits_per_byte
none
0.689631
0.00463
benchmark
null
deu_siqa_bpb
5
bits_per_byte
false
deu
1,954
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "ellamind/siqa-german-preview", "doc_to_choice": null, "doc_to_target": " {{correct_answer}}", "doc_to_text": "Frage: {{context}} {{question}}\nAntwort:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
deu_siqa_easy_rc
acc
none
0.514841
0.011309
benchmark
null
deu_siqa_easy_rc
5
rank_choice
true
deu
1,954
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "ellamind/siqa-german-preview", "doc_to_choice": "{{choices}}", "doc_to_target": "{{answer_idx}}", "doc_to_text": "Frage: {{context}} {{question}}\nAntwort:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": fals...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
deu_siqa_easy_rc
acc_norm
none
0.639713
0.010863
benchmark
null
deu_siqa_easy_rc
5
rank_choice
true
deu
1,954
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "ellamind/siqa-german-preview", "doc_to_choice": "{{choices}}", "doc_to_target": "{{answer_idx}}", "doc_to_text": "Frage: {{context}} {{question}}\nAntwort:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": fals...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
deu_siqa_rc
acc
none
0.574207
0.011189
benchmark
null
deu_siqa_rc
5
rank_choice
true
deu
1,954
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "ellamind/siqa-german-preview", "doc_to_choice": "{{choices}}", "doc_to_target": "{{answer_idx}}", "doc_to_text": "Frage: {{context}} {{question}}\nAntwort:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": fals...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
deu_siqa_rc
acc_norm
none
0.548618
0.01126
benchmark
null
deu_siqa_rc
5
rank_choice
true
deu
1,954
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "ellamind/siqa-german-preview", "doc_to_choice": "{{choices}}", "doc_to_target": "{{answer_idx}}", "doc_to_text": "Frage: {{context}} {{question}}\nAntwort:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": fals...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
drop_bpb
bits_per_byte
none
1.958088
0.010973
benchmark
null
drop_bpb
5
bits_per_byte
false
eng
9,536
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "EleutherAI/drop", "doc_to_choice": null, "doc_to_target": " {% if answer.spans %}{{answer.spans[0]}}{% elif answer.number %}{{answer.number}}{% else %}{{answer.date.day}} {{answer.date.month}} {{answer.date.year}}{% endif %}", "doc_to_text": "Passage: {{passage}}\nQues...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
drop_rc
acc
none
0.281265
0.004912
benchmark
null
drop_rc
5
rank_choice
true
eng
8,380
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "allenai/drop-gen2mc", "doc_to_choice": "{{choices.text}}", "doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}", "doc_to_text": "Passage: {{passage_original}}\nQuestion: {{question_original}}\nAnswer:", "generation_kwargs": null, "output_type": "mult...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
drop_rc
acc_norm
none
0.263126
0.00481
benchmark
null
drop_rc
5
rank_choice
true
eng
8,380
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "allenai/drop-gen2mc", "doc_to_choice": "{{choices.text}}", "doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}", "doc_to_text": "Passage: {{passage_original}}\nQuestion: {{question_original}}\nAnswer:", "generation_kwargs": null, "output_type": "mult...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
eng_base_easy
acc
none
0.411789
null
eval_suite
null
eng_base_easy
null
null
true
eng
86,236
{"eng_base_easy": ["qa_rc"], "qa_rc": ["arc_easy_rc", "arc_challenge_rc", "mmlu_rc", "csqa_rc", "hellaswag_rc", "winogrande_rc", "socialiqa_rc", "piqa_rc", "coqa_rc", "drop_rc", "jeopardy_rc", "naturalqs_rc", "squad_rc", "sciq_rc", "qasper_yesno_rc", "lambada_rc", "medmcqa_rc", "medqa_rc"], "mmlu_rc": ["mmlu_abstract_a...
null
100,000,000,000
2026-02-17T11:43:01.575599+00:00
null
null
null
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
eng_base_easy
acc_norm
none
0.416695
null
eval_suite
null
eng_base_easy
null
null
true
eng
81,083
{"eng_base_easy": ["qa_rc"], "qa_rc": ["arc_easy_rc", "arc_challenge_rc", "mmlu_rc", "csqa_rc", "hellaswag_rc", "winogrande_rc", "socialiqa_rc", "piqa_rc", "coqa_rc", "drop_rc", "jeopardy_rc", "naturalqs_rc", "squad_rc", "sciq_rc", "qasper_yesno_rc", "medmcqa_rc", "medqa_rc"], "mmlu_rc": ["mmlu_abstract_algebra_rc", "m...
null
100,000,000,000
2026-02-17T11:43:01.575599+00:00
null
null
null
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
eng_base_easy
bits_per_byte
none
1.378868
null
eval_suite
null
eng_base_easy
null
null
false
eng
86,964
{"eng_base_easy": ["math_bpb", "code_bpb", "qa_bpb"], "math_bpb": ["minerva_math_algebra_bpb", "minerva_math_counting_and_probability_bpb", "minerva_math_geometry_bpb", "minerva_math_intermediate_algebra_bpb", "minerva_math_number_theory_bpb", "minerva_math_prealgebra_bpb", "minerva_math_precalculus_bpb"], "code_bpb": ...
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
null
null
null
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
hellaswag_bpb
bits_per_byte
none
1.047292
0.002084
benchmark
null
hellaswag_bpb
5
bits_per_byte
false
eng
10,042
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "allenai/hellaswag", "doc_to_choice": null, "doc_to_target": " {{endings[label | int]}}", "doc_to_text": "{{activity_label}}: {{ctx_a}} {{ctx_b}}", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_ha...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
hellaswag_rc
acc
none
0.2525
0.004345
benchmark
null
hellaswag_rc
5
rank_choice
true
eng
10,000
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "allenai/hellaswag", "doc_to_choice": "{{endings}}", "doc_to_target": "{{label | int}}", "doc_to_text": "{{activity_label}}: {{ctx_a}} {{ctx_b}}", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": false, "task_h...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
hellaswag_rc
acc_norm
none
0.2478
0.004318
benchmark
null
hellaswag_rc
5
rank_choice
true
eng
10,000
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "allenai/hellaswag", "doc_to_choice": "{{endings}}", "doc_to_target": "{{label | int}}", "doc_to_text": "{{activity_label}}: {{ctx_a}} {{ctx_b}}", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": false, "task_h...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
jeopardy_bpb
bits_per_byte
none
1.360302
0.018296
benchmark
null
jeopardy_bpb
5
bits_per_byte
false
eng
2,116
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "mosaicml_gauntlet", "dataset_path": "soldni/jeopardy", "doc_to_choice": null, "doc_to_target": " {{continuation}}", "doc_to_text": "Category: {{context.split(': ', 1)[0]}}\nClue: {{context.split(': ', 1)[1]}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repea...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
jeopardy_rc
acc
none
0.462125
0.010885
benchmark
null
jeopardy_rc
5
rank_choice
true
eng
2,099
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "allenai/jeopardy-gen2mc", "doc_to_choice": "{{choices.text}}", "doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}", "doc_to_text": "Category: {{context_original.split(': ', 1)[0]}}\nQuestion: {{context_original.split(': ', 1)[1]}}\nAnswer:", "generati...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
jeopardy_rc
acc_norm
none
0.396379
0.010679
benchmark
null
jeopardy_rc
5
rank_choice
true
eng
2,099
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "allenai/jeopardy-gen2mc", "doc_to_choice": "{{choices.text}}", "doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}", "doc_to_text": "Category: {{context_original.split(': ', 1)[0]}}\nQuestion: {{context_original.split(': ', 1)[1]}}\nAnswer:", "generati...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
lab_bench_dbqa_bpb
bits_per_byte
none
4.576707
0.033691
benchmark
null
lab_bench_dbqa_bpb
3
bits_per_byte
false
eng
520
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "DbQA", "dataset_path": "futurehouse/lab-bench", "doc_to_choice": null, "doc_to_target": " {{ideal}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash": null, "tas...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
lab_bench_protocolqa_bpb
bits_per_byte
none
1.630068
0.038825
benchmark
null
lab_bench_protocolqa_bpb
3
bits_per_byte
false
eng
108
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "ProtocolQA", "dataset_path": "futurehouse/lab-bench", "doc_to_choice": null, "doc_to_target": " {{ideal}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash": null, ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
lambada_bpb
bits_per_byte
none
1.437608
null
benchmark
null
lambada_bpb
0
bits_per_byte
false
eng
5,153
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "default", "dataset_path": "EleutherAI/lambada_openai", "doc_to_choice": null, "doc_to_target": "{{text}}", "doc_to_text": "", "generation_kwargs": null, "output_type": "loglikelihood_rolling", "repeats": 1, "should_decontaminate": false, "task_hash": null, "task_version": "1.0" ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
lambada_rc
acc
none
0.363672
0.006702
benchmark
null
lambada_rc
0
rank_choice
true
eng
5,153
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "default", "dataset_path": "EleutherAI/lambada_openai", "doc_to_choice": null, "doc_to_target": " {{text.rsplit(' ', 1)[1]}}", "doc_to_text": "{{text.rsplit(' ', 1)[0]}}", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
math_bpb
bits_per_byte
none
1.327082
null
task_group
eng_base_easy
math_bpb
null
null
false
eng
5,000
{"math_bpb": ["minerva_math_algebra_bpb", "minerva_math_counting_and_probability_bpb", "minerva_math_geometry_bpb", "minerva_math_intermediate_algebra_bpb", "minerva_math_number_theory_bpb", "minerva_math_prealgebra_bpb", "minerva_math_precalculus_bpb"]}
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
null
null
null
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mbpp_bpb
bits_per_byte
none
1.982307
0.033944
benchmark
null
mbpp_bpb
3
bits_per_byte
false
eng
500
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "full", "dataset_path": "google-research-datasets/mbpp", "doc_to_choice": null, "doc_to_target": "def doc_to_target_bpb(doc):\n \"\"\"Code body after function signature (no overlap with context).\n\n Example:\n \\\\n for i in range(len(s)):\\\\n ...\n \"\"\"\n code ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
medmcqa_bpb
bits_per_byte
none
1.823879
0.014063
benchmark
null
medmcqa_bpb
5
bits_per_byte
false
eng
4,183
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "openlifescienceai/medmcqa", "doc_to_choice": null, "doc_to_target": " {{[opa, opb, opc, opd][cop]}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "ta...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
medmcqa_rc
acc
none
0.220894
0.006415
benchmark
null
medmcqa_rc
5
rank_choice
true
eng
4,183
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "openlifescienceai/medmcqa", "doc_to_choice": "{{[opa, opb, opc, opd]}}", "doc_to_target": "{{cop}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": false, "t...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
medmcqa_rc
acc_norm
none
0.263447
0.006812
benchmark
null
medmcqa_rc
5
rank_choice
true
eng
4,183
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "openlifescienceai/medmcqa", "doc_to_choice": "{{[opa, opb, opc, opd]}}", "doc_to_target": "{{cop}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": false, "t...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
medqa_bpb
bits_per_byte
none
1.359605
0.018932
benchmark
null
medqa_bpb
5
bits_per_byte
false
eng
1,273
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "davidheineman/medqa-en", "doc_to_choice": null, "doc_to_target": " {{answer}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash": null, "tas...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
medqa_rc
acc
none
0.210526
0.011431
benchmark
null
medqa_rc
5
rank_choice
true
eng
1,273
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "davidheineman/medqa-en", "doc_to_choice": "{{choices}}", "doc_to_target": "{{answer_idx}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": false, "task_hash"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
medqa_rc
acc_norm
none
0.209741
0.011415
benchmark
null
medqa_rc
5
rank_choice
true
eng
1,273
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": null, "dataset_path": "davidheineman/medqa-en", "doc_to_choice": "{{choices}}", "doc_to_target": "{{answer_idx}}", "doc_to_text": "Question: {{question}}\nAnswer:", "generation_kwargs": null, "output_type": "multiple_choice", "repeats": 1, "should_decontaminate": false, "task_hash"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
minerva_math_algebra_bpb
bits_per_byte
none
1.434884
0.008872
benchmark
null
minerva_math_algebra_bpb
4
bits_per_byte
false
eng
1,187
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "algebra", "dataset_path": "EleutherAI/hendrycks_math", "doc_to_choice": null, "doc_to_target": " {{solution}}", "doc_to_text": "Problem:\n{{problem}}\n\nSolution:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash":...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
minerva_math_counting_and_probability_bpb
bits_per_byte
none
1.170161
0.014208
benchmark
null
minerva_math_counting_and_probability_bpb
4
bits_per_byte
false
eng
474
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "counting_and_probability", "dataset_path": "EleutherAI/hendrycks_math", "doc_to_choice": null, "doc_to_target": " {{solution}}", "doc_to_text": "Problem:\n{{problem}}\n\nSolution:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": fals...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
minerva_math_geometry_bpb
bits_per_byte
none
1.36389
0.015558
benchmark
null
minerva_math_geometry_bpb
4
bits_per_byte
false
eng
479
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "geometry", "dataset_path": "EleutherAI/hendrycks_math", "doc_to_choice": null, "doc_to_target": " {{solution}}", "doc_to_text": "Problem:\n{{problem}}\n\nSolution:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
minerva_math_intermediate_algebra_bpb
bits_per_byte
none
1.469797
0.00995
benchmark
null
minerva_math_intermediate_algebra_bpb
4
bits_per_byte
false
eng
903
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "intermediate_algebra", "dataset_path": "EleutherAI/hendrycks_math", "doc_to_choice": null, "doc_to_target": " {{solution}}", "doc_to_text": "Problem:\n{{problem}}\n\nSolution:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
minerva_math_number_theory_bpb
bits_per_byte
none
1.336563
0.011864
benchmark
null
minerva_math_number_theory_bpb
4
bits_per_byte
false
eng
540
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "number_theory", "dataset_path": "EleutherAI/hendrycks_math", "doc_to_choice": null, "doc_to_target": " {{solution}}", "doc_to_text": "Problem:\n{{problem}}\n\nSolution:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
minerva_math_prealgebra_bpb
bits_per_byte
none
1.23569
0.010422
benchmark
null
minerva_math_prealgebra_bpb
4
bits_per_byte
false
eng
871
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "prealgebra", "dataset_path": "EleutherAI/hendrycks_math", "doc_to_choice": null, "doc_to_target": " {{solution}}", "doc_to_text": "Problem:\n{{problem}}\n\nSolution:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_has...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
minerva_math_precalculus_bpb
bits_per_byte
none
1.278591
0.015545
benchmark
null
minerva_math_precalculus_bpb
4
bits_per_byte
false
eng
546
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "precalculus", "dataset_path": "EleutherAI/hendrycks_math", "doc_to_choice": null, "doc_to_target": " {{solution}}", "doc_to_text": "Problem:\n{{problem}}\n\nSolution:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_ha...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_abstract_algebra_bpb
bits_per_byte
none
1.480533
0.148554
benchmark
null
mmlu_abstract_algebra_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "abstract_algebra", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_h...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_abstract_algebra_rc
acc
none
0.16
0.036845
benchmark
null
mmlu_abstract_algebra_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "abstract_algebra", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_abstract_algebra_rc
acc_norm
none
0.17
0.037753
benchmark
null
mmlu_abstract_algebra_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "abstract_algebra", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_anatomy_bpb
bits_per_byte
none
1.292732
0.039576
benchmark
null
mmlu_anatomy_bpb
5
bits_per_byte
false
eng
135
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "anatomy", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash": nul...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_anatomy_rc
acc
none
0.296296
0.039446
benchmark
null
mmlu_anatomy_rc
5
rank_choice
true
eng
135
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "anatomy", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: 'Qu...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_anatomy_rc
acc_norm
none
0.318519
0.040248
benchmark
null
mmlu_anatomy_rc
5
rank_choice
true
eng
135
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "anatomy", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: 'Qu...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_astronomy_bpb
bits_per_byte
none
1.361239
0.062982
benchmark
null
mmlu_astronomy_bpb
5
bits_per_byte
false
eng
152
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "astronomy", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash": n...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_astronomy_rc
acc
none
0.210526
0.033177
benchmark
null
mmlu_astronomy_rc
5
rank_choice
true
eng
152
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "astronomy", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: '...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_astronomy_rc
acc_norm
none
0.256579
0.035542
benchmark
null
mmlu_astronomy_rc
5
rank_choice
true
eng
152
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "astronomy", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt: '...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_bpb
bits_per_byte
none
1.433623
null
task_group
qa_bpb
mmlu_bpb
null
null
false
eng
14,042
{"mmlu_bpb": ["mmlu_abstract_algebra_bpb", "mmlu_anatomy_bpb", "mmlu_astronomy_bpb", "mmlu_business_ethics_bpb", "mmlu_clinical_knowledge_bpb", "mmlu_college_biology_bpb", "mmlu_college_chemistry_bpb", "mmlu_college_computer_science_bpb", "mmlu_college_mathematics_bpb", "mmlu_college_medicine_bpb", "mmlu_college_physic...
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
null
null
null
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_business_ethics_bpb
bits_per_byte
none
1.532048
0.126497
benchmark
null
mmlu_business_ethics_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "business_ethics", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_ha...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_business_ethics_rc
acc
none
0.42
0.049604
benchmark
null
mmlu_business_ethics_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "business_ethics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_business_ethics_rc
acc_norm
none
0.33
0.047258
benchmark
null
mmlu_business_ethics_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "business_ethics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_clinical_knowledge_bpb
bits_per_byte
none
1.336823
0.035128
benchmark
null
mmlu_clinical_knowledge_bpb
5
bits_per_byte
false
eng
265
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "clinical_knowledge", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_clinical_knowledge_rc
acc
none
0.218868
0.025448
benchmark
null
mmlu_clinical_knowledge_rc
5
rank_choice
true
eng
265
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "clinical_knowledge", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_clinical_knowledge_rc
acc_norm
none
0.267925
0.027257
benchmark
null
mmlu_clinical_knowledge_rc
5
rank_choice
true
eng
265
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "clinical_knowledge", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_biology_bpb
bits_per_byte
none
1.310637
0.043286
benchmark
null
mmlu_college_biology_bpb
5
bits_per_byte
false
eng
144
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_biology", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_ha...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_biology_rc
acc
none
0.236111
0.035514
benchmark
null
mmlu_college_biology_rc
5
rank_choice
true
eng
144
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_biology", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_biology_rc
acc_norm
none
0.25
0.03621
benchmark
null
mmlu_college_biology_rc
5
rank_choice
true
eng
144
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_biology", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_chemistry_bpb
bits_per_byte
none
2.226741
0.110052
benchmark
null
mmlu_college_chemistry_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_chemistry", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_chemistry_rc
acc
none
0.33
0.047258
benchmark
null
mmlu_college_chemistry_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_chemistry", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_chemistry_rc
acc_norm
none
0.31
0.046482
benchmark
null
mmlu_college_chemistry_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_chemistry", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_computer_science_bpb
bits_per_byte
none
1.980096
0.109393
benchmark
null
mmlu_college_computer_science_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_computer_science", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_computer_science_rc
acc
none
0.32
0.046883
benchmark
null
mmlu_college_computer_science_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_computer_science", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_computer_science_rc
acc_norm
none
0.24
0.042923
benchmark
null
mmlu_college_computer_science_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_computer_science", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_mathematics_bpb
bits_per_byte
none
2.365635
0.106684
benchmark
null
mmlu_college_mathematics_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_mathematics", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "tas...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_mathematics_rc
acc
none
0.14
0.034874
benchmark
null
mmlu_college_mathematics_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_mathematics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_mathematics_rc
acc_norm
none
0.22
0.041633
benchmark
null
mmlu_college_mathematics_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_mathematics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_medicine_bpb
bits_per_byte
none
1.361238
0.047204
benchmark
null
mmlu_college_medicine_bpb
5
bits_per_byte
false
eng
173
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_medicine", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_h...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_medicine_rc
acc
none
0.236994
0.032424
benchmark
null
mmlu_college_medicine_rc
5
rank_choice
true
eng
173
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_medicine", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_medicine_rc
acc_norm
none
0.225434
0.031862
benchmark
null
mmlu_college_medicine_rc
5
rank_choice
true
eng
173
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_medicine", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pr...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_physics_bpb
bits_per_byte
none
2.34413
0.120026
benchmark
null
mmlu_college_physics_bpb
5
bits_per_byte
false
eng
102
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_physics", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_ha...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_physics_rc
acc
none
0.147059
0.035241
benchmark
null
mmlu_college_physics_rc
5
rank_choice
true
eng
102
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_physics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_college_physics_rc
acc_norm
none
0.156863
0.036187
benchmark
null
mmlu_college_physics_rc
5
rank_choice
true
eng
102
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "college_physics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze pro...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_computer_security_bpb
bits_per_byte
none
1.530627
0.072114
benchmark
null
mmlu_computer_security_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "computer_security", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_computer_security_rc
acc
none
0.24
0.042923
benchmark
null
mmlu_computer_security_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "computer_security", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_computer_security_rc
acc_norm
none
0.3
0.046057
benchmark
null
mmlu_computer_security_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "computer_security", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze p...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_conceptual_physics_bpb
bits_per_byte
none
1.415904
0.04384
benchmark
null
mmlu_conceptual_physics_bpb
5
bits_per_byte
false
eng
235
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "conceptual_physics", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_conceptual_physics_rc
acc
none
0.323404
0.030579
benchmark
null
mmlu_conceptual_physics_rc
5
rank_choice
true
eng
235
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "conceptual_physics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_conceptual_physics_rc
acc_norm
none
0.297872
0.029896
benchmark
null
mmlu_conceptual_physics_rc
5
rank_choice
true
eng
235
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "conceptual_physics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze ...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_econometrics_bpb
bits_per_byte
none
1.143167
0.073116
benchmark
null
mmlu_econometrics_bpb
5
bits_per_byte
false
eng
114
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "econometrics", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_econometrics_rc
acc
none
0.254386
0.04097
benchmark
null
mmlu_econometrics_rc
5
rank_choice
true
eng
114
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "econometrics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_econometrics_rc
acc_norm
none
0.245614
0.040493
benchmark
null
mmlu_econometrics_rc
5
rank_choice
true
eng
114
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "econometrics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_electrical_engineering_bpb
bits_per_byte
none
1.917137
0.063081
benchmark
null
mmlu_electrical_engineering_bpb
5
bits_per_byte
false
eng
145
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "electrical_engineering", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_electrical_engineering_rc
acc
none
0.262069
0.036647
benchmark
null
mmlu_electrical_engineering_rc
5
rank_choice
true
eng
145
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "electrical_engineering", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_electrical_engineering_rc
acc_norm
none
0.268966
0.036952
benchmark
null
mmlu_electrical_engineering_rc
5
rank_choice
true
eng
145
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "electrical_engineering", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_elementary_mathematics_bpb
bits_per_byte
none
2.559193
0.053839
benchmark
null
mmlu_elementary_mathematics_bpb
5
bits_per_byte
false
eng
378
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "elementary_mathematics", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_elementary_mathematics_rc
acc
none
0.214286
0.021133
benchmark
null
mmlu_elementary_mathematics_rc
5
rank_choice
true
eng
378
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "elementary_mathematics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_elementary_mathematics_rc
acc_norm
none
0.23545
0.021852
benchmark
null
mmlu_elementary_mathematics_rc
5
rank_choice
true
eng
378
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "elementary_mathematics", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cl...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_formal_logic_bpb
bits_per_byte
none
1.988142
0.116928
benchmark
null
mmlu_formal_logic_bpb
5
bits_per_byte
false
eng
126
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "formal_logic", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_formal_logic_rc
acc
none
0.31746
0.041635
benchmark
null
mmlu_formal_logic_rc
5
rank_choice
true
eng
126
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "formal_logic", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_formal_logic_rc
acc_norm
none
0.261905
0.039325
benchmark
null
mmlu_formal_logic_rc
5
rank_choice
true
eng
126
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "formal_logic", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_global_facts_bpb
bits_per_byte
none
1.821181
0.064829
benchmark
null
mmlu_global_facts_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "global_facts", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "task_hash"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_global_facts_rc
acc
none
0.22
0.041633
benchmark
null
mmlu_global_facts_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "global_facts", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_global_facts_rc
acc_norm
none
0.21
0.040936
benchmark
null
mmlu_global_facts_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "global_facts", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze prompt...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_biology_bpb
bits_per_byte
none
1.258708
0.039439
benchmark
null
mmlu_high_school_biology_bpb
5
bits_per_byte
false
eng
310
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_biology", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "tas...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_biology_rc
acc
none
0.245161
0.024472
benchmark
null
mmlu_high_school_biology_rc
5
rank_choice
true
eng
310
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_biology", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_biology_rc
acc_norm
none
0.312903
0.026378
benchmark
null
mmlu_high_school_biology_rc
5
rank_choice
true
eng
310
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_biology", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Cloze...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_chemistry_bpb
bits_per_byte
none
1.797248
0.066453
benchmark
null
mmlu_high_school_chemistry_bpb
5
bits_per_byte
false
eng
203
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_chemistry", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": false, "t...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_chemistry_rc
acc
none
0.197044
0.027987
benchmark
null
mmlu_high_school_chemistry_rc
5
rank_choice
true
eng
203
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_chemistry", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Clo...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_chemistry_rc
acc_norm
none
0.221675
0.029226
benchmark
null
mmlu_high_school_chemistry_rc
5
rank_choice
true
eng
203
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_chemistry", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"\"\"Clo...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_computer_science_bpb
bits_per_byte
none
1.65856
0.134434
benchmark
null
mmlu_high_school_computer_science_bpb
5
bits_per_byte
false
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_computer_science", "dataset_path": "cais/mmlu", "doc_to_choice": null, "doc_to_target": " {{choices[answer]}}", "doc_to_text": "Question: {{question.strip()}}\nAnswer:", "generation_kwargs": null, "output_type": "loglikelihood", "repeats": 1, "should_decontaminate": fals...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null
HPLT2c_deu
null
mmlu_high_school_computer_science_rc
acc
none
0.22
0.041633
benchmark
null
mmlu_high_school_computer_science_rc
5
rank_choice
true
eng
100
null
null
100,000,000,000
2026-02-17T08:57:35.964148+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive/eval_results/ref_models/HPLT2c_deu/main/compacted/results_compacted.json
{ "chat_template": null, "chat_template_sha": null, "eot_token_id": 1, "git_hash": "7463f94", "lm_eval_version": "0.4.12.dev0", "max_length": 2047, "model_name": "/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main", "model_source": "local-completions", "system_instruction": null, "...
{ "dataset_name": "high_school_computer_science", "dataset_path": "cais/mmlu", "doc_to_choice": "def doc_to_choice_rc(doc):\n \"\"\"Return full answer texts as choices (not letter labels).\"\"\"\n return doc[\"choices\"]\n", "doc_to_target": "{{answer}}", "doc_to_text": "def doc_to_text_rc(doc):\n \"...
/leonardo_work/AIFAC_L01_028/models/HPLT/hplt2c_deu_checkpoints/main
false
HPLT2c deu
null