Datasets:
Formats:
parquet
Size:
10K - 100K
model large_stringclasses 3 values | step int64 10k 2.63M ⌀ | task large_stringclasses 767 values | metric large_stringclasses 20 values | metric_filter large_stringclasses 6 values | score float64 0 4.45 | score_stderr float64 0 0.25 ⌀ | task_type large_stringclasses 3 values | parent_task large_stringclasses 7 values | task_display_name large_stringclasses 767 values | num_fewshot int64 0 8 ⌀ | task_formulation large_stringclasses 3 values | higher_is_better bool 2 classes | language large_stringclasses 2 values | n_samples int64 23 180k | subtask_tree large_stringclasses 44 values | train_batch_size int64 4.19M 8.39M ⌀ | tokens_trained int64 41.9B 22,038B | eval_date large_stringdate 2026-03-10 17:22:50 2026-03-12 12:50:40 | result_source large_stringlengths 160 168 ⌀ | eval_config dict | task_config dict | model_path large_stringclasses 95 values | is_checkpoint bool 2 classes | model_display_name large_stringclasses 3 values |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
apertus_8b | 50,000 | arc_challenge_bpb | bits_per_byte | none | 0.777649 | 0.013966 | benchmark | null | arc_challenge_bpb | 5 | bits_per_byte | false | eng | 1,172 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693991+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2467.716009796597,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "loglikelihood",
"dataset_name": "ARC-Challenge",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{choices.text[choices.label.index(answerKey)]}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_challenge_mc | acc | none | 0.263468 | 0.009039 | benchmark | null | arc_challenge_mc | 5 | rank_choice | true | eng | 2,376 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694601+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3745.1029282175004,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Easy",
"repeats": 1,
"doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:",
"doc_to_target": "{{answerKey}}",
"doc_to_choice": "{{choices.label}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_challenge_mc | acc_norm | none | 0.263468 | 0.009039 | benchmark | null | arc_challenge_mc | 5 | rank_choice | true | eng | 2,376 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694601+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3745.1029282175004,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Easy",
"repeats": 1,
"doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:",
"doc_to_target": "{{answerKey}}",
"doc_to_choice": "{{choices.label}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_challenge_rc | acc | none | 0.41041 | 0.014375 | benchmark | null | arc_challenge_rc | 5 | rank_choice | true | eng | 1,172 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693991+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2467.716009796597,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Challenge",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_challenge_rc | acc_norm | none | 0.449659 | 0.014537 | benchmark | null | arc_challenge_rc | 5 | rank_choice | true | eng | 1,172 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693991+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2467.716009796597,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Challenge",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_easy_bpb | bits_per_byte | none | 0.580533 | 0.007264 | benchmark | null | arc_easy_bpb | 5 | bits_per_byte | false | eng | 2,376 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694601+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3745.1029282175004,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "loglikelihood",
"dataset_name": "ARC-Easy",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{choices.text[choices.label.index(answerKey)]}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_easy_mc | acc | none | 0.244529 | 0.008819 | benchmark | null | arc_easy_mc | 5 | rank_choice | true | eng | 2,376 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694601+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3745.1029282175004,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Easy",
"repeats": 1,
"doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:",
"doc_to_target": "{{answerKey}}",
"doc_to_choice": "{{choices.label}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_easy_mc | acc_norm | none | 0.244529 | 0.008819 | benchmark | null | arc_easy_mc | 5 | rank_choice | true | eng | 2,376 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694601+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3745.1029282175004,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Easy",
"repeats": 1,
"doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:",
"doc_to_target": "{{answerKey}}",
"doc_to_choice": "{{choices.label}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_easy_rc | acc | none | 0.739478 | 0.009006 | benchmark | null | arc_easy_rc | 5 | rank_choice | true | eng | 2,376 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694601+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3745.1029282175004,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Easy",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | arc_easy_rc | acc_norm | none | 0.737374 | 0.00903 | benchmark | null | arc_easy_rc | 5 | rank_choice | true | eng | 2,376 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694601+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3745.1029282175004,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/ai2_arc",
"output_type": "multiple_choice",
"dataset_name": "ARC-Easy",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_arithmetic_bpb | bits_per_byte | none | 1.139799 | 0.029688 | benchmark | null | basic_skills_arithmetic_bpb | 5 | bits_per_byte | false | eng | 1,047 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596432+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1829.1680674208328,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "loglikelihood",
"dataset_name": "arithmetic",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{answer}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_arithmetic_rc | acc | none | 0.56638 | 0.015323 | benchmark | null | basic_skills_arithmetic_rc | 5 | rank_choice | true | eng | 1,047 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596432+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1829.1680674208328,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "arithmetic",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_arithmetic_rc | acc_norm | none | 0.587393 | 0.015222 | benchmark | null | basic_skills_arithmetic_rc | 5 | rank_choice | true | eng | 1,047 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596432+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1829.1680674208328,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "arithmetic",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_coding_bpb | bits_per_byte | none | 0.560042 | 0.009872 | benchmark | null | basic_skills_coding_bpb | 5 | bits_per_byte | false | eng | 1,012 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596733+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 929.7853837599978,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "loglikelihood",
"dataset_name": "coding",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{answer}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_coding_rc | acc | none | 0.523715 | 0.015707 | benchmark | null | basic_skills_coding_rc | 5 | rank_choice | true | eng | 1,012 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596432+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1829.1680674208328,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "coding",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_coding_rc | acc_norm | none | 0.672925 | 0.014755 | benchmark | null | basic_skills_coding_rc | 5 | rank_choice | true | eng | 1,012 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596432+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1829.1680674208328,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "coding",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_common_knowledge_bpb | bits_per_byte | none | 0.172218 | 0.00593 | benchmark | null | basic_skills_common_knowledge_bpb | 5 | bits_per_byte | false | eng | 1,037 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596733+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 929.7853837599978,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "loglikelihood",
"dataset_name": "common_knowledge",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{answer}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_common_knowledge_rc | acc | none | 0.937319 | 0.007531 | benchmark | null | basic_skills_common_knowledge_rc | 5 | rank_choice | true | eng | 1,037 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596733+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 929.7853837599978,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "common_knowledge",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_common_knowledge_rc | acc_norm | none | 0.922854 | 0.00829 | benchmark | null | basic_skills_common_knowledge_rc | 5 | rank_choice | true | eng | 1,037 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596733+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 929.7853837599978,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "common_knowledge",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_logical_reasoning_bpb | bits_per_byte | none | 0.264452 | 0.005103 | benchmark | null | basic_skills_logical_reasoning_bpb | 5 | bits_per_byte | false | eng | 1,118 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990618+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 804.4468631707132,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "loglikelihood",
"dataset_name": "logical_reasoning",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{answer}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_logical_reasoning_rc | acc | none | 0.966011 | 0.005422 | benchmark | null | basic_skills_logical_reasoning_rc | 5 | rank_choice | true | eng | 1,118 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596733+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 929.7853837599978,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "logical_reasoning",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_logical_reasoning_rc | acc_norm | none | 0.974955 | 0.004675 | benchmark | null | basic_skills_logical_reasoning_rc | 5 | rank_choice | true | eng | 1,118 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596733+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 929.7853837599978,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "logical_reasoning",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_pattern_bpb | bits_per_byte | none | 1.300733 | 0.039628 | benchmark | null | basic_skills_pattern_bpb | 5 | bits_per_byte | false | eng | 534 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990618+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 804.4468631707132,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "loglikelihood",
"dataset_name": "pattern",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{answer}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_pattern_rc | acc | none | 0.773408 | 0.018133 | benchmark | null | basic_skills_pattern_rc | 5 | rank_choice | true | eng | 534 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990618+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 804.4468631707132,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "pattern",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_pattern_rc | acc_norm | none | 0.775281 | 0.018079 | benchmark | null | basic_skills_pattern_rc | 5 | rank_choice | true | eng | 534 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990618+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 804.4468631707132,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "pattern",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_rc | acc | none | 0.559694 | 0.015349 | benchmark | null | basic_skills_rc | 5 | rank_choice | true | eng | 1,047 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990618+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 804.4468631707132,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "arithmetic",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_rc | acc_norm | none | 0.582617 | 0.015247 | benchmark | null | basic_skills_rc | 5 | rank_choice | true | eng | 1,047 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990618+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 804.4468631707132,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "arithmetic",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_string_operations_bpb | bits_per_byte | none | 1.185186 | 0.032276 | benchmark | null | basic_skills_string_operations_bpb | 5 | bits_per_byte | false | eng | 1,219 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990878+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 4027.134608012624,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "loglikelihood",
"dataset_name": "string_operations",
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{answer}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_string_operations_rc | acc | none | 0.303527 | 0.013174 | benchmark | null | basic_skills_string_operations_rc | 5 | rank_choice | true | eng | 1,219 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990878+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 4027.134608012624,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "string_operations",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | basic_skills_string_operations_rc | acc_norm | none | 0.401148 | 0.014044 | benchmark | null | basic_skills_string_operations_rc | 5 | rank_choice | true | eng | 1,219 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990878+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 4027.134608012624,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/basic-skills",
"output_type": "multiple_choice",
"dataset_name": "string_operations",
"repeats": 1,
"doc_to_text": "Question: {{question}}",
"doc_to_target": "0",
"doc_to_choice": "{{[answer] + wrong_answers[:3]}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_boolean_expressions_cot | exact_match | get-answer | 0.528 | 0.031636 | benchmark | null | bbh_boolean_expressions_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693991+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2467.716009796597,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "boolean_expressions",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_causal_judgement_cot | exact_match | get-answer | 0.475936 | 0.036619 | benchmark | null | bbh_causal_judgement_cot | 3 | generative | true | eng | 187 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693991+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2467.716009796597,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "causal_judgement",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_date_understanding_cot | exact_match | get-answer | 0.212 | 0.025902 | benchmark | null | bbh_date_understanding_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692865+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1844.1305395988747,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "date_understanding",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_disambiguation_qa_cot | exact_match | get-answer | 0.408 | 0.031145 | benchmark | null | bbh_disambiguation_qa_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692865+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1844.1305395988747,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "disambiguation_qa",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_dyck_languages_cot | exact_match | get-answer | 0.008 | 0.005645 | benchmark | null | bbh_dyck_languages_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692865+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1844.1305395988747,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "dyck_languages",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_formal_fallacies_cot | exact_match | get-answer | 0.508 | 0.031682 | benchmark | null | bbh_formal_fallacies_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692865+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1844.1305395988747,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "formal_fallacies",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_geometric_shapes_cot | exact_match | get-answer | 0.14 | 0.021989 | benchmark | null | bbh_geometric_shapes_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692869+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 444.9072736380622,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "geometric_shapes",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_hyperbaton_cot | exact_match | get-answer | 0.528 | 0.031636 | benchmark | null | bbh_hyperbaton_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692869+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 444.9072736380622,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "hyperbaton",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_logical_deduction_five_objects_cot | exact_match | get-answer | 0.256 | 0.027657 | benchmark | null | bbh_logical_deduction_five_objects_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692869+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 444.9072736380622,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "logical_deduction_five_objects",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_logical_deduction_seven_objects_cot | exact_match | get-answer | 0.164 | 0.023465 | benchmark | null | bbh_logical_deduction_seven_objects_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692869+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 444.9072736380622,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "logical_deduction_seven_objects",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_logical_deduction_three_objects_cot | exact_match | get-answer | 0.324 | 0.029658 | benchmark | null | bbh_logical_deduction_three_objects_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694015+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 406.35265475604683,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "logical_deduction_three_objects",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_movie_recommendation_cot | exact_match | get-answer | 0.308 | 0.029257 | benchmark | null | bbh_movie_recommendation_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694015+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 406.35265475604683,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "movie_recommendation",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_multistep_arithmetic_two_cot | exact_match | get-answer | 0.008 | 0.005645 | benchmark | null | bbh_multistep_arithmetic_two_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694015+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 406.35265475604683,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "multistep_arithmetic_two",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_navigate_cot | exact_match | get-answer | 0.6 | 0.031046 | benchmark | null | bbh_navigate_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694015+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 406.35265475604683,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "navigate",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_object_counting_cot | exact_match | get-answer | 0.276 | 0.028329 | benchmark | null | bbh_object_counting_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694321+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.3605205435306,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "object_counting",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_penguins_in_a_table_cot | exact_match | get-answer | 0.157534 | 0.030254 | benchmark | null | bbh_penguins_in_a_table_cot | 3 | generative | true | eng | 146 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694321+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.3605205435306,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "penguins_in_a_table",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_reasoning_about_colored_objects_cot | exact_match | get-answer | 0.16 | 0.023233 | benchmark | null | bbh_reasoning_about_colored_objects_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694321+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.3605205435306,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "reasoning_about_colored_objects",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_ruin_names_cot | exact_match | get-answer | 0.224 | 0.026421 | benchmark | null | bbh_ruin_names_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.694321+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.3605205435306,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "ruin_names",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_salient_translation_error_detection_cot | exact_match | get-answer | 0.2 | 0.025349 | benchmark | null | bbh_salient_translation_error_detection_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692898+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.1463885176927,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "salient_translation_error_detection",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_snarks_cot | exact_match | get-answer | 0.550562 | 0.03739 | benchmark | null | bbh_snarks_cot | 3 | generative | true | eng | 178 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692898+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.1463885176927,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "snarks",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_sports_understanding_cot | exact_match | get-answer | 0.504 | 0.031685 | benchmark | null | bbh_sports_understanding_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692898+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.1463885176927,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "sports_understanding",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_temporal_sequences_cot | exact_match | get-answer | 0.252 | 0.027514 | benchmark | null | bbh_temporal_sequences_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.692898+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 394.1463885176927,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "temporal_sequences",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_tracking_shuffled_objects_five_objects_cot | exact_match | get-answer | 0.2 | 0.025349 | benchmark | null | bbh_tracking_shuffled_objects_five_objects_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693084+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 403.64010308589786,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "tracking_shuffled_objects_five_objects",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_tracking_shuffled_objects_seven_objects_cot | exact_match | get-answer | 0.096 | 0.018669 | benchmark | null | bbh_tracking_shuffled_objects_seven_objects_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693084+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 403.64010308589786,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "tracking_shuffled_objects_seven_objects",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_tracking_shuffled_objects_three_objects_cot | exact_match | get-answer | 0.332 | 0.029844 | benchmark | null | bbh_tracking_shuffled_objects_three_objects_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693084+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 403.64010308589786,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "tracking_shuffled_objects_three_objects",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_web_of_lies_cot | exact_match | get-answer | 0.532 | 0.031621 | benchmark | null | bbh_web_of_lies_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:30:48.693084+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 403.64010308589786,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "web_of_lies",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | bbh_word_sorting_cot | exact_match | get-answer | 0.028 | 0.010455 | benchmark | null | bbh_word_sorting_cot | 3 | generative | true | eng | 250 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:35.596432+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1829.1680674208328,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "lukaemon/bbh",
"output_type": "generate_until",
"dataset_name": "word_sorting",
"repeats": 1,
"doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n",
"doc_to_target": "{{target}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | coqa_bpb | bits_per_byte | none | 0.695591 | 0.030972 | benchmark | null | coqa_bpb | 0 | bits_per_byte | false | eng | 500 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:37:36.061508+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3691.1114229094237,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "EleutherAI/coqa",
"output_type": "loglikelihood",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "Story: {{story}}\n\nQ: {{questions.input_text[0]}}\nA:",
"doc_to_target": " {{answers.input_text[0]}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | coqa_gen | em | none | 0.593762 | 0.005497 | benchmark | null | coqa_gen | 0 | generative | true | eng | 7,983 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990878+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 4027.134608012624,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "EleutherAI/coqa",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "{{query}}",
"doc_to_target": " {{answers[0]}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 50, \"until\": [\"\\n\\n\", \"Answer:\", \"Question:\"], \"do_sample\": false, \"temperature\": 0.0}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | coqa_gen | f1 | none | 0.7099 | 0.004537 | benchmark | null | coqa_gen | 0 | generative | true | eng | 7,983 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990878+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 4027.134608012624,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "EleutherAI/coqa",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "{{query}}",
"doc_to_target": " {{answers[0]}}",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 50, \"until\": [\"\\n\\n\", \"Answer:\", \"Question:\"], \"do_sample\": false, \"temperature\": 0.0}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | coqa_mc | acc | none | 0.261646 | 0.005 | benchmark | null | coqa_mc | 0 | rank_choice | true | eng | 7,728 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990878+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 4027.134608012624,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/coqa-gen2mc",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "{{query_processed}}",
"doc_to_target": "{{answerKey}}",
"doc_to_choice": "{{choices.label}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | coqa_mc | acc_norm | none | 0.261646 | 0.005 | benchmark | null | coqa_mc | 0 | rank_choice | true | eng | 7,728 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:36:54.990878+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 4027.134608012624,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/coqa-gen2mc",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "{{query_processed}}",
"doc_to_target": "{{answerKey}}",
"doc_to_choice": "{{choices.label}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | coqa_rc | acc | none | 0.861931 | 0.003924 | benchmark | null | coqa_rc | 0 | rank_choice | true | eng | 7,728 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:37:36.061508+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3691.1114229094237,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/coqa-gen2mc",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "{{query_original}}",
"doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | coqa_rc | acc_norm | none | 0.794772 | 0.004594 | benchmark | null | coqa_rc | 0 | rank_choice | true | eng | 7,728 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:37:36.061508+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3691.1114229094237,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "allenai/coqa-gen2mc",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "{{query_original}}",
"doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | csqa_bpb | bits_per_byte | none | 0.821162 | 0.016462 | benchmark | null | csqa_bpb | 5 | bits_per_byte | false | eng | 1,221 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:50:06.069934+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 586.0866446048021,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "tau/commonsense_qa",
"output_type": "loglikelihood",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": " {{choices.text[choices.label.index(answerKey)]}}",
"doc_to_choice": null,
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | csqa_mc | acc | none | 0.208026 | 0.011621 | benchmark | null | csqa_mc | 5 | rank_choice | true | eng | 1,221 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:37:36.061508+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3691.1114229094237,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "tau/commonsense_qa",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "Question: {{question}}\n A. {{choices.text[0]}}\n B. {{choices.text[1]}}\n C. {{choices.text[2]}}\n D. {{choices.text[3]}}\n E. {{choices.text[4]}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "[\"A\", \"B\", \"C\", \"D\", \"E\"]",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | csqa_mc | acc_norm | none | 0.208026 | 0.011621 | benchmark | null | csqa_mc | 5 | rank_choice | true | eng | 1,221 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:37:36.061508+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3691.1114229094237,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "tau/commonsense_qa",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "Question: {{question}}\n A. {{choices.text[0]}}\n B. {{choices.text[1]}}\n C. {{choices.text[2]}}\n D. {{choices.text[3]}}\n E. {{choices.text[4]}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "[\"A\", \"B\", \"C\", \"D\", \"E\"]",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | csqa_rc | acc | none | 0.660934 | 0.013553 | benchmark | null | csqa_rc | 5 | rank_choice | true | eng | 1,221 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:37:36.061508+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3691.1114229094237,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "tau/commonsense_qa",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | csqa_rc | acc_norm | none | 0.6724 | 0.013437 | benchmark | null | csqa_rc | 5 | rank_choice | true | eng | 1,221 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:37:36.061508+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 3691.1114229094237,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "tau/commonsense_qa",
"output_type": "multiple_choice",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "Question: {{question}}\nAnswer:",
"doc_to_target": "{{choices.label.index(answerKey)}}",
"doc_to_choice": "{{choices.text}}",
"generation_kwargs": null,
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__linear_1d_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__linear_1d_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:50:06.069934+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 586.0866446048021,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__linear_1d_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__linear_1d_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:50:06.069934+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 586.0866446048021,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__linear_2d_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__linear_2d_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:51:57.737758+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1291.7352589257061,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__linear_2d_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__linear_2d_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:50:06.069934+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 586.0866446048021,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__polynomial_roots_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__polynomial_roots_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:51:57.737758+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1291.7352589257061,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__polynomial_roots_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__polynomial_roots_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:51:57.737758+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1291.7352589257061,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__sequence_next_term_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__sequence_next_term_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:51:57.737758+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1291.7352589257061,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_algebra__sequence_nth_term_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_algebra__sequence_nth_term_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:59:59.312265+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 7977.657158617862,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__add_or_sub_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__add_or_sub_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:59:59.312265+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 7977.657158617862,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__add_or_sub_in_base_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__add_or_sub_in_base_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:59:59.312265+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 7977.657158617862,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__add_sub_multiple_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__add_sub_multiple_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-11T23:59:59.312265+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 7977.657158617862,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__div_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__div_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:00:39.698333+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1407.1599106695503,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__mixed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__mixed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:00:39.698333+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1407.1599106695503,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__mul_cot | exact_match | none | 0.01 | 0.01 | benchmark | null | deepmind_math_arithmetic__mul_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:00:39.698333+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1407.1599106695503,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__mul_div_multiple_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__mul_div_multiple_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:00:39.698333+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 1407.1599106695503,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__nearest_integer_root_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__nearest_integer_root_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:06:53.886686+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2344.535878072493,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_arithmetic__simplify_surd_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_arithmetic__simplify_surd_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:06:53.886686+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2344.535878072493,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_calculus__differentiate_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_calculus__differentiate_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:06:53.886686+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2344.535878072493,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_calculus__differentiate_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_calculus__differentiate_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:06:53.886686+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2344.535878072493,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__closest_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__closest_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:10:56.171717+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 373.42632418405265,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__closest_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__closest_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:10:56.171717+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 373.42632418405265,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__kth_biggest_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__kth_biggest_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:10:56.171717+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 373.42632418405265,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__kth_biggest_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__kth_biggest_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:10:56.171717+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 373.42632418405265,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__pair_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__pair_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:13:39.802607+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2702.589003068395,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__pair_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__pair_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:13:39.802607+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2702.589003068395,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__sort_composed_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__sort_composed_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:13:39.802607+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2702.589003068395,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_comparison__sort_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_comparison__sort_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:13:39.802607+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 2702.589003068395,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_cot | exact_match | none | 0.000179 | null | task_group | eng_base_full | deepmind_math_cot | null | null | true | eng | 5,600 | {"deepmind_math_cot": ["deepmind_math_algebra__linear_1d_cot", "deepmind_math_algebra__linear_1d_composed_cot", "deepmind_math_algebra__linear_2d_cot", "deepmind_math_algebra__linear_2d_composed_cot", "deepmind_math_algebra__polynomial_roots_cot", "deepmind_math_algebra__polynomial_roots_composed_cot", "deepmind_math_algebra__sequence_next_term_cot", "deepmind_math_algebra__sequence_nth_term_cot", "deepmind_math_arithmetic__add_or_sub_cot", "deepmind_math_arithmetic__add_or_sub_in_base_cot", "deepmind_math_arithmetic__add_sub_multiple_cot", "deepmind_math_arithmetic__div_cot", "deepmind_math_arithmetic__mixed_cot", "deepmind_math_arithmetic__mul_cot", "deepmind_math_arithmetic__mul_div_multiple_cot", "deepmind_math_arithmetic__nearest_integer_root_cot", "deepmind_math_arithmetic__simplify_surd_cot", "deepmind_math_calculus__differentiate_cot", "deepmind_math_calculus__differentiate_composed_cot", "deepmind_math_comparison__closest_cot", "deepmind_math_comparison__closest_composed_cot", "deepmind_math_comparison__kth_biggest_cot", "deepmind_math_comparison__kth_biggest_composed_cot", "deepmind_math_comparison__pair_cot", "deepmind_math_comparison__pair_composed_cot", "deepmind_math_comparison__sort_cot", "deepmind_math_comparison__sort_composed_cot", "deepmind_math_measurement__conversion_cot", "deepmind_math_measurement__time_cot", "deepmind_math_numbers__base_conversion_cot", "deepmind_math_numbers__div_remainder_cot", "deepmind_math_numbers__div_remainder_composed_cot", "deepmind_math_numbers__gcd_cot", "deepmind_math_numbers__gcd_composed_cot", "deepmind_math_numbers__is_factor_cot", "deepmind_math_numbers__is_factor_composed_cot", "deepmind_math_numbers__is_prime_cot", "deepmind_math_numbers__is_prime_composed_cot", "deepmind_math_numbers__lcm_cot", "deepmind_math_numbers__lcm_composed_cot", "deepmind_math_numbers__list_prime_factors_cot", "deepmind_math_numbers__list_prime_factors_composed_cot", "deepmind_math_numbers__place_value_cot", "deepmind_math_numbers__place_value_composed_cot", "deepmind_math_numbers__round_number_cot", "deepmind_math_numbers__round_number_composed_cot", "deepmind_math_polynomials__add_cot", "deepmind_math_polynomials__coefficient_named_cot", "deepmind_math_polynomials__collect_cot", "deepmind_math_polynomials__compose_cot", "deepmind_math_polynomials__evaluate_cot", "deepmind_math_polynomials__evaluate_composed_cot", "deepmind_math_polynomials__expand_cot", "deepmind_math_polynomials__simplify_power_cot", "deepmind_math_probability__swr_p_level_set_cot", "deepmind_math_probability__swr_p_sequence_cot"]} | 4,194,304 | 209,715,200,000 | 2026-03-12T00:56:59.492351+00:00 | null | null | null | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_measurement__conversion_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_measurement__conversion_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:17:14.071702+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_020/results_2026-03-12T04-24-26.208957.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 11260.004700644873,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_measurement__time_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_measurement__time_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:17:14.071702+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_020/results_2026-03-12T04-24-26.208957.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 11260.004700644873,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
apertus_8b | 50,000 | deepmind_math_numbers__base_conversion_cot | exact_match | none | 0 | 0 | benchmark | null | deepmind_math_numbers__base_conversion_cot | 5 | generative | true | eng | 100 | null | 4,194,304 | 209,715,200,000 | 2026-03-12T00:17:14.071702+00:00 | /leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_020/results_2026-03-12T04-24-26.208957.json | {
"lm_eval_version": "0.4.12.dev0",
"git_hash": "2b2cde4",
"transformers_version": "5.3.0",
"model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B",
"model_source": "local-completions",
"max_length": 4095,
"total_eval_time_seconds": 11260.004700644873,
"tokenizer_pad_token": "<pad>",
"tokenizer_eos_token": "</s>",
"tokenizer_bos_token": "<s>",
"eot_token_id": 2,
"chat_template": null,
"chat_template_sha": null,
"system_instruction": null,
"system_instruction_sha": null
} | {
"task_version": "1.0",
"task_hash": null,
"dataset_path": "ellamind/deepmind-math-sample",
"output_type": "generate_until",
"dataset_name": null,
"repeats": 1,
"doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n",
"doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n",
"doc_to_choice": null,
"generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}",
"should_decontaminate": false
} | /leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B | true | Apertus 8B |
End of preview. Expand
in Data Studio
No dataset card yet
- Downloads last month
- -