Dataset Viewer
Auto-converted to Parquet Duplicate
model
large_stringclasses
3 values
step
int64
10k
2.63M
task
large_stringclasses
767 values
metric
large_stringclasses
20 values
metric_filter
large_stringclasses
6 values
score
float64
0
4.45
score_stderr
float64
0
0.25
task_type
large_stringclasses
3 values
parent_task
large_stringclasses
7 values
task_display_name
large_stringclasses
767 values
num_fewshot
int64
0
8
task_formulation
large_stringclasses
3 values
higher_is_better
bool
2 classes
language
large_stringclasses
2 values
n_samples
int64
23
180k
subtask_tree
large_stringclasses
44 values
train_batch_size
int64
4.19M
8.39M
tokens_trained
int64
41.9B
22,038B
eval_date
large_stringdate
2026-03-10 17:22:50
2026-03-12 12:50:40
result_source
large_stringlengths
160
168
eval_config
dict
task_config
dict
model_path
large_stringclasses
95 values
is_checkpoint
bool
2 classes
model_display_name
large_stringclasses
3 values
apertus_8b
50,000
arc_challenge_bpb
bits_per_byte
none
0.777649
0.013966
benchmark
null
arc_challenge_bpb
5
bits_per_byte
false
eng
1,172
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693991+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2467.716009796597, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "loglikelihood", "dataset_name": "ARC-Challenge", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{choices.text[choices.label.index(answerKey)]}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_challenge_mc
acc
none
0.263468
0.009039
benchmark
null
arc_challenge_mc
5
rank_choice
true
eng
2,376
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694601+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3745.1029282175004, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Easy", "repeats": 1, "doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:", "doc_to_target": "{{answerKey}}", "doc_to_choice": "{{choices.label}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_challenge_mc
acc_norm
none
0.263468
0.009039
benchmark
null
arc_challenge_mc
5
rank_choice
true
eng
2,376
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694601+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3745.1029282175004, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Easy", "repeats": 1, "doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:", "doc_to_target": "{{answerKey}}", "doc_to_choice": "{{choices.label}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_challenge_rc
acc
none
0.41041
0.014375
benchmark
null
arc_challenge_rc
5
rank_choice
true
eng
1,172
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693991+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2467.716009796597, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Challenge", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_challenge_rc
acc_norm
none
0.449659
0.014537
benchmark
null
arc_challenge_rc
5
rank_choice
true
eng
1,172
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693991+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2467.716009796597, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Challenge", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_easy_bpb
bits_per_byte
none
0.580533
0.007264
benchmark
null
arc_easy_bpb
5
bits_per_byte
false
eng
2,376
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694601+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3745.1029282175004, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "loglikelihood", "dataset_name": "ARC-Easy", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{choices.text[choices.label.index(answerKey)]}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_easy_mc
acc
none
0.244529
0.008819
benchmark
null
arc_easy_mc
5
rank_choice
true
eng
2,376
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694601+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3745.1029282175004, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Easy", "repeats": 1, "doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:", "doc_to_target": "{{answerKey}}", "doc_to_choice": "{{choices.label}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_easy_mc
acc_norm
none
0.244529
0.008819
benchmark
null
arc_easy_mc
5
rank_choice
true
eng
2,376
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694601+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3745.1029282175004, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Easy", "repeats": 1, "doc_to_text": "Question: {{question}}\n{% for i in range(choices.text|length) %} {{'ABCDE'[i]}}. {{choices.text[i]}}\n{% endfor %}Answer:", "doc_to_target": "{{answerKey}}", "doc_to_choice": "{{choices.label}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_easy_rc
acc
none
0.739478
0.009006
benchmark
null
arc_easy_rc
5
rank_choice
true
eng
2,376
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694601+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3745.1029282175004, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Easy", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
arc_easy_rc
acc_norm
none
0.737374
0.00903
benchmark
null
arc_easy_rc
5
rank_choice
true
eng
2,376
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694601+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_000/results_2026-03-12T01-31-41.965560.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3745.1029282175004, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/ai2_arc", "output_type": "multiple_choice", "dataset_name": "ARC-Easy", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_arithmetic_bpb
bits_per_byte
none
1.139799
0.029688
benchmark
null
basic_skills_arithmetic_bpb
5
bits_per_byte
false
eng
1,047
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596432+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1829.1680674208328, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "loglikelihood", "dataset_name": "arithmetic", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{answer}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_arithmetic_rc
acc
none
0.56638
0.015323
benchmark
null
basic_skills_arithmetic_rc
5
rank_choice
true
eng
1,047
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596432+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1829.1680674208328, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "arithmetic", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_arithmetic_rc
acc_norm
none
0.587393
0.015222
benchmark
null
basic_skills_arithmetic_rc
5
rank_choice
true
eng
1,047
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596432+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1829.1680674208328, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "arithmetic", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_coding_bpb
bits_per_byte
none
0.560042
0.009872
benchmark
null
basic_skills_coding_bpb
5
bits_per_byte
false
eng
1,012
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596733+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 929.7853837599978, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "loglikelihood", "dataset_name": "coding", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{answer}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_coding_rc
acc
none
0.523715
0.015707
benchmark
null
basic_skills_coding_rc
5
rank_choice
true
eng
1,012
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596432+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1829.1680674208328, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "coding", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_coding_rc
acc_norm
none
0.672925
0.014755
benchmark
null
basic_skills_coding_rc
5
rank_choice
true
eng
1,012
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596432+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1829.1680674208328, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "coding", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_common_knowledge_bpb
bits_per_byte
none
0.172218
0.00593
benchmark
null
basic_skills_common_knowledge_bpb
5
bits_per_byte
false
eng
1,037
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596733+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 929.7853837599978, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "loglikelihood", "dataset_name": "common_knowledge", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{answer}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_common_knowledge_rc
acc
none
0.937319
0.007531
benchmark
null
basic_skills_common_knowledge_rc
5
rank_choice
true
eng
1,037
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596733+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 929.7853837599978, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "common_knowledge", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_common_knowledge_rc
acc_norm
none
0.922854
0.00829
benchmark
null
basic_skills_common_knowledge_rc
5
rank_choice
true
eng
1,037
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596733+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 929.7853837599978, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "common_knowledge", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_logical_reasoning_bpb
bits_per_byte
none
0.264452
0.005103
benchmark
null
basic_skills_logical_reasoning_bpb
5
bits_per_byte
false
eng
1,118
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990618+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 804.4468631707132, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "loglikelihood", "dataset_name": "logical_reasoning", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{answer}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_logical_reasoning_rc
acc
none
0.966011
0.005422
benchmark
null
basic_skills_logical_reasoning_rc
5
rank_choice
true
eng
1,118
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596733+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 929.7853837599978, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "logical_reasoning", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_logical_reasoning_rc
acc_norm
none
0.974955
0.004675
benchmark
null
basic_skills_logical_reasoning_rc
5
rank_choice
true
eng
1,118
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596733+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_009/results_2026-03-12T00-51-24.407230.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 929.7853837599978, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "logical_reasoning", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_pattern_bpb
bits_per_byte
none
1.300733
0.039628
benchmark
null
basic_skills_pattern_bpb
5
bits_per_byte
false
eng
534
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990618+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 804.4468631707132, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "loglikelihood", "dataset_name": "pattern", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{answer}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_pattern_rc
acc
none
0.773408
0.018133
benchmark
null
basic_skills_pattern_rc
5
rank_choice
true
eng
534
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990618+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 804.4468631707132, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "pattern", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_pattern_rc
acc_norm
none
0.775281
0.018079
benchmark
null
basic_skills_pattern_rc
5
rank_choice
true
eng
534
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990618+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 804.4468631707132, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "pattern", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_rc
acc
none
0.559694
0.015349
benchmark
null
basic_skills_rc
5
rank_choice
true
eng
1,047
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990618+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 804.4468631707132, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "arithmetic", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_rc
acc_norm
none
0.582617
0.015247
benchmark
null
basic_skills_rc
5
rank_choice
true
eng
1,047
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990618+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_010/results_2026-03-12T00-49-30.385033.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 804.4468631707132, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "arithmetic", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_string_operations_bpb
bits_per_byte
none
1.185186
0.032276
benchmark
null
basic_skills_string_operations_bpb
5
bits_per_byte
false
eng
1,219
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990878+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 4027.134608012624, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "loglikelihood", "dataset_name": "string_operations", "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{answer}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_string_operations_rc
acc
none
0.303527
0.013174
benchmark
null
basic_skills_string_operations_rc
5
rank_choice
true
eng
1,219
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990878+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 4027.134608012624, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "string_operations", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
basic_skills_string_operations_rc
acc_norm
none
0.401148
0.014044
benchmark
null
basic_skills_string_operations_rc
5
rank_choice
true
eng
1,219
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990878+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 4027.134608012624, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/basic-skills", "output_type": "multiple_choice", "dataset_name": "string_operations", "repeats": 1, "doc_to_text": "Question: {{question}}", "doc_to_target": "0", "doc_to_choice": "{{[answer] + wrong_answers[:3]}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_boolean_expressions_cot
exact_match
get-answer
0.528
0.031636
benchmark
null
bbh_boolean_expressions_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693991+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2467.716009796597, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "boolean_expressions", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_causal_judgement_cot
exact_match
get-answer
0.475936
0.036619
benchmark
null
bbh_causal_judgement_cot
3
generative
true
eng
187
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693991+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_001/results_2026-03-12T01-10-24.578504.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2467.716009796597, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "causal_judgement", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_date_understanding_cot
exact_match
get-answer
0.212
0.025902
benchmark
null
bbh_date_understanding_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692865+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1844.1305395988747, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "date_understanding", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_disambiguation_qa_cot
exact_match
get-answer
0.408
0.031145
benchmark
null
bbh_disambiguation_qa_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692865+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1844.1305395988747, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "disambiguation_qa", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_dyck_languages_cot
exact_match
get-answer
0.008
0.005645
benchmark
null
bbh_dyck_languages_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692865+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1844.1305395988747, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "dyck_languages", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_formal_fallacies_cot
exact_match
get-answer
0.508
0.031682
benchmark
null
bbh_formal_fallacies_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692865+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_002/results_2026-03-12T01-00-00.993112.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1844.1305395988747, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "formal_fallacies", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_geometric_shapes_cot
exact_match
get-answer
0.14
0.021989
benchmark
null
bbh_geometric_shapes_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692869+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 444.9072736380622, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "geometric_shapes", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_hyperbaton_cot
exact_match
get-answer
0.528
0.031636
benchmark
null
bbh_hyperbaton_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692869+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 444.9072736380622, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "hyperbaton", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_logical_deduction_five_objects_cot
exact_match
get-answer
0.256
0.027657
benchmark
null
bbh_logical_deduction_five_objects_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692869+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 444.9072736380622, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "logical_deduction_five_objects", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_logical_deduction_seven_objects_cot
exact_match
get-answer
0.164
0.023465
benchmark
null
bbh_logical_deduction_seven_objects_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692869+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_003/results_2026-03-12T00-36-41.769791.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 444.9072736380622, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "logical_deduction_seven_objects", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_logical_deduction_three_objects_cot
exact_match
get-answer
0.324
0.029658
benchmark
null
bbh_logical_deduction_three_objects_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694015+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 406.35265475604683, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "logical_deduction_three_objects", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_movie_recommendation_cot
exact_match
get-answer
0.308
0.029257
benchmark
null
bbh_movie_recommendation_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694015+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 406.35265475604683, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "movie_recommendation", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_multistep_arithmetic_two_cot
exact_match
get-answer
0.008
0.005645
benchmark
null
bbh_multistep_arithmetic_two_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694015+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 406.35265475604683, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "multistep_arithmetic_two", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_navigate_cot
exact_match
get-answer
0.6
0.031046
benchmark
null
bbh_navigate_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694015+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_004/results_2026-03-12T00-36-03.215208.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 406.35265475604683, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "navigate", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_object_counting_cot
exact_match
get-answer
0.276
0.028329
benchmark
null
bbh_object_counting_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694321+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.3605205435306, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "object_counting", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_penguins_in_a_table_cot
exact_match
get-answer
0.157534
0.030254
benchmark
null
bbh_penguins_in_a_table_cot
3
generative
true
eng
146
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694321+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.3605205435306, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "penguins_in_a_table", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_reasoning_about_colored_objects_cot
exact_match
get-answer
0.16
0.023233
benchmark
null
bbh_reasoning_about_colored_objects_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694321+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.3605205435306, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "reasoning_about_colored_objects", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_ruin_names_cot
exact_match
get-answer
0.224
0.026421
benchmark
null
bbh_ruin_names_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.694321+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_005/results_2026-03-12T00-35-51.223020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.3605205435306, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "ruin_names", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_salient_translation_error_detection_cot
exact_match
get-answer
0.2
0.025349
benchmark
null
bbh_salient_translation_error_detection_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692898+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.1463885176927, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "salient_translation_error_detection", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_snarks_cot
exact_match
get-answer
0.550562
0.03739
benchmark
null
bbh_snarks_cot
3
generative
true
eng
178
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692898+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.1463885176927, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "snarks", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_sports_understanding_cot
exact_match
get-answer
0.504
0.031685
benchmark
null
bbh_sports_understanding_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692898+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.1463885176927, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "sports_understanding", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_temporal_sequences_cot
exact_match
get-answer
0.252
0.027514
benchmark
null
bbh_temporal_sequences_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.692898+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_006/results_2026-03-12T00-35-51.008960.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 394.1463885176927, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "temporal_sequences", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_tracking_shuffled_objects_five_objects_cot
exact_match
get-answer
0.2
0.025349
benchmark
null
bbh_tracking_shuffled_objects_five_objects_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693084+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 403.64010308589786, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "tracking_shuffled_objects_five_objects", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_tracking_shuffled_objects_seven_objects_cot
exact_match
get-answer
0.096
0.018669
benchmark
null
bbh_tracking_shuffled_objects_seven_objects_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693084+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 403.64010308589786, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "tracking_shuffled_objects_seven_objects", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_tracking_shuffled_objects_three_objects_cot
exact_match
get-answer
0.332
0.029844
benchmark
null
bbh_tracking_shuffled_objects_three_objects_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693084+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 403.64010308589786, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "tracking_shuffled_objects_three_objects", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_web_of_lies_cot
exact_match
get-answer
0.532
0.031621
benchmark
null
bbh_web_of_lies_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:30:48.693084+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_007/results_2026-03-12T00-36-00.502617.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 403.64010308589786, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "web_of_lies", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
bbh_word_sorting_cot
exact_match
get-answer
0.028
0.010455
benchmark
null
bbh_word_sorting_cot
3
generative
true
eng
250
null
4,194,304
209,715,200,000
2026-03-11T23:36:35.596432+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_008/results_2026-03-12T01-06-23.789690.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1829.1680674208328, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "lukaemon/bbh", "output_type": "generate_until", "dataset_name": "word_sorting", "repeats": 1, "doc_to_text": "Q: {{input}}\nA: Let's think step by step.\n", "doc_to_target": "{{target}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 512, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Q:\", \"\\n\\n\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
coqa_bpb
bits_per_byte
none
0.695591
0.030972
benchmark
null
coqa_bpb
0
bits_per_byte
false
eng
500
null
4,194,304
209,715,200,000
2026-03-11T23:37:36.061508+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3691.1114229094237, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "EleutherAI/coqa", "output_type": "loglikelihood", "dataset_name": null, "repeats": 1, "doc_to_text": "Story: {{story}}\n\nQ: {{questions.input_text[0]}}\nA:", "doc_to_target": " {{answers.input_text[0]}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
coqa_gen
em
none
0.593762
0.005497
benchmark
null
coqa_gen
0
generative
true
eng
7,983
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990878+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 4027.134608012624, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "EleutherAI/coqa", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "{{query}}", "doc_to_target": " {{answers[0]}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 50, \"until\": [\"\\n\\n\", \"Answer:\", \"Question:\"], \"do_sample\": false, \"temperature\": 0.0}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
coqa_gen
f1
none
0.7099
0.004537
benchmark
null
coqa_gen
0
generative
true
eng
7,983
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990878+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 4027.134608012624, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "EleutherAI/coqa", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "{{query}}", "doc_to_target": " {{answers[0]}}", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 50, \"until\": [\"\\n\\n\", \"Answer:\", \"Question:\"], \"do_sample\": false, \"temperature\": 0.0}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
coqa_mc
acc
none
0.261646
0.005
benchmark
null
coqa_mc
0
rank_choice
true
eng
7,728
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990878+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 4027.134608012624, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/coqa-gen2mc", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "{{query_processed}}", "doc_to_target": "{{answerKey}}", "doc_to_choice": "{{choices.label}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
coqa_mc
acc_norm
none
0.261646
0.005
benchmark
null
coqa_mc
0
rank_choice
true
eng
7,728
null
4,194,304
209,715,200,000
2026-03-11T23:36:54.990878+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_011/results_2026-03-12T01-43-14.998530.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 4027.134608012624, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/coqa-gen2mc", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "{{query_processed}}", "doc_to_target": "{{answerKey}}", "doc_to_choice": "{{choices.label}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
coqa_rc
acc
none
0.861931
0.003924
benchmark
null
coqa_rc
0
rank_choice
true
eng
7,728
null
4,194,304
209,715,200,000
2026-03-11T23:37:36.061508+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3691.1114229094237, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/coqa-gen2mc", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "{{query_original}}", "doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
coqa_rc
acc_norm
none
0.794772
0.004594
benchmark
null
coqa_rc
0
rank_choice
true
eng
7,728
null
4,194,304
209,715,200,000
2026-03-11T23:37:36.061508+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3691.1114229094237, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "allenai/coqa-gen2mc", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "{{query_original}}", "doc_to_target": "{{ ['A', 'B', 'C', 'D', 'E'].index(answerKey) }}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
csqa_bpb
bits_per_byte
none
0.821162
0.016462
benchmark
null
csqa_bpb
5
bits_per_byte
false
eng
1,221
null
4,194,304
209,715,200,000
2026-03-11T23:50:06.069934+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 586.0866446048021, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "tau/commonsense_qa", "output_type": "loglikelihood", "dataset_name": null, "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": " {{choices.text[choices.label.index(answerKey)]}}", "doc_to_choice": null, "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
csqa_mc
acc
none
0.208026
0.011621
benchmark
null
csqa_mc
5
rank_choice
true
eng
1,221
null
4,194,304
209,715,200,000
2026-03-11T23:37:36.061508+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3691.1114229094237, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "tau/commonsense_qa", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "Question: {{question}}\n A. {{choices.text[0]}}\n B. {{choices.text[1]}}\n C. {{choices.text[2]}}\n D. {{choices.text[3]}}\n E. {{choices.text[4]}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "[\"A\", \"B\", \"C\", \"D\", \"E\"]", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
csqa_mc
acc_norm
none
0.208026
0.011621
benchmark
null
csqa_mc
5
rank_choice
true
eng
1,221
null
4,194,304
209,715,200,000
2026-03-11T23:37:36.061508+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3691.1114229094237, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "tau/commonsense_qa", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "Question: {{question}}\n A. {{choices.text[0]}}\n B. {{choices.text[1]}}\n C. {{choices.text[2]}}\n D. {{choices.text[3]}}\n E. {{choices.text[4]}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "[\"A\", \"B\", \"C\", \"D\", \"E\"]", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
csqa_rc
acc
none
0.660934
0.013553
benchmark
null
csqa_rc
5
rank_choice
true
eng
1,221
null
4,194,304
209,715,200,000
2026-03-11T23:37:36.061508+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3691.1114229094237, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "tau/commonsense_qa", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
csqa_rc
acc_norm
none
0.6724
0.013437
benchmark
null
csqa_rc
5
rank_choice
true
eng
1,221
null
4,194,304
209,715,200,000
2026-03-11T23:37:36.061508+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_012/results_2026-03-12T01-38-20.072777.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 3691.1114229094237, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "tau/commonsense_qa", "output_type": "multiple_choice", "dataset_name": null, "repeats": 1, "doc_to_text": "Question: {{question}}\nAnswer:", "doc_to_target": "{{choices.label.index(answerKey)}}", "doc_to_choice": "{{choices.text}}", "generation_kwargs": null, "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__linear_1d_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__linear_1d_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:50:06.069934+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 586.0866446048021, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__linear_1d_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__linear_1d_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:50:06.069934+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 586.0866446048021, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__linear_2d_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__linear_2d_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:51:57.737758+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1291.7352589257061, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__linear_2d_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__linear_2d_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:50:06.069934+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_013/results_2026-03-12T00-59-20.291272.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 586.0866446048021, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__polynomial_roots_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__polynomial_roots_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:51:57.737758+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1291.7352589257061, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__polynomial_roots_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__polynomial_roots_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:51:57.737758+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1291.7352589257061, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__sequence_next_term_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__sequence_next_term_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:51:57.737758+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_014/results_2026-03-12T01-12-59.709648.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1291.7352589257061, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_algebra__sequence_nth_term_cot
exact_match
none
0
0
benchmark
null
deepmind_math_algebra__sequence_nth_term_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:59:59.312265+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 7977.657158617862, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__add_or_sub_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__add_or_sub_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:59:59.312265+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 7977.657158617862, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__add_or_sub_in_base_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__add_or_sub_in_base_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:59:59.312265+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 7977.657158617862, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__add_sub_multiple_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__add_sub_multiple_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-11T23:59:59.312265+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_015/results_2026-03-12T03-12-22.431042.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 7977.657158617862, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__div_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__div_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:00:39.698333+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1407.1599106695503, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__mixed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__mixed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:00:39.698333+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1407.1599106695503, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__mul_cot
exact_match
none
0.01
0.01
benchmark
null
deepmind_math_arithmetic__mul_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:00:39.698333+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1407.1599106695503, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__mul_div_multiple_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__mul_div_multiple_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:00:39.698333+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_016/results_2026-03-12T01-23-32.386020.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 1407.1599106695503, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__nearest_integer_root_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__nearest_integer_root_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:06:53.886686+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2344.535878072493, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_arithmetic__simplify_surd_cot
exact_match
none
0
0
benchmark
null
deepmind_math_arithmetic__simplify_surd_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:06:53.886686+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2344.535878072493, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_calculus__differentiate_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_calculus__differentiate_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:06:53.886686+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2344.535878072493, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_calculus__differentiate_cot
exact_match
none
0
0
benchmark
null
deepmind_math_calculus__differentiate_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:06:53.886686+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_017/results_2026-03-12T01-45-31.963338.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2344.535878072493, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__closest_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__closest_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:10:56.171717+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 373.42632418405265, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__closest_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__closest_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:10:56.171717+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 373.42632418405265, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__kth_biggest_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__kth_biggest_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:10:56.171717+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 373.42632418405265, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__kth_biggest_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__kth_biggest_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:10:56.171717+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_018/results_2026-03-12T01-16-42.254263.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 373.42632418405265, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__pair_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__pair_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:13:39.802607+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2702.589003068395, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__pair_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__pair_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:13:39.802607+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2702.589003068395, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__sort_composed_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__sort_composed_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:13:39.802607+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2702.589003068395, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_comparison__sort_cot
exact_match
none
0
0
benchmark
null
deepmind_math_comparison__sort_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:13:39.802607+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_019/results_2026-03-12T01-58-06.733521.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 2702.589003068395, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_cot
exact_match
none
0.000179
null
task_group
eng_base_full
deepmind_math_cot
null
null
true
eng
5,600
{"deepmind_math_cot": ["deepmind_math_algebra__linear_1d_cot", "deepmind_math_algebra__linear_1d_composed_cot", "deepmind_math_algebra__linear_2d_cot", "deepmind_math_algebra__linear_2d_composed_cot", "deepmind_math_algebra__polynomial_roots_cot", "deepmind_math_algebra__polynomial_roots_composed_cot", "deepmind_math_algebra__sequence_next_term_cot", "deepmind_math_algebra__sequence_nth_term_cot", "deepmind_math_arithmetic__add_or_sub_cot", "deepmind_math_arithmetic__add_or_sub_in_base_cot", "deepmind_math_arithmetic__add_sub_multiple_cot", "deepmind_math_arithmetic__div_cot", "deepmind_math_arithmetic__mixed_cot", "deepmind_math_arithmetic__mul_cot", "deepmind_math_arithmetic__mul_div_multiple_cot", "deepmind_math_arithmetic__nearest_integer_root_cot", "deepmind_math_arithmetic__simplify_surd_cot", "deepmind_math_calculus__differentiate_cot", "deepmind_math_calculus__differentiate_composed_cot", "deepmind_math_comparison__closest_cot", "deepmind_math_comparison__closest_composed_cot", "deepmind_math_comparison__kth_biggest_cot", "deepmind_math_comparison__kth_biggest_composed_cot", "deepmind_math_comparison__pair_cot", "deepmind_math_comparison__pair_composed_cot", "deepmind_math_comparison__sort_cot", "deepmind_math_comparison__sort_composed_cot", "deepmind_math_measurement__conversion_cot", "deepmind_math_measurement__time_cot", "deepmind_math_numbers__base_conversion_cot", "deepmind_math_numbers__div_remainder_cot", "deepmind_math_numbers__div_remainder_composed_cot", "deepmind_math_numbers__gcd_cot", "deepmind_math_numbers__gcd_composed_cot", "deepmind_math_numbers__is_factor_cot", "deepmind_math_numbers__is_factor_composed_cot", "deepmind_math_numbers__is_prime_cot", "deepmind_math_numbers__is_prime_composed_cot", "deepmind_math_numbers__lcm_cot", "deepmind_math_numbers__lcm_composed_cot", "deepmind_math_numbers__list_prime_factors_cot", "deepmind_math_numbers__list_prime_factors_composed_cot", "deepmind_math_numbers__place_value_cot", "deepmind_math_numbers__place_value_composed_cot", "deepmind_math_numbers__round_number_cot", "deepmind_math_numbers__round_number_composed_cot", "deepmind_math_polynomials__add_cot", "deepmind_math_polynomials__coefficient_named_cot", "deepmind_math_polynomials__collect_cot", "deepmind_math_polynomials__compose_cot", "deepmind_math_polynomials__evaluate_cot", "deepmind_math_polynomials__evaluate_composed_cot", "deepmind_math_polynomials__expand_cot", "deepmind_math_polynomials__simplify_power_cot", "deepmind_math_probability__swr_p_level_set_cot", "deepmind_math_probability__swr_p_sequence_cot"]}
4,194,304
209,715,200,000
2026-03-12T00:56:59.492351+00:00
null
null
null
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_measurement__conversion_cot
exact_match
none
0
0
benchmark
null
deepmind_math_measurement__conversion_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:17:14.071702+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_020/results_2026-03-12T04-24-26.208957.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 11260.004700644873, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_measurement__time_cot
exact_match
none
0
0
benchmark
null
deepmind_math_measurement__time_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:17:14.071702+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_020/results_2026-03-12T04-24-26.208957.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 11260.004700644873, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
apertus_8b
50,000
deepmind_math_numbers__base_conversion_cot
exact_match
none
0
0
benchmark
null
deepmind_math_numbers__base_conversion_cot
5
generative
true
eng
100
null
4,194,304
209,715,200,000
2026-03-12T00:17:14.071702+00:00
/leonardo_work/AIFAC_L01_028/midahl00/eval-hive-2/eval_results/ref_models/apertus_8b/step50000-tokens210B/batch_37416037_020/results_2026-03-12T04-24-26.208957.json
{ "lm_eval_version": "0.4.12.dev0", "git_hash": "2b2cde4", "transformers_version": "5.3.0", "model_name": "/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B", "model_source": "local-completions", "max_length": 4095, "total_eval_time_seconds": 11260.004700644873, "tokenizer_pad_token": "<pad>", "tokenizer_eos_token": "</s>", "tokenizer_bos_token": "<s>", "eot_token_id": 2, "chat_template": null, "chat_template_sha": null, "system_instruction": null, "system_instruction_sha": null }
{ "task_version": "1.0", "task_hash": null, "dataset_path": "ellamind/deepmind-math-sample", "output_type": "generate_until", "dataset_name": null, "repeats": 1, "doc_to_text": "def doc_to_text(doc):\n \"\"\"Build prompt for a math question.\"\"\"\n return _clean_bytes(doc[\"question\"])\n", "doc_to_target": "def doc_to_target(doc):\n \"\"\"Return the gold answer.\"\"\"\n return \" \" + _clean_bytes(doc[\"answer\"])\n", "doc_to_choice": null, "generation_kwargs": "{\"max_gen_toks\": 1024, \"do_sample\": true, \"temperature\": 0.6, \"top_p\": 0.6, \"until\": [\"Problem:\", \"Answer:\", \"Question:\", \"</s>\", \"<|eot_id|>\"]}", "should_decontaminate": false }
/leonardo_work/AIFAC_L01_028/models/swiss-ai/Apertus-8B-2509/step50000-tokens210B
true
Apertus 8B
End of preview. Expand in Data Studio

No dataset card yet

Downloads last month
-