tyzhu commited on
Commit
25de0f9
·
verified ·
1 Parent(s): e6aaa32

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. 10000_hf/config.json +35 -0
  2. 10000_hf/generation_config.json +6 -0
  3. 10000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/results_2026-01-07T03-56-14.560763.json +240 -0
  4. 10000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_arc_challenge_2026-01-07T03-56-14.560763.jsonl +0 -0
  5. 10000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_arc_easy_2026-01-07T03-56-14.560763.jsonl +0 -0
  6. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/results_2026-01-07T03-56-07.752198.json +478 -0
  7. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_algebra_2026-01-07T03-56-07.752198.jsonl +0 -0
  8. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_counting_and_prob_2026-01-07T03-56-07.752198.jsonl +0 -0
  9. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_geometry_2026-01-07T03-56-07.752198.jsonl +0 -0
  10. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_intermediate_algebra_2026-01-07T03-56-07.752198.jsonl +0 -0
  11. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_num_theory_2026-01-07T03-56-07.752198.jsonl +0 -0
  12. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_prealgebra_2026-01-07T03-56-07.752198.jsonl +0 -0
  13. 10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_precalc_2026-01-07T03-56-07.752198.jsonl +0 -0
  14. 10000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/results_2026-01-07T07-13-00.293068.json +290 -0
  15. 10000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_gsm8k_cot_2026-01-07T07-13-00.293068.jsonl +0 -0
  16. 10000_hf/special_tokens_map.json +16 -0
  17. 10000_hf/tokenizer_config.json +2062 -0
  18. 12500/checkpoint_metadata.json +24 -0
  19. 12500/config.yaml +141 -0
  20. 12500/model_config.json +1 -0
  21. 12500/optimizer/optimizer_config.json +1 -0
  22. 12500_hf/config.json +35 -0
  23. 12500_hf/generation_config.json +6 -0
  24. 12500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/results_2026-01-07T03-51-35.016035.json +240 -0
  25. 12500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_arc_challenge_2026-01-07T03-51-35.016035.jsonl +0 -0
  26. 12500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_arc_easy_2026-01-07T03-51-35.016035.jsonl +0 -0
  27. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/results_2026-01-07T03-56-06.178713.json +478 -0
  28. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_algebra_2026-01-07T03-56-06.178713.jsonl +0 -0
  29. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_counting_and_prob_2026-01-07T03-56-06.178713.jsonl +0 -0
  30. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_geometry_2026-01-07T03-56-06.178713.jsonl +0 -0
  31. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_intermediate_algebra_2026-01-07T03-56-06.178713.jsonl +0 -0
  32. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_num_theory_2026-01-07T03-56-06.178713.jsonl +0 -0
  33. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_prealgebra_2026-01-07T03-56-06.178713.jsonl +0 -0
  34. 12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_precalc_2026-01-07T03-56-06.178713.jsonl +0 -0
  35. 12500_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/results_2026-01-07T07-12-17.705915.json +290 -0
  36. 12500_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_gsm8k_cot_2026-01-07T07-12-17.705915.jsonl +0 -0
  37. 12500_hf/special_tokens_map.json +16 -0
  38. 12500_hf/tokenizer_config.json +2062 -0
  39. 2500_hf/config.json +35 -0
  40. 2500_hf/generation_config.json +6 -0
  41. 2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2025-12-29T05-34-34.570482.json +240 -0
  42. 2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2026-01-07T03-51-34.430609.json +240 -0
  43. 2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_challenge_2025-12-29T05-34-34.570482.jsonl +0 -0
  44. 2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_challenge_2026-01-07T03-51-34.430609.jsonl +0 -0
  45. 2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_easy_2025-12-29T05-34-34.570482.jsonl +0 -0
  46. 2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_easy_2026-01-07T03-51-34.430609.jsonl +0 -0
  47. 2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2025-12-29T06-24-24.342056.json +478 -0
  48. 2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2026-01-07T03-55-15.968865.json +478 -0
  49. 2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_hendrycks_math_algebra_2025-12-29T06-24-24.342056.jsonl +0 -0
  50. 2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_hendrycks_math_algebra_2026-01-07T03-55-15.968865.jsonl +0 -0
10000_hf/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 128001,
10
+ "head_dim": 64,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 2048,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 8192,
15
+ "max_position_embeddings": 131072,
16
+ "mlp_bias": false,
17
+ "model_type": "llama",
18
+ "num_attention_heads": 32,
19
+ "num_hidden_layers": 16,
20
+ "num_key_value_heads": 8,
21
+ "pretraining_tp": 1,
22
+ "rms_norm_eps": 1e-05,
23
+ "rope_scaling": {
24
+ "factor": 32.0,
25
+ "high_freq_factor": 4.0,
26
+ "low_freq_factor": 1.0,
27
+ "original_max_position_embeddings": 8192,
28
+ "rope_type": "llama3"
29
+ },
30
+ "rope_theta": 500000.0,
31
+ "tie_word_embeddings": true,
32
+ "transformers_version": "4.57.3",
33
+ "use_cache": true,
34
+ "vocab_size": 128256
35
+ }
10000_hf/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "eos_token_id": 128001,
5
+ "transformers_version": "4.57.3"
6
+ }
10000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/results_2026-01-07T03-56-14.560763.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "alias": "arc_challenge",
5
+ "acc,none": 0.35580204778157,
6
+ "acc_stderr,none": 0.013990571137918763,
7
+ "acc_norm,none": 0.3660409556313993,
8
+ "acc_norm_stderr,none": 0.01407722310847014
9
+ },
10
+ "arc_easy": {
11
+ "alias": "arc_easy",
12
+ "acc,none": 0.6405723905723906,
13
+ "acc_stderr,none": 0.009845958893373764,
14
+ "acc_norm,none": 0.6106902356902357,
15
+ "acc_norm_stderr,none": 0.010005212782878142
16
+ },
17
+ "hellaswag": {
18
+ "alias": "hellaswag",
19
+ "acc,none": 0.40967934674367656,
20
+ "acc_stderr,none": 0.004907694727935687,
21
+ "acc_norm,none": 0.5283808006373233,
22
+ "acc_norm_stderr,none": 0.004981736689518745
23
+ }
24
+ },
25
+ "group_subtasks": {
26
+ "arc_challenge": [],
27
+ "arc_easy": [],
28
+ "hellaswag": []
29
+ },
30
+ "configs": {
31
+ "arc_challenge": {
32
+ "task": "arc_challenge",
33
+ "tag": [
34
+ "ai2_arc"
35
+ ],
36
+ "dataset_path": "allenai/ai2_arc",
37
+ "dataset_name": "ARC-Challenge",
38
+ "training_split": "train",
39
+ "validation_split": "validation",
40
+ "test_split": "test",
41
+ "doc_to_text": "Question: {{question}}\nAnswer:",
42
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
43
+ "unsafe_code": false,
44
+ "doc_to_choice": "{{choices.text}}",
45
+ "description": "",
46
+ "target_delimiter": " ",
47
+ "fewshot_delimiter": "\n\n",
48
+ "num_fewshot": 0,
49
+ "metric_list": [
50
+ {
51
+ "metric": "acc",
52
+ "aggregation": "mean",
53
+ "higher_is_better": true
54
+ },
55
+ {
56
+ "metric": "acc_norm",
57
+ "aggregation": "mean",
58
+ "higher_is_better": true
59
+ }
60
+ ],
61
+ "output_type": "multiple_choice",
62
+ "repeats": 1,
63
+ "should_decontaminate": true,
64
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
65
+ "metadata": {
66
+ "version": 1.0,
67
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
68
+ }
69
+ },
70
+ "arc_easy": {
71
+ "task": "arc_easy",
72
+ "tag": [
73
+ "ai2_arc"
74
+ ],
75
+ "dataset_path": "allenai/ai2_arc",
76
+ "dataset_name": "ARC-Easy",
77
+ "training_split": "train",
78
+ "validation_split": "validation",
79
+ "test_split": "test",
80
+ "doc_to_text": "Question: {{question}}\nAnswer:",
81
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
82
+ "unsafe_code": false,
83
+ "doc_to_choice": "{{choices.text}}",
84
+ "description": "",
85
+ "target_delimiter": " ",
86
+ "fewshot_delimiter": "\n\n",
87
+ "num_fewshot": 0,
88
+ "metric_list": [
89
+ {
90
+ "metric": "acc",
91
+ "aggregation": "mean",
92
+ "higher_is_better": true
93
+ },
94
+ {
95
+ "metric": "acc_norm",
96
+ "aggregation": "mean",
97
+ "higher_is_better": true
98
+ }
99
+ ],
100
+ "output_type": "multiple_choice",
101
+ "repeats": 1,
102
+ "should_decontaminate": true,
103
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
104
+ "metadata": {
105
+ "version": 1.0,
106
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
107
+ }
108
+ },
109
+ "hellaswag": {
110
+ "task": "hellaswag",
111
+ "tag": [
112
+ "multiple_choice"
113
+ ],
114
+ "dataset_path": "Rowan/hellaswag",
115
+ "training_split": "train",
116
+ "validation_split": "validation",
117
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc):\n ctx = doc[\"ctx_a\"] + \" \" + doc[\"ctx_b\"].capitalize()\n out_doc = {\n \"query\": preprocess(doc[\"activity_label\"] + \": \" + ctx),\n \"choices\": [preprocess(ending) for ending in doc[\"endings\"]],\n \"gold\": int(doc[\"label\"]),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
118
+ "doc_to_text": "{{query}}",
119
+ "doc_to_target": "{{label}}",
120
+ "unsafe_code": false,
121
+ "doc_to_choice": "choices",
122
+ "description": "",
123
+ "target_delimiter": " ",
124
+ "fewshot_delimiter": "\n\n",
125
+ "num_fewshot": 0,
126
+ "metric_list": [
127
+ {
128
+ "metric": "acc",
129
+ "aggregation": "mean",
130
+ "higher_is_better": true
131
+ },
132
+ {
133
+ "metric": "acc_norm",
134
+ "aggregation": "mean",
135
+ "higher_is_better": true
136
+ }
137
+ ],
138
+ "output_type": "multiple_choice",
139
+ "repeats": 1,
140
+ "should_decontaminate": false,
141
+ "metadata": {
142
+ "version": 1.0,
143
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
144
+ }
145
+ }
146
+ },
147
+ "versions": {
148
+ "arc_challenge": 1.0,
149
+ "arc_easy": 1.0,
150
+ "hellaswag": 1.0
151
+ },
152
+ "n-shot": {
153
+ "arc_challenge": 0,
154
+ "arc_easy": 0,
155
+ "hellaswag": 0
156
+ },
157
+ "higher_is_better": {
158
+ "arc_challenge": {
159
+ "acc": true,
160
+ "acc_norm": true
161
+ },
162
+ "arc_easy": {
163
+ "acc": true,
164
+ "acc_norm": true
165
+ },
166
+ "hellaswag": {
167
+ "acc": true,
168
+ "acc_norm": true
169
+ }
170
+ },
171
+ "n-samples": {
172
+ "hellaswag": {
173
+ "original": 10042,
174
+ "effective": 10042
175
+ },
176
+ "arc_easy": {
177
+ "original": 2376,
178
+ "effective": 2376
179
+ },
180
+ "arc_challenge": {
181
+ "original": 1172,
182
+ "effective": 1172
183
+ }
184
+ },
185
+ "config": {
186
+ "model": "hf",
187
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/10000_hf",
188
+ "model_num_parameters": 1235814400,
189
+ "model_dtype": "torch.bfloat16",
190
+ "model_revision": "main",
191
+ "model_sha": "",
192
+ "batch_size": "32",
193
+ "batch_sizes": [],
194
+ "device": null,
195
+ "use_cache": null,
196
+ "limit": null,
197
+ "bootstrap_iters": 100000,
198
+ "gen_kwargs": null,
199
+ "random_seed": 0,
200
+ "numpy_seed": 1234,
201
+ "torch_seed": 1234,
202
+ "fewshot_seed": 1234
203
+ },
204
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
205
+ "date": 1767758037.6056852,
206
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-166-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.129.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 256\nOn-line CPU(s) list: 0-255\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7742 64-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 64\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2250.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4491.61\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca sme sev sev_es\nVirtualization: AMD-V\nL1d cache: 4 MiB (128 instances)\nL1i cache: 4 MiB (128 instances)\nL2 cache: 64 MiB (128 instances)\nL3 cache: 512 MiB (32 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-15,128-143\nNUMA node1 CPU(s): 16-31,144-159\nNUMA node2 CPU(s): 32-47,160-175\nNUMA node3 CPU(s): 48-63,176-191\nNUMA node4 CPU(s): 64-79,192-207\nNUMA node5 CPU(s): 80-95,208-223\nNUMA node6 CPU(s): 96-111,224-239\nNUMA node7 CPU(s): 112-127,240-255\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Retbleed: Vulnerable\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Retpolines, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling, PBRSB-eIBRS Not affected\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
207
+ "transformers_version": "4.57.3",
208
+ "lm_eval_version": "0.4.9.1",
209
+ "upper_git_hash": null,
210
+ "tokenizer_pad_token": [
211
+ "<|end_of_text|>",
212
+ "128001"
213
+ ],
214
+ "tokenizer_eos_token": [
215
+ "<|end_of_text|>",
216
+ "128001"
217
+ ],
218
+ "tokenizer_bos_token": [
219
+ "<|begin_of_text|>",
220
+ "128000"
221
+ ],
222
+ "eot_token_id": 128001,
223
+ "max_length": 131072,
224
+ "task_hashes": {
225
+ "hellaswag": "edcc7edd27a555d3f7cbca0641152b2c5e4eb6eb79c5e62d7fe5887f47814323",
226
+ "arc_easy": "dce0d9b0f0cecd55bf2ac264042c5e45487df708d13123af3ae9e67bbbefdeb1",
227
+ "arc_challenge": "09f9ae87a0905d63512cffc4aa91a55e44258fc35160e40fa1eb66fb75473e34"
228
+ },
229
+ "model_source": "hf",
230
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf",
231
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__10000_hf",
232
+ "system_instruction": null,
233
+ "system_instruction_sha": null,
234
+ "fewshot_as_multiturn": false,
235
+ "chat_template": null,
236
+ "chat_template_sha": null,
237
+ "start_time": 77054.817714131,
238
+ "end_time": 77242.890718665,
239
+ "total_evaluation_time_seconds": "188.0730045340024"
240
+ }
10000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_arc_challenge_2026-01-07T03-56-14.560763.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_arc_easy_2026-01-07T03-56-14.560763.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/results_2026-01-07T03-56-07.752198.json ADDED
@@ -0,0 +1,478 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hendrycks_math": {
4
+ "exact_match,none": 0.102,
5
+ "exact_match_stderr,none": 0.004234538823160173,
6
+ "alias": "hendrycks_math"
7
+ },
8
+ "hendrycks_math_algebra": {
9
+ "alias": " - hendrycks_math_algebra",
10
+ "exact_match,none": 0.09856781802864364,
11
+ "exact_match_stderr,none": 0.008655491939014121
12
+ },
13
+ "hendrycks_math_counting_and_prob": {
14
+ "alias": " - hendrycks_math_counting_and_prob",
15
+ "exact_match,none": 0.10970464135021098,
16
+ "exact_match_stderr,none": 0.014369742481257587
17
+ },
18
+ "hendrycks_math_geometry": {
19
+ "alias": " - hendrycks_math_geometry",
20
+ "exact_match,none": 0.10438413361169102,
21
+ "exact_match_stderr,none": 0.013985055252741458
22
+ },
23
+ "hendrycks_math_intermediate_algebra": {
24
+ "alias": " - hendrycks_math_intermediate_algebra",
25
+ "exact_match,none": 0.05647840531561462,
26
+ "exact_match_stderr,none": 0.0076862386783551956
27
+ },
28
+ "hendrycks_math_num_theory": {
29
+ "alias": " - hendrycks_math_num_theory",
30
+ "exact_match,none": 0.07592592592592592,
31
+ "exact_match_stderr,none": 0.011409170195973909
32
+ },
33
+ "hendrycks_math_prealgebra": {
34
+ "alias": " - hendrycks_math_prealgebra",
35
+ "exact_match,none": 0.19173363949483352,
36
+ "exact_match_stderr,none": 0.01334648134700147
37
+ },
38
+ "hendrycks_math_precalc": {
39
+ "alias": " - hendrycks_math_precalc",
40
+ "exact_match,none": 0.05860805860805861,
41
+ "exact_match_stderr,none": 0.010061567725278794
42
+ }
43
+ },
44
+ "groups": {
45
+ "hendrycks_math": {
46
+ "exact_match,none": 0.102,
47
+ "exact_match_stderr,none": 0.004234538823160173,
48
+ "alias": "hendrycks_math"
49
+ }
50
+ },
51
+ "group_subtasks": {
52
+ "hendrycks_math": [
53
+ "hendrycks_math_algebra",
54
+ "hendrycks_math_counting_and_prob",
55
+ "hendrycks_math_geometry",
56
+ "hendrycks_math_intermediate_algebra",
57
+ "hendrycks_math_num_theory",
58
+ "hendrycks_math_prealgebra",
59
+ "hendrycks_math_precalc"
60
+ ]
61
+ },
62
+ "configs": {
63
+ "hendrycks_math_algebra": {
64
+ "task": "hendrycks_math_algebra",
65
+ "tag": [
66
+ "math_word_problems"
67
+ ],
68
+ "dataset_path": "EleutherAI/hendrycks_math",
69
+ "dataset_name": "algebra",
70
+ "training_split": "train",
71
+ "test_split": "test",
72
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
73
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
74
+ "doc_to_target": "{{answer}}",
75
+ "unsafe_code": false,
76
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
77
+ "description": "",
78
+ "target_delimiter": " ",
79
+ "fewshot_delimiter": "\n\n",
80
+ "num_fewshot": 4,
81
+ "metric_list": [
82
+ {
83
+ "metric": "exact_match",
84
+ "aggregation": "mean",
85
+ "higher_is_better": true
86
+ }
87
+ ],
88
+ "output_type": "generate_until",
89
+ "generation_kwargs": {
90
+ "until": [
91
+ "Problem:"
92
+ ],
93
+ "do_sample": false,
94
+ "temperature": 0.0
95
+ },
96
+ "repeats": 1,
97
+ "should_decontaminate": false,
98
+ "metadata": {
99
+ "version": 1.0,
100
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
101
+ }
102
+ },
103
+ "hendrycks_math_counting_and_prob": {
104
+ "task": "hendrycks_math_counting_and_prob",
105
+ "tag": [
106
+ "math_word_problems"
107
+ ],
108
+ "dataset_path": "EleutherAI/hendrycks_math",
109
+ "dataset_name": "counting_and_probability",
110
+ "training_split": "train",
111
+ "test_split": "test",
112
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
113
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
114
+ "doc_to_target": "{{answer}}",
115
+ "unsafe_code": false,
116
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
117
+ "description": "",
118
+ "target_delimiter": " ",
119
+ "fewshot_delimiter": "\n\n",
120
+ "num_fewshot": 4,
121
+ "metric_list": [
122
+ {
123
+ "metric": "exact_match",
124
+ "aggregation": "mean",
125
+ "higher_is_better": true
126
+ }
127
+ ],
128
+ "output_type": "generate_until",
129
+ "generation_kwargs": {
130
+ "until": [
131
+ "Problem:"
132
+ ],
133
+ "do_sample": false,
134
+ "temperature": 0.0
135
+ },
136
+ "repeats": 1,
137
+ "should_decontaminate": false,
138
+ "metadata": {
139
+ "version": 1.0,
140
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
141
+ }
142
+ },
143
+ "hendrycks_math_geometry": {
144
+ "task": "hendrycks_math_geometry",
145
+ "tag": [
146
+ "math_word_problems"
147
+ ],
148
+ "dataset_path": "EleutherAI/hendrycks_math",
149
+ "dataset_name": "geometry",
150
+ "training_split": "train",
151
+ "test_split": "test",
152
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
153
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
154
+ "doc_to_target": "{{answer}}",
155
+ "unsafe_code": false,
156
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
157
+ "description": "",
158
+ "target_delimiter": " ",
159
+ "fewshot_delimiter": "\n\n",
160
+ "num_fewshot": 4,
161
+ "metric_list": [
162
+ {
163
+ "metric": "exact_match",
164
+ "aggregation": "mean",
165
+ "higher_is_better": true
166
+ }
167
+ ],
168
+ "output_type": "generate_until",
169
+ "generation_kwargs": {
170
+ "until": [
171
+ "Problem:"
172
+ ],
173
+ "do_sample": false,
174
+ "temperature": 0.0
175
+ },
176
+ "repeats": 1,
177
+ "should_decontaminate": false,
178
+ "metadata": {
179
+ "version": 1.0,
180
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
181
+ }
182
+ },
183
+ "hendrycks_math_intermediate_algebra": {
184
+ "task": "hendrycks_math_intermediate_algebra",
185
+ "tag": [
186
+ "math_word_problems"
187
+ ],
188
+ "dataset_path": "EleutherAI/hendrycks_math",
189
+ "dataset_name": "intermediate_algebra",
190
+ "training_split": "train",
191
+ "test_split": "test",
192
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
193
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
194
+ "doc_to_target": "{{answer}}",
195
+ "unsafe_code": false,
196
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
197
+ "description": "",
198
+ "target_delimiter": " ",
199
+ "fewshot_delimiter": "\n\n",
200
+ "num_fewshot": 4,
201
+ "metric_list": [
202
+ {
203
+ "metric": "exact_match",
204
+ "aggregation": "mean",
205
+ "higher_is_better": true
206
+ }
207
+ ],
208
+ "output_type": "generate_until",
209
+ "generation_kwargs": {
210
+ "until": [
211
+ "Problem:"
212
+ ],
213
+ "do_sample": false,
214
+ "temperature": 0.0
215
+ },
216
+ "repeats": 1,
217
+ "should_decontaminate": false,
218
+ "metadata": {
219
+ "version": 1.0,
220
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
221
+ }
222
+ },
223
+ "hendrycks_math_num_theory": {
224
+ "task": "hendrycks_math_num_theory",
225
+ "tag": [
226
+ "math_word_problems"
227
+ ],
228
+ "dataset_path": "EleutherAI/hendrycks_math",
229
+ "dataset_name": "number_theory",
230
+ "training_split": "train",
231
+ "test_split": "test",
232
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
233
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
234
+ "doc_to_target": "{{answer}}",
235
+ "unsafe_code": false,
236
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
237
+ "description": "",
238
+ "target_delimiter": " ",
239
+ "fewshot_delimiter": "\n\n",
240
+ "num_fewshot": 4,
241
+ "metric_list": [
242
+ {
243
+ "metric": "exact_match",
244
+ "aggregation": "mean",
245
+ "higher_is_better": true
246
+ }
247
+ ],
248
+ "output_type": "generate_until",
249
+ "generation_kwargs": {
250
+ "until": [
251
+ "Problem:"
252
+ ],
253
+ "do_sample": false,
254
+ "temperature": 0.0
255
+ },
256
+ "repeats": 1,
257
+ "should_decontaminate": false,
258
+ "metadata": {
259
+ "version": 1.0,
260
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
261
+ }
262
+ },
263
+ "hendrycks_math_prealgebra": {
264
+ "task": "hendrycks_math_prealgebra",
265
+ "tag": [
266
+ "math_word_problems"
267
+ ],
268
+ "dataset_path": "EleutherAI/hendrycks_math",
269
+ "dataset_name": "prealgebra",
270
+ "training_split": "train",
271
+ "test_split": "test",
272
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
273
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
274
+ "doc_to_target": "{{answer}}",
275
+ "unsafe_code": false,
276
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
277
+ "description": "",
278
+ "target_delimiter": " ",
279
+ "fewshot_delimiter": "\n\n",
280
+ "num_fewshot": 4,
281
+ "metric_list": [
282
+ {
283
+ "metric": "exact_match",
284
+ "aggregation": "mean",
285
+ "higher_is_better": true
286
+ }
287
+ ],
288
+ "output_type": "generate_until",
289
+ "generation_kwargs": {
290
+ "until": [
291
+ "Problem:"
292
+ ],
293
+ "do_sample": false,
294
+ "temperature": 0.0
295
+ },
296
+ "repeats": 1,
297
+ "should_decontaminate": false,
298
+ "metadata": {
299
+ "version": 1.0,
300
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
301
+ }
302
+ },
303
+ "hendrycks_math_precalc": {
304
+ "task": "hendrycks_math_precalc",
305
+ "tag": [
306
+ "math_word_problems"
307
+ ],
308
+ "dataset_path": "EleutherAI/hendrycks_math",
309
+ "dataset_name": "precalculus",
310
+ "training_split": "train",
311
+ "test_split": "test",
312
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
313
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
314
+ "doc_to_target": "{{answer}}",
315
+ "unsafe_code": false,
316
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
317
+ "description": "",
318
+ "target_delimiter": " ",
319
+ "fewshot_delimiter": "\n\n",
320
+ "num_fewshot": 4,
321
+ "metric_list": [
322
+ {
323
+ "metric": "exact_match",
324
+ "aggregation": "mean",
325
+ "higher_is_better": true
326
+ }
327
+ ],
328
+ "output_type": "generate_until",
329
+ "generation_kwargs": {
330
+ "until": [
331
+ "Problem:"
332
+ ],
333
+ "do_sample": false,
334
+ "temperature": 0.0
335
+ },
336
+ "repeats": 1,
337
+ "should_decontaminate": false,
338
+ "metadata": {
339
+ "version": 1.0,
340
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
341
+ }
342
+ }
343
+ },
344
+ "versions": {
345
+ "hendrycks_math": 1.0,
346
+ "hendrycks_math_algebra": 1.0,
347
+ "hendrycks_math_counting_and_prob": 1.0,
348
+ "hendrycks_math_geometry": 1.0,
349
+ "hendrycks_math_intermediate_algebra": 1.0,
350
+ "hendrycks_math_num_theory": 1.0,
351
+ "hendrycks_math_prealgebra": 1.0,
352
+ "hendrycks_math_precalc": 1.0
353
+ },
354
+ "n-shot": {
355
+ "hendrycks_math_algebra": 4,
356
+ "hendrycks_math_counting_and_prob": 4,
357
+ "hendrycks_math_geometry": 4,
358
+ "hendrycks_math_intermediate_algebra": 4,
359
+ "hendrycks_math_num_theory": 4,
360
+ "hendrycks_math_prealgebra": 4,
361
+ "hendrycks_math_precalc": 4
362
+ },
363
+ "higher_is_better": {
364
+ "hendrycks_math": {
365
+ "exact_match": true
366
+ },
367
+ "hendrycks_math_algebra": {
368
+ "exact_match": true
369
+ },
370
+ "hendrycks_math_counting_and_prob": {
371
+ "exact_match": true
372
+ },
373
+ "hendrycks_math_geometry": {
374
+ "exact_match": true
375
+ },
376
+ "hendrycks_math_intermediate_algebra": {
377
+ "exact_match": true
378
+ },
379
+ "hendrycks_math_num_theory": {
380
+ "exact_match": true
381
+ },
382
+ "hendrycks_math_prealgebra": {
383
+ "exact_match": true
384
+ },
385
+ "hendrycks_math_precalc": {
386
+ "exact_match": true
387
+ }
388
+ },
389
+ "n-samples": {
390
+ "hendrycks_math_algebra": {
391
+ "original": 1187,
392
+ "effective": 1187
393
+ },
394
+ "hendrycks_math_counting_and_prob": {
395
+ "original": 474,
396
+ "effective": 474
397
+ },
398
+ "hendrycks_math_geometry": {
399
+ "original": 479,
400
+ "effective": 479
401
+ },
402
+ "hendrycks_math_intermediate_algebra": {
403
+ "original": 903,
404
+ "effective": 903
405
+ },
406
+ "hendrycks_math_num_theory": {
407
+ "original": 540,
408
+ "effective": 540
409
+ },
410
+ "hendrycks_math_prealgebra": {
411
+ "original": 871,
412
+ "effective": 871
413
+ },
414
+ "hendrycks_math_precalc": {
415
+ "original": 546,
416
+ "effective": 546
417
+ }
418
+ },
419
+ "config": {
420
+ "model": "hf",
421
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/10000_hf,trust_remote_code=True",
422
+ "model_num_parameters": 1235814400,
423
+ "model_dtype": "torch.bfloat16",
424
+ "model_revision": "main",
425
+ "model_sha": "",
426
+ "batch_size": "64",
427
+ "batch_sizes": [],
428
+ "device": null,
429
+ "use_cache": null,
430
+ "limit": null,
431
+ "bootstrap_iters": 100000,
432
+ "gen_kwargs": null,
433
+ "random_seed": 0,
434
+ "numpy_seed": 1234,
435
+ "torch_seed": 1234,
436
+ "fewshot_seed": 1234
437
+ },
438
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
439
+ "date": 1767757751.052034,
440
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-166-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.129.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 256\nOn-line CPU(s) list: 0-255\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7742 64-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 64\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2250.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4491.61\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca sme sev sev_es\nVirtualization: AMD-V\nL1d cache: 4 MiB (128 instances)\nL1i cache: 4 MiB (128 instances)\nL2 cache: 64 MiB (128 instances)\nL3 cache: 512 MiB (32 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-15,128-143\nNUMA node1 CPU(s): 16-31,144-159\nNUMA node2 CPU(s): 32-47,160-175\nNUMA node3 CPU(s): 48-63,176-191\nNUMA node4 CPU(s): 64-79,192-207\nNUMA node5 CPU(s): 80-95,208-223\nNUMA node6 CPU(s): 96-111,224-239\nNUMA node7 CPU(s): 112-127,240-255\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Retbleed: Vulnerable\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Retpolines, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling, PBRSB-eIBRS Not affected\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
441
+ "transformers_version": "4.57.3",
442
+ "lm_eval_version": "0.4.9.1",
443
+ "upper_git_hash": null,
444
+ "tokenizer_pad_token": [
445
+ "<|end_of_text|>",
446
+ "128001"
447
+ ],
448
+ "tokenizer_eos_token": [
449
+ "<|end_of_text|>",
450
+ "128001"
451
+ ],
452
+ "tokenizer_bos_token": [
453
+ "<|begin_of_text|>",
454
+ "128000"
455
+ ],
456
+ "eot_token_id": 128001,
457
+ "max_length": 131072,
458
+ "task_hashes": {
459
+ "hendrycks_math_algebra": "7b50860e4db85f9d690d9679b543412837815a3b9854f0d0d166cfefc428cc23",
460
+ "hendrycks_math_counting_and_prob": "826f9744663d8e2e04e0050aa87b6edc5640c41ca1aa53fee8ef747f7c93b100",
461
+ "hendrycks_math_geometry": "4b67d6f45ba7f89a0c8a0a01fc9b1885bf942346a45b6a4fd03afd3d0e23ab6e",
462
+ "hendrycks_math_intermediate_algebra": "d7bc95882c4f58ef0b7252e6026e98b8bc77bbfbe36de6da6d97127e9744141e",
463
+ "hendrycks_math_num_theory": "ff82fd9f09e0a7c370270b3e7778773f7f1dec4aa40564e1a1318417af819030",
464
+ "hendrycks_math_prealgebra": "e4296905fe5f21ca452a0596371350ffedec630103ac7cce4b76cb806c55be05",
465
+ "hendrycks_math_precalc": "065216e339084dad7a3260b23a5151c8bf63771c7f0d23e88b6f3df4a9f0175a"
466
+ },
467
+ "model_source": "hf",
468
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf",
469
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__10000_hf",
470
+ "system_instruction": null,
471
+ "system_instruction_sha": null,
472
+ "fewshot_as_multiturn": false,
473
+ "chat_template": null,
474
+ "chat_template_sha": null,
475
+ "start_time": 76759.655376049,
476
+ "end_time": 77236.088412527,
477
+ "total_evaluation_time_seconds": "476.43303647798894"
478
+ }
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_algebra_2026-01-07T03-56-07.752198.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_counting_and_prob_2026-01-07T03-56-07.752198.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_geometry_2026-01-07T03-56-07.752198.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_intermediate_algebra_2026-01-07T03-56-07.752198.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_num_theory_2026-01-07T03-56-07.752198.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_prealgebra_2026-01-07T03-56-07.752198.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_hendrycks_math_precalc_2026-01-07T03-56-07.752198.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/results_2026-01-07T07-13-00.293068.json ADDED
@@ -0,0 +1,290 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "gsm8k": {
4
+ "alias": "gsm8k",
5
+ "exact_match,strict-match": 0.21531463229719486,
6
+ "exact_match_stderr,strict-match": 0.011322096294579675,
7
+ "exact_match,flexible-extract": 0.21455648218347234,
8
+ "exact_match_stderr,flexible-extract": 0.011307604104052883
9
+ },
10
+ "gsm8k_cot": {
11
+ "alias": "gsm8k_cot",
12
+ "exact_match,strict-match": 0.20090978013646701,
13
+ "exact_match_stderr,strict-match": 0.011036738221872374,
14
+ "exact_match,flexible-extract": 0.22971948445792267,
15
+ "exact_match_stderr,flexible-extract": 0.011586857544997501
16
+ }
17
+ },
18
+ "group_subtasks": {
19
+ "gsm8k": [],
20
+ "gsm8k_cot": []
21
+ },
22
+ "configs": {
23
+ "gsm8k": {
24
+ "task": "gsm8k",
25
+ "tag": [
26
+ "math_word_problems"
27
+ ],
28
+ "dataset_path": "gsm8k",
29
+ "dataset_name": "main",
30
+ "training_split": "train",
31
+ "test_split": "test",
32
+ "fewshot_split": "train",
33
+ "doc_to_text": "Question: {{question}}\nAnswer:",
34
+ "doc_to_target": "{{answer}}",
35
+ "unsafe_code": false,
36
+ "description": "",
37
+ "target_delimiter": " ",
38
+ "fewshot_delimiter": "\n\n",
39
+ "num_fewshot": 8,
40
+ "metric_list": [
41
+ {
42
+ "metric": "exact_match",
43
+ "aggregation": "mean",
44
+ "higher_is_better": true,
45
+ "ignore_case": true,
46
+ "ignore_punctuation": false,
47
+ "regexes_to_ignore": [
48
+ ",",
49
+ "\\$",
50
+ "(?s).*#### ",
51
+ "\\.$"
52
+ ]
53
+ }
54
+ ],
55
+ "output_type": "generate_until",
56
+ "generation_kwargs": {
57
+ "until": [
58
+ "Question:",
59
+ "</s>",
60
+ "<|im_end|>"
61
+ ],
62
+ "do_sample": false,
63
+ "temperature": 0.0
64
+ },
65
+ "repeats": 1,
66
+ "filter_list": [
67
+ {
68
+ "name": "strict-match",
69
+ "filter": [
70
+ {
71
+ "function": "regex",
72
+ "regex_pattern": "#### (\\-?[0-9\\.\\,]+)"
73
+ },
74
+ {
75
+ "function": "take_first"
76
+ }
77
+ ]
78
+ },
79
+ {
80
+ "name": "flexible-extract",
81
+ "filter": [
82
+ {
83
+ "function": "regex",
84
+ "group_select": -1,
85
+ "regex_pattern": "(-?[$0-9.,]{2,})|(-?[0-9]+)"
86
+ },
87
+ {
88
+ "function": "take_first"
89
+ }
90
+ ]
91
+ }
92
+ ],
93
+ "should_decontaminate": false,
94
+ "metadata": {
95
+ "version": 3.0,
96
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
97
+ }
98
+ },
99
+ "gsm8k_cot": {
100
+ "task": "gsm8k_cot",
101
+ "tag": [
102
+ "chain_of_thought"
103
+ ],
104
+ "dataset_path": "gsm8k",
105
+ "dataset_name": "main",
106
+ "test_split": "test",
107
+ "doc_to_text": "Q: {{question}}\nA:",
108
+ "doc_to_target": "{{answer.split('####')[-1].strip() if answer is defined else target}}",
109
+ "unsafe_code": false,
110
+ "description": "",
111
+ "target_delimiter": " ",
112
+ "fewshot_delimiter": "\n\n",
113
+ "fewshot_config": {
114
+ "sampler": "first_n",
115
+ "samples": [
116
+ {
117
+ "question": "There are 15 trees in the grove. Grove workers will plant trees in the grove today. After they are done, there will be 21 trees. How many trees did the grove workers plant today?",
118
+ "target": "There are 15 trees originally. Then there were 21 trees after some more were planted. So there must have been 21 - 15 = 6. The answer is 6."
119
+ },
120
+ {
121
+ "question": "If there are 3 cars in the parking lot and 2 more cars arrive, how many cars are in the parking lot?",
122
+ "target": "There are originally 3 cars. 2 more cars arrive. 3 + 2 = 5. The answer is 5."
123
+ },
124
+ {
125
+ "question": "Leah had 32 chocolates and her sister had 42. If they ate 35, how many pieces do they have left in total?",
126
+ "target": "Originally, Leah had 32 chocolates. Her sister had 42. So in total they had 32 + 42 = 74. After eating 35, they had 74 - 35 = 39. The answer is 39."
127
+ },
128
+ {
129
+ "question": "Jason had 20 lollipops. He gave Denny some lollipops. Now Jason has 12 lollipops. How many lollipops did Jason give to Denny?",
130
+ "target": "Jason started with 20 lollipops. Then he had 12 after giving some to Denny. So he gave Denny 20 - 12 = 8. The answer is 8."
131
+ },
132
+ {
133
+ "question": "Shawn has five toys. For Christmas, he got two toys each from his mom and dad. How many toys does he have now?",
134
+ "target": "Shawn started with 5 toys. If he got 2 toys each from his mom and dad, then that is 4 more toys. 5 + 4 = 9. The answer is 9."
135
+ },
136
+ {
137
+ "question": "There were nine computers in the server room. Five more computers were installed each day, from monday to thursday. How many computers are now in the server room?",
138
+ "target": "There were originally 9 computers. For each of 4 days, 5 more computers were added. So 5 * 4 = 20 computers were added. 9 + 20 is 29. The answer is 29."
139
+ },
140
+ {
141
+ "question": "Michael had 58 golf balls. On tuesday, he lost 23 golf balls. On wednesday, he lost 2 more. How many golf balls did he have at the end of wednesday?",
142
+ "target": "Michael started with 58 golf balls. After losing 23 on tuesday, he had 58 - 23 = 35. After losing 2 more, he had 35 - 2 = 33 golf balls. The answer is 33."
143
+ },
144
+ {
145
+ "question": "Olivia has $23. She bought five bagels for $3 each. How much money does she have left?",
146
+ "target": "Olivia had 23 dollars. 5 bagels for 3 dollars each will be 5 x 3 = 15 dollars. So she has 23 - 15 dollars left. 23 - 15 is 8. The answer is 8."
147
+ }
148
+ ]
149
+ },
150
+ "num_fewshot": 8,
151
+ "metric_list": [
152
+ {
153
+ "aggregation": "mean",
154
+ "higher_is_better": true,
155
+ "ignore_case": true,
156
+ "ignore_punctuation": false,
157
+ "metric": "exact_match",
158
+ "regexes_to_ignore": [
159
+ ",",
160
+ "\\$",
161
+ "(?s).*#### ",
162
+ "\\.$"
163
+ ]
164
+ }
165
+ ],
166
+ "output_type": "generate_until",
167
+ "generation_kwargs": {
168
+ "do_sample": false,
169
+ "until": [
170
+ "Q:",
171
+ "</s>",
172
+ "<|im_end|>"
173
+ ]
174
+ },
175
+ "repeats": 1,
176
+ "filter_list": [
177
+ {
178
+ "filter": [
179
+ {
180
+ "function": "regex",
181
+ "regex_pattern": "The answer is (\\-?[0-9\\.\\,]+)."
182
+ },
183
+ {
184
+ "function": "take_first"
185
+ }
186
+ ],
187
+ "name": "strict-match"
188
+ },
189
+ {
190
+ "filter": [
191
+ {
192
+ "function": "regex",
193
+ "group_select": -1,
194
+ "regex_pattern": "(-?[$0-9.,]{2,})|(-?[0-9]+)"
195
+ },
196
+ {
197
+ "function": "take_first"
198
+ }
199
+ ],
200
+ "name": "flexible-extract"
201
+ }
202
+ ],
203
+ "should_decontaminate": false,
204
+ "metadata": {
205
+ "version": 3.0,
206
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf"
207
+ }
208
+ }
209
+ },
210
+ "versions": {
211
+ "gsm8k": 3.0,
212
+ "gsm8k_cot": 3.0
213
+ },
214
+ "n-shot": {
215
+ "gsm8k": 8,
216
+ "gsm8k_cot": 8
217
+ },
218
+ "higher_is_better": {
219
+ "gsm8k": {
220
+ "exact_match": true
221
+ },
222
+ "gsm8k_cot": {
223
+ "exact_match": true
224
+ }
225
+ },
226
+ "n-samples": {
227
+ "gsm8k_cot": {
228
+ "original": 1319,
229
+ "effective": 1319
230
+ },
231
+ "gsm8k": {
232
+ "original": 1319,
233
+ "effective": 1319
234
+ }
235
+ },
236
+ "config": {
237
+ "model": "hf",
238
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/10000_hf,trust_remote_code=True",
239
+ "model_num_parameters": 1235814400,
240
+ "model_dtype": "torch.bfloat16",
241
+ "model_revision": "main",
242
+ "model_sha": "",
243
+ "batch_size": "64",
244
+ "batch_sizes": [],
245
+ "device": null,
246
+ "use_cache": null,
247
+ "limit": null,
248
+ "bootstrap_iters": 100000,
249
+ "gen_kwargs": null,
250
+ "random_seed": 0,
251
+ "numpy_seed": 1234,
252
+ "torch_seed": 1234,
253
+ "fewshot_seed": 1234
254
+ },
255
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
256
+ "date": 1767769329.570305,
257
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-77-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.129.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 48 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 96\nOn-line CPU(s) list: 0-95\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7352 24-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 24\nSocket(s): 2\nStepping: 0\nBogoMIPS: 4591.57\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca\nVirtualization: AMD-V\nL1d cache: 1.5 MiB (48 instances)\nL1i cache: 1.5 MiB (48 instances)\nL2 cache: 24 MiB (48 instances)\nL3 cache: 256 MiB (16 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-5,48-53\nNUMA node1 CPU(s): 6-11,54-59\nNUMA node2 CPU(s): 12-17,60-65\nNUMA node3 CPU(s): 18-23,66-71\nNUMA node4 CPU(s): 24-29,72-77\nNUMA node5 CPU(s): 30-35,78-83\nNUMA node6 CPU(s): 36-41,84-89\nNUMA node7 CPU(s): 42-47,90-95\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Full AMD retpoline, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
258
+ "transformers_version": "4.57.3",
259
+ "lm_eval_version": "0.4.9.1",
260
+ "upper_git_hash": null,
261
+ "tokenizer_pad_token": [
262
+ "<|end_of_text|>",
263
+ "128001"
264
+ ],
265
+ "tokenizer_eos_token": [
266
+ "<|end_of_text|>",
267
+ "128001"
268
+ ],
269
+ "tokenizer_bos_token": [
270
+ "<|begin_of_text|>",
271
+ "128000"
272
+ ],
273
+ "eot_token_id": 128001,
274
+ "max_length": 131072,
275
+ "task_hashes": {
276
+ "gsm8k_cot": "fc360963b39ee52c26a82795124f9ad7da4d6a8fecf1b77e2502823b1669b3d0",
277
+ "gsm8k": "5edaa24ff4f3d939c3e1c5fd65a53cead84d4a52171818c453ec47099bd2a422"
278
+ },
279
+ "model_source": "hf",
280
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/10000_hf",
281
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__10000_hf",
282
+ "system_instruction": null,
283
+ "system_instruction_sha": null,
284
+ "fewshot_as_multiturn": false,
285
+ "chat_template": null,
286
+ "chat_template_sha": null,
287
+ "start_time": 618876.622582658,
288
+ "end_time": 619576.918757243,
289
+ "total_evaluation_time_seconds": "700.2961745850043"
290
+ }
10000_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__10000_hf/samples_gsm8k_cot_2026-01-07T07-13-00.293068.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
10000_hf/special_tokens_map.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|end_of_text|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ }
16
+ }
10000_hf/tokenizer_config.json ADDED
@@ -0,0 +1,2062 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "128000": {
4
+ "content": "<|begin_of_text|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "128001": {
12
+ "content": "<|end_of_text|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "128002": {
20
+ "content": "<|reserved_special_token_0|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "128003": {
28
+ "content": "<|reserved_special_token_1|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128004": {
36
+ "content": "<|finetune_right_pad_id|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "128005": {
44
+ "content": "<|reserved_special_token_2|>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "128006": {
52
+ "content": "<|start_header_id|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "128007": {
60
+ "content": "<|end_header_id|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "128008": {
68
+ "content": "<|eom_id|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "128009": {
76
+ "content": "<|eot_id|>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "128010": {
84
+ "content": "<|python_tag|>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "128011": {
92
+ "content": "<|reserved_special_token_3|>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "128012": {
100
+ "content": "<|reserved_special_token_4|>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "128013": {
108
+ "content": "<|reserved_special_token_5|>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "128014": {
116
+ "content": "<|reserved_special_token_6|>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "128015": {
124
+ "content": "<|reserved_special_token_7|>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "128016": {
132
+ "content": "<|reserved_special_token_8|>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "128017": {
140
+ "content": "<|reserved_special_token_9|>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "128018": {
148
+ "content": "<|reserved_special_token_10|>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "128019": {
156
+ "content": "<|reserved_special_token_11|>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "128020": {
164
+ "content": "<|reserved_special_token_12|>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "128021": {
172
+ "content": "<|reserved_special_token_13|>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "128022": {
180
+ "content": "<|reserved_special_token_14|>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "128023": {
188
+ "content": "<|reserved_special_token_15|>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "128024": {
196
+ "content": "<|reserved_special_token_16|>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "128025": {
204
+ "content": "<|reserved_special_token_17|>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "128026": {
212
+ "content": "<|reserved_special_token_18|>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "128027": {
220
+ "content": "<|reserved_special_token_19|>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "128028": {
228
+ "content": "<|reserved_special_token_20|>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "128029": {
236
+ "content": "<|reserved_special_token_21|>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "128030": {
244
+ "content": "<|reserved_special_token_22|>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "128031": {
252
+ "content": "<|reserved_special_token_23|>",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "128032": {
260
+ "content": "<|reserved_special_token_24|>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "128033": {
268
+ "content": "<|reserved_special_token_25|>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "128034": {
276
+ "content": "<|reserved_special_token_26|>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "128035": {
284
+ "content": "<|reserved_special_token_27|>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "128036": {
292
+ "content": "<|reserved_special_token_28|>",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "128037": {
300
+ "content": "<|reserved_special_token_29|>",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "128038": {
308
+ "content": "<|reserved_special_token_30|>",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "128039": {
316
+ "content": "<|reserved_special_token_31|>",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "128040": {
324
+ "content": "<|reserved_special_token_32|>",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "128041": {
332
+ "content": "<|reserved_special_token_33|>",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "128042": {
340
+ "content": "<|reserved_special_token_34|>",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "128043": {
348
+ "content": "<|reserved_special_token_35|>",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "128044": {
356
+ "content": "<|reserved_special_token_36|>",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "128045": {
364
+ "content": "<|reserved_special_token_37|>",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "128046": {
372
+ "content": "<|reserved_special_token_38|>",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "128047": {
380
+ "content": "<|reserved_special_token_39|>",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "128048": {
388
+ "content": "<|reserved_special_token_40|>",
389
+ "lstrip": false,
390
+ "normalized": false,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "128049": {
396
+ "content": "<|reserved_special_token_41|>",
397
+ "lstrip": false,
398
+ "normalized": false,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "128050": {
404
+ "content": "<|reserved_special_token_42|>",
405
+ "lstrip": false,
406
+ "normalized": false,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "128051": {
412
+ "content": "<|reserved_special_token_43|>",
413
+ "lstrip": false,
414
+ "normalized": false,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "128052": {
420
+ "content": "<|reserved_special_token_44|>",
421
+ "lstrip": false,
422
+ "normalized": false,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "128053": {
428
+ "content": "<|reserved_special_token_45|>",
429
+ "lstrip": false,
430
+ "normalized": false,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "128054": {
436
+ "content": "<|reserved_special_token_46|>",
437
+ "lstrip": false,
438
+ "normalized": false,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": true
442
+ },
443
+ "128055": {
444
+ "content": "<|reserved_special_token_47|>",
445
+ "lstrip": false,
446
+ "normalized": false,
447
+ "rstrip": false,
448
+ "single_word": false,
449
+ "special": true
450
+ },
451
+ "128056": {
452
+ "content": "<|reserved_special_token_48|>",
453
+ "lstrip": false,
454
+ "normalized": false,
455
+ "rstrip": false,
456
+ "single_word": false,
457
+ "special": true
458
+ },
459
+ "128057": {
460
+ "content": "<|reserved_special_token_49|>",
461
+ "lstrip": false,
462
+ "normalized": false,
463
+ "rstrip": false,
464
+ "single_word": false,
465
+ "special": true
466
+ },
467
+ "128058": {
468
+ "content": "<|reserved_special_token_50|>",
469
+ "lstrip": false,
470
+ "normalized": false,
471
+ "rstrip": false,
472
+ "single_word": false,
473
+ "special": true
474
+ },
475
+ "128059": {
476
+ "content": "<|reserved_special_token_51|>",
477
+ "lstrip": false,
478
+ "normalized": false,
479
+ "rstrip": false,
480
+ "single_word": false,
481
+ "special": true
482
+ },
483
+ "128060": {
484
+ "content": "<|reserved_special_token_52|>",
485
+ "lstrip": false,
486
+ "normalized": false,
487
+ "rstrip": false,
488
+ "single_word": false,
489
+ "special": true
490
+ },
491
+ "128061": {
492
+ "content": "<|reserved_special_token_53|>",
493
+ "lstrip": false,
494
+ "normalized": false,
495
+ "rstrip": false,
496
+ "single_word": false,
497
+ "special": true
498
+ },
499
+ "128062": {
500
+ "content": "<|reserved_special_token_54|>",
501
+ "lstrip": false,
502
+ "normalized": false,
503
+ "rstrip": false,
504
+ "single_word": false,
505
+ "special": true
506
+ },
507
+ "128063": {
508
+ "content": "<|reserved_special_token_55|>",
509
+ "lstrip": false,
510
+ "normalized": false,
511
+ "rstrip": false,
512
+ "single_word": false,
513
+ "special": true
514
+ },
515
+ "128064": {
516
+ "content": "<|reserved_special_token_56|>",
517
+ "lstrip": false,
518
+ "normalized": false,
519
+ "rstrip": false,
520
+ "single_word": false,
521
+ "special": true
522
+ },
523
+ "128065": {
524
+ "content": "<|reserved_special_token_57|>",
525
+ "lstrip": false,
526
+ "normalized": false,
527
+ "rstrip": false,
528
+ "single_word": false,
529
+ "special": true
530
+ },
531
+ "128066": {
532
+ "content": "<|reserved_special_token_58|>",
533
+ "lstrip": false,
534
+ "normalized": false,
535
+ "rstrip": false,
536
+ "single_word": false,
537
+ "special": true
538
+ },
539
+ "128067": {
540
+ "content": "<|reserved_special_token_59|>",
541
+ "lstrip": false,
542
+ "normalized": false,
543
+ "rstrip": false,
544
+ "single_word": false,
545
+ "special": true
546
+ },
547
+ "128068": {
548
+ "content": "<|reserved_special_token_60|>",
549
+ "lstrip": false,
550
+ "normalized": false,
551
+ "rstrip": false,
552
+ "single_word": false,
553
+ "special": true
554
+ },
555
+ "128069": {
556
+ "content": "<|reserved_special_token_61|>",
557
+ "lstrip": false,
558
+ "normalized": false,
559
+ "rstrip": false,
560
+ "single_word": false,
561
+ "special": true
562
+ },
563
+ "128070": {
564
+ "content": "<|reserved_special_token_62|>",
565
+ "lstrip": false,
566
+ "normalized": false,
567
+ "rstrip": false,
568
+ "single_word": false,
569
+ "special": true
570
+ },
571
+ "128071": {
572
+ "content": "<|reserved_special_token_63|>",
573
+ "lstrip": false,
574
+ "normalized": false,
575
+ "rstrip": false,
576
+ "single_word": false,
577
+ "special": true
578
+ },
579
+ "128072": {
580
+ "content": "<|reserved_special_token_64|>",
581
+ "lstrip": false,
582
+ "normalized": false,
583
+ "rstrip": false,
584
+ "single_word": false,
585
+ "special": true
586
+ },
587
+ "128073": {
588
+ "content": "<|reserved_special_token_65|>",
589
+ "lstrip": false,
590
+ "normalized": false,
591
+ "rstrip": false,
592
+ "single_word": false,
593
+ "special": true
594
+ },
595
+ "128074": {
596
+ "content": "<|reserved_special_token_66|>",
597
+ "lstrip": false,
598
+ "normalized": false,
599
+ "rstrip": false,
600
+ "single_word": false,
601
+ "special": true
602
+ },
603
+ "128075": {
604
+ "content": "<|reserved_special_token_67|>",
605
+ "lstrip": false,
606
+ "normalized": false,
607
+ "rstrip": false,
608
+ "single_word": false,
609
+ "special": true
610
+ },
611
+ "128076": {
612
+ "content": "<|reserved_special_token_68|>",
613
+ "lstrip": false,
614
+ "normalized": false,
615
+ "rstrip": false,
616
+ "single_word": false,
617
+ "special": true
618
+ },
619
+ "128077": {
620
+ "content": "<|reserved_special_token_69|>",
621
+ "lstrip": false,
622
+ "normalized": false,
623
+ "rstrip": false,
624
+ "single_word": false,
625
+ "special": true
626
+ },
627
+ "128078": {
628
+ "content": "<|reserved_special_token_70|>",
629
+ "lstrip": false,
630
+ "normalized": false,
631
+ "rstrip": false,
632
+ "single_word": false,
633
+ "special": true
634
+ },
635
+ "128079": {
636
+ "content": "<|reserved_special_token_71|>",
637
+ "lstrip": false,
638
+ "normalized": false,
639
+ "rstrip": false,
640
+ "single_word": false,
641
+ "special": true
642
+ },
643
+ "128080": {
644
+ "content": "<|reserved_special_token_72|>",
645
+ "lstrip": false,
646
+ "normalized": false,
647
+ "rstrip": false,
648
+ "single_word": false,
649
+ "special": true
650
+ },
651
+ "128081": {
652
+ "content": "<|reserved_special_token_73|>",
653
+ "lstrip": false,
654
+ "normalized": false,
655
+ "rstrip": false,
656
+ "single_word": false,
657
+ "special": true
658
+ },
659
+ "128082": {
660
+ "content": "<|reserved_special_token_74|>",
661
+ "lstrip": false,
662
+ "normalized": false,
663
+ "rstrip": false,
664
+ "single_word": false,
665
+ "special": true
666
+ },
667
+ "128083": {
668
+ "content": "<|reserved_special_token_75|>",
669
+ "lstrip": false,
670
+ "normalized": false,
671
+ "rstrip": false,
672
+ "single_word": false,
673
+ "special": true
674
+ },
675
+ "128084": {
676
+ "content": "<|reserved_special_token_76|>",
677
+ "lstrip": false,
678
+ "normalized": false,
679
+ "rstrip": false,
680
+ "single_word": false,
681
+ "special": true
682
+ },
683
+ "128085": {
684
+ "content": "<|reserved_special_token_77|>",
685
+ "lstrip": false,
686
+ "normalized": false,
687
+ "rstrip": false,
688
+ "single_word": false,
689
+ "special": true
690
+ },
691
+ "128086": {
692
+ "content": "<|reserved_special_token_78|>",
693
+ "lstrip": false,
694
+ "normalized": false,
695
+ "rstrip": false,
696
+ "single_word": false,
697
+ "special": true
698
+ },
699
+ "128087": {
700
+ "content": "<|reserved_special_token_79|>",
701
+ "lstrip": false,
702
+ "normalized": false,
703
+ "rstrip": false,
704
+ "single_word": false,
705
+ "special": true
706
+ },
707
+ "128088": {
708
+ "content": "<|reserved_special_token_80|>",
709
+ "lstrip": false,
710
+ "normalized": false,
711
+ "rstrip": false,
712
+ "single_word": false,
713
+ "special": true
714
+ },
715
+ "128089": {
716
+ "content": "<|reserved_special_token_81|>",
717
+ "lstrip": false,
718
+ "normalized": false,
719
+ "rstrip": false,
720
+ "single_word": false,
721
+ "special": true
722
+ },
723
+ "128090": {
724
+ "content": "<|reserved_special_token_82|>",
725
+ "lstrip": false,
726
+ "normalized": false,
727
+ "rstrip": false,
728
+ "single_word": false,
729
+ "special": true
730
+ },
731
+ "128091": {
732
+ "content": "<|reserved_special_token_83|>",
733
+ "lstrip": false,
734
+ "normalized": false,
735
+ "rstrip": false,
736
+ "single_word": false,
737
+ "special": true
738
+ },
739
+ "128092": {
740
+ "content": "<|reserved_special_token_84|>",
741
+ "lstrip": false,
742
+ "normalized": false,
743
+ "rstrip": false,
744
+ "single_word": false,
745
+ "special": true
746
+ },
747
+ "128093": {
748
+ "content": "<|reserved_special_token_85|>",
749
+ "lstrip": false,
750
+ "normalized": false,
751
+ "rstrip": false,
752
+ "single_word": false,
753
+ "special": true
754
+ },
755
+ "128094": {
756
+ "content": "<|reserved_special_token_86|>",
757
+ "lstrip": false,
758
+ "normalized": false,
759
+ "rstrip": false,
760
+ "single_word": false,
761
+ "special": true
762
+ },
763
+ "128095": {
764
+ "content": "<|reserved_special_token_87|>",
765
+ "lstrip": false,
766
+ "normalized": false,
767
+ "rstrip": false,
768
+ "single_word": false,
769
+ "special": true
770
+ },
771
+ "128096": {
772
+ "content": "<|reserved_special_token_88|>",
773
+ "lstrip": false,
774
+ "normalized": false,
775
+ "rstrip": false,
776
+ "single_word": false,
777
+ "special": true
778
+ },
779
+ "128097": {
780
+ "content": "<|reserved_special_token_89|>",
781
+ "lstrip": false,
782
+ "normalized": false,
783
+ "rstrip": false,
784
+ "single_word": false,
785
+ "special": true
786
+ },
787
+ "128098": {
788
+ "content": "<|reserved_special_token_90|>",
789
+ "lstrip": false,
790
+ "normalized": false,
791
+ "rstrip": false,
792
+ "single_word": false,
793
+ "special": true
794
+ },
795
+ "128099": {
796
+ "content": "<|reserved_special_token_91|>",
797
+ "lstrip": false,
798
+ "normalized": false,
799
+ "rstrip": false,
800
+ "single_word": false,
801
+ "special": true
802
+ },
803
+ "128100": {
804
+ "content": "<|reserved_special_token_92|>",
805
+ "lstrip": false,
806
+ "normalized": false,
807
+ "rstrip": false,
808
+ "single_word": false,
809
+ "special": true
810
+ },
811
+ "128101": {
812
+ "content": "<|reserved_special_token_93|>",
813
+ "lstrip": false,
814
+ "normalized": false,
815
+ "rstrip": false,
816
+ "single_word": false,
817
+ "special": true
818
+ },
819
+ "128102": {
820
+ "content": "<|reserved_special_token_94|>",
821
+ "lstrip": false,
822
+ "normalized": false,
823
+ "rstrip": false,
824
+ "single_word": false,
825
+ "special": true
826
+ },
827
+ "128103": {
828
+ "content": "<|reserved_special_token_95|>",
829
+ "lstrip": false,
830
+ "normalized": false,
831
+ "rstrip": false,
832
+ "single_word": false,
833
+ "special": true
834
+ },
835
+ "128104": {
836
+ "content": "<|reserved_special_token_96|>",
837
+ "lstrip": false,
838
+ "normalized": false,
839
+ "rstrip": false,
840
+ "single_word": false,
841
+ "special": true
842
+ },
843
+ "128105": {
844
+ "content": "<|reserved_special_token_97|>",
845
+ "lstrip": false,
846
+ "normalized": false,
847
+ "rstrip": false,
848
+ "single_word": false,
849
+ "special": true
850
+ },
851
+ "128106": {
852
+ "content": "<|reserved_special_token_98|>",
853
+ "lstrip": false,
854
+ "normalized": false,
855
+ "rstrip": false,
856
+ "single_word": false,
857
+ "special": true
858
+ },
859
+ "128107": {
860
+ "content": "<|reserved_special_token_99|>",
861
+ "lstrip": false,
862
+ "normalized": false,
863
+ "rstrip": false,
864
+ "single_word": false,
865
+ "special": true
866
+ },
867
+ "128108": {
868
+ "content": "<|reserved_special_token_100|>",
869
+ "lstrip": false,
870
+ "normalized": false,
871
+ "rstrip": false,
872
+ "single_word": false,
873
+ "special": true
874
+ },
875
+ "128109": {
876
+ "content": "<|reserved_special_token_101|>",
877
+ "lstrip": false,
878
+ "normalized": false,
879
+ "rstrip": false,
880
+ "single_word": false,
881
+ "special": true
882
+ },
883
+ "128110": {
884
+ "content": "<|reserved_special_token_102|>",
885
+ "lstrip": false,
886
+ "normalized": false,
887
+ "rstrip": false,
888
+ "single_word": false,
889
+ "special": true
890
+ },
891
+ "128111": {
892
+ "content": "<|reserved_special_token_103|>",
893
+ "lstrip": false,
894
+ "normalized": false,
895
+ "rstrip": false,
896
+ "single_word": false,
897
+ "special": true
898
+ },
899
+ "128112": {
900
+ "content": "<|reserved_special_token_104|>",
901
+ "lstrip": false,
902
+ "normalized": false,
903
+ "rstrip": false,
904
+ "single_word": false,
905
+ "special": true
906
+ },
907
+ "128113": {
908
+ "content": "<|reserved_special_token_105|>",
909
+ "lstrip": false,
910
+ "normalized": false,
911
+ "rstrip": false,
912
+ "single_word": false,
913
+ "special": true
914
+ },
915
+ "128114": {
916
+ "content": "<|reserved_special_token_106|>",
917
+ "lstrip": false,
918
+ "normalized": false,
919
+ "rstrip": false,
920
+ "single_word": false,
921
+ "special": true
922
+ },
923
+ "128115": {
924
+ "content": "<|reserved_special_token_107|>",
925
+ "lstrip": false,
926
+ "normalized": false,
927
+ "rstrip": false,
928
+ "single_word": false,
929
+ "special": true
930
+ },
931
+ "128116": {
932
+ "content": "<|reserved_special_token_108|>",
933
+ "lstrip": false,
934
+ "normalized": false,
935
+ "rstrip": false,
936
+ "single_word": false,
937
+ "special": true
938
+ },
939
+ "128117": {
940
+ "content": "<|reserved_special_token_109|>",
941
+ "lstrip": false,
942
+ "normalized": false,
943
+ "rstrip": false,
944
+ "single_word": false,
945
+ "special": true
946
+ },
947
+ "128118": {
948
+ "content": "<|reserved_special_token_110|>",
949
+ "lstrip": false,
950
+ "normalized": false,
951
+ "rstrip": false,
952
+ "single_word": false,
953
+ "special": true
954
+ },
955
+ "128119": {
956
+ "content": "<|reserved_special_token_111|>",
957
+ "lstrip": false,
958
+ "normalized": false,
959
+ "rstrip": false,
960
+ "single_word": false,
961
+ "special": true
962
+ },
963
+ "128120": {
964
+ "content": "<|reserved_special_token_112|>",
965
+ "lstrip": false,
966
+ "normalized": false,
967
+ "rstrip": false,
968
+ "single_word": false,
969
+ "special": true
970
+ },
971
+ "128121": {
972
+ "content": "<|reserved_special_token_113|>",
973
+ "lstrip": false,
974
+ "normalized": false,
975
+ "rstrip": false,
976
+ "single_word": false,
977
+ "special": true
978
+ },
979
+ "128122": {
980
+ "content": "<|reserved_special_token_114|>",
981
+ "lstrip": false,
982
+ "normalized": false,
983
+ "rstrip": false,
984
+ "single_word": false,
985
+ "special": true
986
+ },
987
+ "128123": {
988
+ "content": "<|reserved_special_token_115|>",
989
+ "lstrip": false,
990
+ "normalized": false,
991
+ "rstrip": false,
992
+ "single_word": false,
993
+ "special": true
994
+ },
995
+ "128124": {
996
+ "content": "<|reserved_special_token_116|>",
997
+ "lstrip": false,
998
+ "normalized": false,
999
+ "rstrip": false,
1000
+ "single_word": false,
1001
+ "special": true
1002
+ },
1003
+ "128125": {
1004
+ "content": "<|reserved_special_token_117|>",
1005
+ "lstrip": false,
1006
+ "normalized": false,
1007
+ "rstrip": false,
1008
+ "single_word": false,
1009
+ "special": true
1010
+ },
1011
+ "128126": {
1012
+ "content": "<|reserved_special_token_118|>",
1013
+ "lstrip": false,
1014
+ "normalized": false,
1015
+ "rstrip": false,
1016
+ "single_word": false,
1017
+ "special": true
1018
+ },
1019
+ "128127": {
1020
+ "content": "<|reserved_special_token_119|>",
1021
+ "lstrip": false,
1022
+ "normalized": false,
1023
+ "rstrip": false,
1024
+ "single_word": false,
1025
+ "special": true
1026
+ },
1027
+ "128128": {
1028
+ "content": "<|reserved_special_token_120|>",
1029
+ "lstrip": false,
1030
+ "normalized": false,
1031
+ "rstrip": false,
1032
+ "single_word": false,
1033
+ "special": true
1034
+ },
1035
+ "128129": {
1036
+ "content": "<|reserved_special_token_121|>",
1037
+ "lstrip": false,
1038
+ "normalized": false,
1039
+ "rstrip": false,
1040
+ "single_word": false,
1041
+ "special": true
1042
+ },
1043
+ "128130": {
1044
+ "content": "<|reserved_special_token_122|>",
1045
+ "lstrip": false,
1046
+ "normalized": false,
1047
+ "rstrip": false,
1048
+ "single_word": false,
1049
+ "special": true
1050
+ },
1051
+ "128131": {
1052
+ "content": "<|reserved_special_token_123|>",
1053
+ "lstrip": false,
1054
+ "normalized": false,
1055
+ "rstrip": false,
1056
+ "single_word": false,
1057
+ "special": true
1058
+ },
1059
+ "128132": {
1060
+ "content": "<|reserved_special_token_124|>",
1061
+ "lstrip": false,
1062
+ "normalized": false,
1063
+ "rstrip": false,
1064
+ "single_word": false,
1065
+ "special": true
1066
+ },
1067
+ "128133": {
1068
+ "content": "<|reserved_special_token_125|>",
1069
+ "lstrip": false,
1070
+ "normalized": false,
1071
+ "rstrip": false,
1072
+ "single_word": false,
1073
+ "special": true
1074
+ },
1075
+ "128134": {
1076
+ "content": "<|reserved_special_token_126|>",
1077
+ "lstrip": false,
1078
+ "normalized": false,
1079
+ "rstrip": false,
1080
+ "single_word": false,
1081
+ "special": true
1082
+ },
1083
+ "128135": {
1084
+ "content": "<|reserved_special_token_127|>",
1085
+ "lstrip": false,
1086
+ "normalized": false,
1087
+ "rstrip": false,
1088
+ "single_word": false,
1089
+ "special": true
1090
+ },
1091
+ "128136": {
1092
+ "content": "<|reserved_special_token_128|>",
1093
+ "lstrip": false,
1094
+ "normalized": false,
1095
+ "rstrip": false,
1096
+ "single_word": false,
1097
+ "special": true
1098
+ },
1099
+ "128137": {
1100
+ "content": "<|reserved_special_token_129|>",
1101
+ "lstrip": false,
1102
+ "normalized": false,
1103
+ "rstrip": false,
1104
+ "single_word": false,
1105
+ "special": true
1106
+ },
1107
+ "128138": {
1108
+ "content": "<|reserved_special_token_130|>",
1109
+ "lstrip": false,
1110
+ "normalized": false,
1111
+ "rstrip": false,
1112
+ "single_word": false,
1113
+ "special": true
1114
+ },
1115
+ "128139": {
1116
+ "content": "<|reserved_special_token_131|>",
1117
+ "lstrip": false,
1118
+ "normalized": false,
1119
+ "rstrip": false,
1120
+ "single_word": false,
1121
+ "special": true
1122
+ },
1123
+ "128140": {
1124
+ "content": "<|reserved_special_token_132|>",
1125
+ "lstrip": false,
1126
+ "normalized": false,
1127
+ "rstrip": false,
1128
+ "single_word": false,
1129
+ "special": true
1130
+ },
1131
+ "128141": {
1132
+ "content": "<|reserved_special_token_133|>",
1133
+ "lstrip": false,
1134
+ "normalized": false,
1135
+ "rstrip": false,
1136
+ "single_word": false,
1137
+ "special": true
1138
+ },
1139
+ "128142": {
1140
+ "content": "<|reserved_special_token_134|>",
1141
+ "lstrip": false,
1142
+ "normalized": false,
1143
+ "rstrip": false,
1144
+ "single_word": false,
1145
+ "special": true
1146
+ },
1147
+ "128143": {
1148
+ "content": "<|reserved_special_token_135|>",
1149
+ "lstrip": false,
1150
+ "normalized": false,
1151
+ "rstrip": false,
1152
+ "single_word": false,
1153
+ "special": true
1154
+ },
1155
+ "128144": {
1156
+ "content": "<|reserved_special_token_136|>",
1157
+ "lstrip": false,
1158
+ "normalized": false,
1159
+ "rstrip": false,
1160
+ "single_word": false,
1161
+ "special": true
1162
+ },
1163
+ "128145": {
1164
+ "content": "<|reserved_special_token_137|>",
1165
+ "lstrip": false,
1166
+ "normalized": false,
1167
+ "rstrip": false,
1168
+ "single_word": false,
1169
+ "special": true
1170
+ },
1171
+ "128146": {
1172
+ "content": "<|reserved_special_token_138|>",
1173
+ "lstrip": false,
1174
+ "normalized": false,
1175
+ "rstrip": false,
1176
+ "single_word": false,
1177
+ "special": true
1178
+ },
1179
+ "128147": {
1180
+ "content": "<|reserved_special_token_139|>",
1181
+ "lstrip": false,
1182
+ "normalized": false,
1183
+ "rstrip": false,
1184
+ "single_word": false,
1185
+ "special": true
1186
+ },
1187
+ "128148": {
1188
+ "content": "<|reserved_special_token_140|>",
1189
+ "lstrip": false,
1190
+ "normalized": false,
1191
+ "rstrip": false,
1192
+ "single_word": false,
1193
+ "special": true
1194
+ },
1195
+ "128149": {
1196
+ "content": "<|reserved_special_token_141|>",
1197
+ "lstrip": false,
1198
+ "normalized": false,
1199
+ "rstrip": false,
1200
+ "single_word": false,
1201
+ "special": true
1202
+ },
1203
+ "128150": {
1204
+ "content": "<|reserved_special_token_142|>",
1205
+ "lstrip": false,
1206
+ "normalized": false,
1207
+ "rstrip": false,
1208
+ "single_word": false,
1209
+ "special": true
1210
+ },
1211
+ "128151": {
1212
+ "content": "<|reserved_special_token_143|>",
1213
+ "lstrip": false,
1214
+ "normalized": false,
1215
+ "rstrip": false,
1216
+ "single_word": false,
1217
+ "special": true
1218
+ },
1219
+ "128152": {
1220
+ "content": "<|reserved_special_token_144|>",
1221
+ "lstrip": false,
1222
+ "normalized": false,
1223
+ "rstrip": false,
1224
+ "single_word": false,
1225
+ "special": true
1226
+ },
1227
+ "128153": {
1228
+ "content": "<|reserved_special_token_145|>",
1229
+ "lstrip": false,
1230
+ "normalized": false,
1231
+ "rstrip": false,
1232
+ "single_word": false,
1233
+ "special": true
1234
+ },
1235
+ "128154": {
1236
+ "content": "<|reserved_special_token_146|>",
1237
+ "lstrip": false,
1238
+ "normalized": false,
1239
+ "rstrip": false,
1240
+ "single_word": false,
1241
+ "special": true
1242
+ },
1243
+ "128155": {
1244
+ "content": "<|reserved_special_token_147|>",
1245
+ "lstrip": false,
1246
+ "normalized": false,
1247
+ "rstrip": false,
1248
+ "single_word": false,
1249
+ "special": true
1250
+ },
1251
+ "128156": {
1252
+ "content": "<|reserved_special_token_148|>",
1253
+ "lstrip": false,
1254
+ "normalized": false,
1255
+ "rstrip": false,
1256
+ "single_word": false,
1257
+ "special": true
1258
+ },
1259
+ "128157": {
1260
+ "content": "<|reserved_special_token_149|>",
1261
+ "lstrip": false,
1262
+ "normalized": false,
1263
+ "rstrip": false,
1264
+ "single_word": false,
1265
+ "special": true
1266
+ },
1267
+ "128158": {
1268
+ "content": "<|reserved_special_token_150|>",
1269
+ "lstrip": false,
1270
+ "normalized": false,
1271
+ "rstrip": false,
1272
+ "single_word": false,
1273
+ "special": true
1274
+ },
1275
+ "128159": {
1276
+ "content": "<|reserved_special_token_151|>",
1277
+ "lstrip": false,
1278
+ "normalized": false,
1279
+ "rstrip": false,
1280
+ "single_word": false,
1281
+ "special": true
1282
+ },
1283
+ "128160": {
1284
+ "content": "<|reserved_special_token_152|>",
1285
+ "lstrip": false,
1286
+ "normalized": false,
1287
+ "rstrip": false,
1288
+ "single_word": false,
1289
+ "special": true
1290
+ },
1291
+ "128161": {
1292
+ "content": "<|reserved_special_token_153|>",
1293
+ "lstrip": false,
1294
+ "normalized": false,
1295
+ "rstrip": false,
1296
+ "single_word": false,
1297
+ "special": true
1298
+ },
1299
+ "128162": {
1300
+ "content": "<|reserved_special_token_154|>",
1301
+ "lstrip": false,
1302
+ "normalized": false,
1303
+ "rstrip": false,
1304
+ "single_word": false,
1305
+ "special": true
1306
+ },
1307
+ "128163": {
1308
+ "content": "<|reserved_special_token_155|>",
1309
+ "lstrip": false,
1310
+ "normalized": false,
1311
+ "rstrip": false,
1312
+ "single_word": false,
1313
+ "special": true
1314
+ },
1315
+ "128164": {
1316
+ "content": "<|reserved_special_token_156|>",
1317
+ "lstrip": false,
1318
+ "normalized": false,
1319
+ "rstrip": false,
1320
+ "single_word": false,
1321
+ "special": true
1322
+ },
1323
+ "128165": {
1324
+ "content": "<|reserved_special_token_157|>",
1325
+ "lstrip": false,
1326
+ "normalized": false,
1327
+ "rstrip": false,
1328
+ "single_word": false,
1329
+ "special": true
1330
+ },
1331
+ "128166": {
1332
+ "content": "<|reserved_special_token_158|>",
1333
+ "lstrip": false,
1334
+ "normalized": false,
1335
+ "rstrip": false,
1336
+ "single_word": false,
1337
+ "special": true
1338
+ },
1339
+ "128167": {
1340
+ "content": "<|reserved_special_token_159|>",
1341
+ "lstrip": false,
1342
+ "normalized": false,
1343
+ "rstrip": false,
1344
+ "single_word": false,
1345
+ "special": true
1346
+ },
1347
+ "128168": {
1348
+ "content": "<|reserved_special_token_160|>",
1349
+ "lstrip": false,
1350
+ "normalized": false,
1351
+ "rstrip": false,
1352
+ "single_word": false,
1353
+ "special": true
1354
+ },
1355
+ "128169": {
1356
+ "content": "<|reserved_special_token_161|>",
1357
+ "lstrip": false,
1358
+ "normalized": false,
1359
+ "rstrip": false,
1360
+ "single_word": false,
1361
+ "special": true
1362
+ },
1363
+ "128170": {
1364
+ "content": "<|reserved_special_token_162|>",
1365
+ "lstrip": false,
1366
+ "normalized": false,
1367
+ "rstrip": false,
1368
+ "single_word": false,
1369
+ "special": true
1370
+ },
1371
+ "128171": {
1372
+ "content": "<|reserved_special_token_163|>",
1373
+ "lstrip": false,
1374
+ "normalized": false,
1375
+ "rstrip": false,
1376
+ "single_word": false,
1377
+ "special": true
1378
+ },
1379
+ "128172": {
1380
+ "content": "<|reserved_special_token_164|>",
1381
+ "lstrip": false,
1382
+ "normalized": false,
1383
+ "rstrip": false,
1384
+ "single_word": false,
1385
+ "special": true
1386
+ },
1387
+ "128173": {
1388
+ "content": "<|reserved_special_token_165|>",
1389
+ "lstrip": false,
1390
+ "normalized": false,
1391
+ "rstrip": false,
1392
+ "single_word": false,
1393
+ "special": true
1394
+ },
1395
+ "128174": {
1396
+ "content": "<|reserved_special_token_166|>",
1397
+ "lstrip": false,
1398
+ "normalized": false,
1399
+ "rstrip": false,
1400
+ "single_word": false,
1401
+ "special": true
1402
+ },
1403
+ "128175": {
1404
+ "content": "<|reserved_special_token_167|>",
1405
+ "lstrip": false,
1406
+ "normalized": false,
1407
+ "rstrip": false,
1408
+ "single_word": false,
1409
+ "special": true
1410
+ },
1411
+ "128176": {
1412
+ "content": "<|reserved_special_token_168|>",
1413
+ "lstrip": false,
1414
+ "normalized": false,
1415
+ "rstrip": false,
1416
+ "single_word": false,
1417
+ "special": true
1418
+ },
1419
+ "128177": {
1420
+ "content": "<|reserved_special_token_169|>",
1421
+ "lstrip": false,
1422
+ "normalized": false,
1423
+ "rstrip": false,
1424
+ "single_word": false,
1425
+ "special": true
1426
+ },
1427
+ "128178": {
1428
+ "content": "<|reserved_special_token_170|>",
1429
+ "lstrip": false,
1430
+ "normalized": false,
1431
+ "rstrip": false,
1432
+ "single_word": false,
1433
+ "special": true
1434
+ },
1435
+ "128179": {
1436
+ "content": "<|reserved_special_token_171|>",
1437
+ "lstrip": false,
1438
+ "normalized": false,
1439
+ "rstrip": false,
1440
+ "single_word": false,
1441
+ "special": true
1442
+ },
1443
+ "128180": {
1444
+ "content": "<|reserved_special_token_172|>",
1445
+ "lstrip": false,
1446
+ "normalized": false,
1447
+ "rstrip": false,
1448
+ "single_word": false,
1449
+ "special": true
1450
+ },
1451
+ "128181": {
1452
+ "content": "<|reserved_special_token_173|>",
1453
+ "lstrip": false,
1454
+ "normalized": false,
1455
+ "rstrip": false,
1456
+ "single_word": false,
1457
+ "special": true
1458
+ },
1459
+ "128182": {
1460
+ "content": "<|reserved_special_token_174|>",
1461
+ "lstrip": false,
1462
+ "normalized": false,
1463
+ "rstrip": false,
1464
+ "single_word": false,
1465
+ "special": true
1466
+ },
1467
+ "128183": {
1468
+ "content": "<|reserved_special_token_175|>",
1469
+ "lstrip": false,
1470
+ "normalized": false,
1471
+ "rstrip": false,
1472
+ "single_word": false,
1473
+ "special": true
1474
+ },
1475
+ "128184": {
1476
+ "content": "<|reserved_special_token_176|>",
1477
+ "lstrip": false,
1478
+ "normalized": false,
1479
+ "rstrip": false,
1480
+ "single_word": false,
1481
+ "special": true
1482
+ },
1483
+ "128185": {
1484
+ "content": "<|reserved_special_token_177|>",
1485
+ "lstrip": false,
1486
+ "normalized": false,
1487
+ "rstrip": false,
1488
+ "single_word": false,
1489
+ "special": true
1490
+ },
1491
+ "128186": {
1492
+ "content": "<|reserved_special_token_178|>",
1493
+ "lstrip": false,
1494
+ "normalized": false,
1495
+ "rstrip": false,
1496
+ "single_word": false,
1497
+ "special": true
1498
+ },
1499
+ "128187": {
1500
+ "content": "<|reserved_special_token_179|>",
1501
+ "lstrip": false,
1502
+ "normalized": false,
1503
+ "rstrip": false,
1504
+ "single_word": false,
1505
+ "special": true
1506
+ },
1507
+ "128188": {
1508
+ "content": "<|reserved_special_token_180|>",
1509
+ "lstrip": false,
1510
+ "normalized": false,
1511
+ "rstrip": false,
1512
+ "single_word": false,
1513
+ "special": true
1514
+ },
1515
+ "128189": {
1516
+ "content": "<|reserved_special_token_181|>",
1517
+ "lstrip": false,
1518
+ "normalized": false,
1519
+ "rstrip": false,
1520
+ "single_word": false,
1521
+ "special": true
1522
+ },
1523
+ "128190": {
1524
+ "content": "<|reserved_special_token_182|>",
1525
+ "lstrip": false,
1526
+ "normalized": false,
1527
+ "rstrip": false,
1528
+ "single_word": false,
1529
+ "special": true
1530
+ },
1531
+ "128191": {
1532
+ "content": "<|reserved_special_token_183|>",
1533
+ "lstrip": false,
1534
+ "normalized": false,
1535
+ "rstrip": false,
1536
+ "single_word": false,
1537
+ "special": true
1538
+ },
1539
+ "128192": {
1540
+ "content": "<|reserved_special_token_184|>",
1541
+ "lstrip": false,
1542
+ "normalized": false,
1543
+ "rstrip": false,
1544
+ "single_word": false,
1545
+ "special": true
1546
+ },
1547
+ "128193": {
1548
+ "content": "<|reserved_special_token_185|>",
1549
+ "lstrip": false,
1550
+ "normalized": false,
1551
+ "rstrip": false,
1552
+ "single_word": false,
1553
+ "special": true
1554
+ },
1555
+ "128194": {
1556
+ "content": "<|reserved_special_token_186|>",
1557
+ "lstrip": false,
1558
+ "normalized": false,
1559
+ "rstrip": false,
1560
+ "single_word": false,
1561
+ "special": true
1562
+ },
1563
+ "128195": {
1564
+ "content": "<|reserved_special_token_187|>",
1565
+ "lstrip": false,
1566
+ "normalized": false,
1567
+ "rstrip": false,
1568
+ "single_word": false,
1569
+ "special": true
1570
+ },
1571
+ "128196": {
1572
+ "content": "<|reserved_special_token_188|>",
1573
+ "lstrip": false,
1574
+ "normalized": false,
1575
+ "rstrip": false,
1576
+ "single_word": false,
1577
+ "special": true
1578
+ },
1579
+ "128197": {
1580
+ "content": "<|reserved_special_token_189|>",
1581
+ "lstrip": false,
1582
+ "normalized": false,
1583
+ "rstrip": false,
1584
+ "single_word": false,
1585
+ "special": true
1586
+ },
1587
+ "128198": {
1588
+ "content": "<|reserved_special_token_190|>",
1589
+ "lstrip": false,
1590
+ "normalized": false,
1591
+ "rstrip": false,
1592
+ "single_word": false,
1593
+ "special": true
1594
+ },
1595
+ "128199": {
1596
+ "content": "<|reserved_special_token_191|>",
1597
+ "lstrip": false,
1598
+ "normalized": false,
1599
+ "rstrip": false,
1600
+ "single_word": false,
1601
+ "special": true
1602
+ },
1603
+ "128200": {
1604
+ "content": "<|reserved_special_token_192|>",
1605
+ "lstrip": false,
1606
+ "normalized": false,
1607
+ "rstrip": false,
1608
+ "single_word": false,
1609
+ "special": true
1610
+ },
1611
+ "128201": {
1612
+ "content": "<|reserved_special_token_193|>",
1613
+ "lstrip": false,
1614
+ "normalized": false,
1615
+ "rstrip": false,
1616
+ "single_word": false,
1617
+ "special": true
1618
+ },
1619
+ "128202": {
1620
+ "content": "<|reserved_special_token_194|>",
1621
+ "lstrip": false,
1622
+ "normalized": false,
1623
+ "rstrip": false,
1624
+ "single_word": false,
1625
+ "special": true
1626
+ },
1627
+ "128203": {
1628
+ "content": "<|reserved_special_token_195|>",
1629
+ "lstrip": false,
1630
+ "normalized": false,
1631
+ "rstrip": false,
1632
+ "single_word": false,
1633
+ "special": true
1634
+ },
1635
+ "128204": {
1636
+ "content": "<|reserved_special_token_196|>",
1637
+ "lstrip": false,
1638
+ "normalized": false,
1639
+ "rstrip": false,
1640
+ "single_word": false,
1641
+ "special": true
1642
+ },
1643
+ "128205": {
1644
+ "content": "<|reserved_special_token_197|>",
1645
+ "lstrip": false,
1646
+ "normalized": false,
1647
+ "rstrip": false,
1648
+ "single_word": false,
1649
+ "special": true
1650
+ },
1651
+ "128206": {
1652
+ "content": "<|reserved_special_token_198|>",
1653
+ "lstrip": false,
1654
+ "normalized": false,
1655
+ "rstrip": false,
1656
+ "single_word": false,
1657
+ "special": true
1658
+ },
1659
+ "128207": {
1660
+ "content": "<|reserved_special_token_199|>",
1661
+ "lstrip": false,
1662
+ "normalized": false,
1663
+ "rstrip": false,
1664
+ "single_word": false,
1665
+ "special": true
1666
+ },
1667
+ "128208": {
1668
+ "content": "<|reserved_special_token_200|>",
1669
+ "lstrip": false,
1670
+ "normalized": false,
1671
+ "rstrip": false,
1672
+ "single_word": false,
1673
+ "special": true
1674
+ },
1675
+ "128209": {
1676
+ "content": "<|reserved_special_token_201|>",
1677
+ "lstrip": false,
1678
+ "normalized": false,
1679
+ "rstrip": false,
1680
+ "single_word": false,
1681
+ "special": true
1682
+ },
1683
+ "128210": {
1684
+ "content": "<|reserved_special_token_202|>",
1685
+ "lstrip": false,
1686
+ "normalized": false,
1687
+ "rstrip": false,
1688
+ "single_word": false,
1689
+ "special": true
1690
+ },
1691
+ "128211": {
1692
+ "content": "<|reserved_special_token_203|>",
1693
+ "lstrip": false,
1694
+ "normalized": false,
1695
+ "rstrip": false,
1696
+ "single_word": false,
1697
+ "special": true
1698
+ },
1699
+ "128212": {
1700
+ "content": "<|reserved_special_token_204|>",
1701
+ "lstrip": false,
1702
+ "normalized": false,
1703
+ "rstrip": false,
1704
+ "single_word": false,
1705
+ "special": true
1706
+ },
1707
+ "128213": {
1708
+ "content": "<|reserved_special_token_205|>",
1709
+ "lstrip": false,
1710
+ "normalized": false,
1711
+ "rstrip": false,
1712
+ "single_word": false,
1713
+ "special": true
1714
+ },
1715
+ "128214": {
1716
+ "content": "<|reserved_special_token_206|>",
1717
+ "lstrip": false,
1718
+ "normalized": false,
1719
+ "rstrip": false,
1720
+ "single_word": false,
1721
+ "special": true
1722
+ },
1723
+ "128215": {
1724
+ "content": "<|reserved_special_token_207|>",
1725
+ "lstrip": false,
1726
+ "normalized": false,
1727
+ "rstrip": false,
1728
+ "single_word": false,
1729
+ "special": true
1730
+ },
1731
+ "128216": {
1732
+ "content": "<|reserved_special_token_208|>",
1733
+ "lstrip": false,
1734
+ "normalized": false,
1735
+ "rstrip": false,
1736
+ "single_word": false,
1737
+ "special": true
1738
+ },
1739
+ "128217": {
1740
+ "content": "<|reserved_special_token_209|>",
1741
+ "lstrip": false,
1742
+ "normalized": false,
1743
+ "rstrip": false,
1744
+ "single_word": false,
1745
+ "special": true
1746
+ },
1747
+ "128218": {
1748
+ "content": "<|reserved_special_token_210|>",
1749
+ "lstrip": false,
1750
+ "normalized": false,
1751
+ "rstrip": false,
1752
+ "single_word": false,
1753
+ "special": true
1754
+ },
1755
+ "128219": {
1756
+ "content": "<|reserved_special_token_211|>",
1757
+ "lstrip": false,
1758
+ "normalized": false,
1759
+ "rstrip": false,
1760
+ "single_word": false,
1761
+ "special": true
1762
+ },
1763
+ "128220": {
1764
+ "content": "<|reserved_special_token_212|>",
1765
+ "lstrip": false,
1766
+ "normalized": false,
1767
+ "rstrip": false,
1768
+ "single_word": false,
1769
+ "special": true
1770
+ },
1771
+ "128221": {
1772
+ "content": "<|reserved_special_token_213|>",
1773
+ "lstrip": false,
1774
+ "normalized": false,
1775
+ "rstrip": false,
1776
+ "single_word": false,
1777
+ "special": true
1778
+ },
1779
+ "128222": {
1780
+ "content": "<|reserved_special_token_214|>",
1781
+ "lstrip": false,
1782
+ "normalized": false,
1783
+ "rstrip": false,
1784
+ "single_word": false,
1785
+ "special": true
1786
+ },
1787
+ "128223": {
1788
+ "content": "<|reserved_special_token_215|>",
1789
+ "lstrip": false,
1790
+ "normalized": false,
1791
+ "rstrip": false,
1792
+ "single_word": false,
1793
+ "special": true
1794
+ },
1795
+ "128224": {
1796
+ "content": "<|reserved_special_token_216|>",
1797
+ "lstrip": false,
1798
+ "normalized": false,
1799
+ "rstrip": false,
1800
+ "single_word": false,
1801
+ "special": true
1802
+ },
1803
+ "128225": {
1804
+ "content": "<|reserved_special_token_217|>",
1805
+ "lstrip": false,
1806
+ "normalized": false,
1807
+ "rstrip": false,
1808
+ "single_word": false,
1809
+ "special": true
1810
+ },
1811
+ "128226": {
1812
+ "content": "<|reserved_special_token_218|>",
1813
+ "lstrip": false,
1814
+ "normalized": false,
1815
+ "rstrip": false,
1816
+ "single_word": false,
1817
+ "special": true
1818
+ },
1819
+ "128227": {
1820
+ "content": "<|reserved_special_token_219|>",
1821
+ "lstrip": false,
1822
+ "normalized": false,
1823
+ "rstrip": false,
1824
+ "single_word": false,
1825
+ "special": true
1826
+ },
1827
+ "128228": {
1828
+ "content": "<|reserved_special_token_220|>",
1829
+ "lstrip": false,
1830
+ "normalized": false,
1831
+ "rstrip": false,
1832
+ "single_word": false,
1833
+ "special": true
1834
+ },
1835
+ "128229": {
1836
+ "content": "<|reserved_special_token_221|>",
1837
+ "lstrip": false,
1838
+ "normalized": false,
1839
+ "rstrip": false,
1840
+ "single_word": false,
1841
+ "special": true
1842
+ },
1843
+ "128230": {
1844
+ "content": "<|reserved_special_token_222|>",
1845
+ "lstrip": false,
1846
+ "normalized": false,
1847
+ "rstrip": false,
1848
+ "single_word": false,
1849
+ "special": true
1850
+ },
1851
+ "128231": {
1852
+ "content": "<|reserved_special_token_223|>",
1853
+ "lstrip": false,
1854
+ "normalized": false,
1855
+ "rstrip": false,
1856
+ "single_word": false,
1857
+ "special": true
1858
+ },
1859
+ "128232": {
1860
+ "content": "<|reserved_special_token_224|>",
1861
+ "lstrip": false,
1862
+ "normalized": false,
1863
+ "rstrip": false,
1864
+ "single_word": false,
1865
+ "special": true
1866
+ },
1867
+ "128233": {
1868
+ "content": "<|reserved_special_token_225|>",
1869
+ "lstrip": false,
1870
+ "normalized": false,
1871
+ "rstrip": false,
1872
+ "single_word": false,
1873
+ "special": true
1874
+ },
1875
+ "128234": {
1876
+ "content": "<|reserved_special_token_226|>",
1877
+ "lstrip": false,
1878
+ "normalized": false,
1879
+ "rstrip": false,
1880
+ "single_word": false,
1881
+ "special": true
1882
+ },
1883
+ "128235": {
1884
+ "content": "<|reserved_special_token_227|>",
1885
+ "lstrip": false,
1886
+ "normalized": false,
1887
+ "rstrip": false,
1888
+ "single_word": false,
1889
+ "special": true
1890
+ },
1891
+ "128236": {
1892
+ "content": "<|reserved_special_token_228|>",
1893
+ "lstrip": false,
1894
+ "normalized": false,
1895
+ "rstrip": false,
1896
+ "single_word": false,
1897
+ "special": true
1898
+ },
1899
+ "128237": {
1900
+ "content": "<|reserved_special_token_229|>",
1901
+ "lstrip": false,
1902
+ "normalized": false,
1903
+ "rstrip": false,
1904
+ "single_word": false,
1905
+ "special": true
1906
+ },
1907
+ "128238": {
1908
+ "content": "<|reserved_special_token_230|>",
1909
+ "lstrip": false,
1910
+ "normalized": false,
1911
+ "rstrip": false,
1912
+ "single_word": false,
1913
+ "special": true
1914
+ },
1915
+ "128239": {
1916
+ "content": "<|reserved_special_token_231|>",
1917
+ "lstrip": false,
1918
+ "normalized": false,
1919
+ "rstrip": false,
1920
+ "single_word": false,
1921
+ "special": true
1922
+ },
1923
+ "128240": {
1924
+ "content": "<|reserved_special_token_232|>",
1925
+ "lstrip": false,
1926
+ "normalized": false,
1927
+ "rstrip": false,
1928
+ "single_word": false,
1929
+ "special": true
1930
+ },
1931
+ "128241": {
1932
+ "content": "<|reserved_special_token_233|>",
1933
+ "lstrip": false,
1934
+ "normalized": false,
1935
+ "rstrip": false,
1936
+ "single_word": false,
1937
+ "special": true
1938
+ },
1939
+ "128242": {
1940
+ "content": "<|reserved_special_token_234|>",
1941
+ "lstrip": false,
1942
+ "normalized": false,
1943
+ "rstrip": false,
1944
+ "single_word": false,
1945
+ "special": true
1946
+ },
1947
+ "128243": {
1948
+ "content": "<|reserved_special_token_235|>",
1949
+ "lstrip": false,
1950
+ "normalized": false,
1951
+ "rstrip": false,
1952
+ "single_word": false,
1953
+ "special": true
1954
+ },
1955
+ "128244": {
1956
+ "content": "<|reserved_special_token_236|>",
1957
+ "lstrip": false,
1958
+ "normalized": false,
1959
+ "rstrip": false,
1960
+ "single_word": false,
1961
+ "special": true
1962
+ },
1963
+ "128245": {
1964
+ "content": "<|reserved_special_token_237|>",
1965
+ "lstrip": false,
1966
+ "normalized": false,
1967
+ "rstrip": false,
1968
+ "single_word": false,
1969
+ "special": true
1970
+ },
1971
+ "128246": {
1972
+ "content": "<|reserved_special_token_238|>",
1973
+ "lstrip": false,
1974
+ "normalized": false,
1975
+ "rstrip": false,
1976
+ "single_word": false,
1977
+ "special": true
1978
+ },
1979
+ "128247": {
1980
+ "content": "<|reserved_special_token_239|>",
1981
+ "lstrip": false,
1982
+ "normalized": false,
1983
+ "rstrip": false,
1984
+ "single_word": false,
1985
+ "special": true
1986
+ },
1987
+ "128248": {
1988
+ "content": "<|reserved_special_token_240|>",
1989
+ "lstrip": false,
1990
+ "normalized": false,
1991
+ "rstrip": false,
1992
+ "single_word": false,
1993
+ "special": true
1994
+ },
1995
+ "128249": {
1996
+ "content": "<|reserved_special_token_241|>",
1997
+ "lstrip": false,
1998
+ "normalized": false,
1999
+ "rstrip": false,
2000
+ "single_word": false,
2001
+ "special": true
2002
+ },
2003
+ "128250": {
2004
+ "content": "<|reserved_special_token_242|>",
2005
+ "lstrip": false,
2006
+ "normalized": false,
2007
+ "rstrip": false,
2008
+ "single_word": false,
2009
+ "special": true
2010
+ },
2011
+ "128251": {
2012
+ "content": "<|reserved_special_token_243|>",
2013
+ "lstrip": false,
2014
+ "normalized": false,
2015
+ "rstrip": false,
2016
+ "single_word": false,
2017
+ "special": true
2018
+ },
2019
+ "128252": {
2020
+ "content": "<|reserved_special_token_244|>",
2021
+ "lstrip": false,
2022
+ "normalized": false,
2023
+ "rstrip": false,
2024
+ "single_word": false,
2025
+ "special": true
2026
+ },
2027
+ "128253": {
2028
+ "content": "<|reserved_special_token_245|>",
2029
+ "lstrip": false,
2030
+ "normalized": false,
2031
+ "rstrip": false,
2032
+ "single_word": false,
2033
+ "special": true
2034
+ },
2035
+ "128254": {
2036
+ "content": "<|reserved_special_token_246|>",
2037
+ "lstrip": false,
2038
+ "normalized": false,
2039
+ "rstrip": false,
2040
+ "single_word": false,
2041
+ "special": true
2042
+ },
2043
+ "128255": {
2044
+ "content": "<|reserved_special_token_247|>",
2045
+ "lstrip": false,
2046
+ "normalized": false,
2047
+ "rstrip": false,
2048
+ "single_word": false,
2049
+ "special": true
2050
+ }
2051
+ },
2052
+ "bos_token": "<|begin_of_text|>",
2053
+ "clean_up_tokenization_spaces": true,
2054
+ "eos_token": "<|end_of_text|>",
2055
+ "extra_special_tokens": {},
2056
+ "model_input_names": [
2057
+ "input_ids",
2058
+ "attention_mask"
2059
+ ],
2060
+ "model_max_length": 131072,
2061
+ "tokenizer_class": "PreTrainedTokenizerFast"
2062
+ }
12500/checkpoint_metadata.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cp": 1,
3
+ "custom_metas": null,
4
+ "dp": 8,
5
+ "metas": {
6
+ "consumed_tokens_total": 52428800000,
7
+ "consumed_train_samples": 6400000,
8
+ "data_stages": [
9
+ {
10
+ "consumed_tokens_per_dataset_folder": {
11
+ "/home/aiops/zhuty/cont_data/mathprosf/tokenized": 52428800000
12
+ },
13
+ "consumed_train_samples": 6400000,
14
+ "name": "mathprosf",
15
+ "sequence_length": 8192,
16
+ "start_training_step": 1
17
+ }
18
+ ],
19
+ "last_stage_idx": 0,
20
+ "last_train_step": 12500
21
+ },
22
+ "tp": 1,
23
+ "version": "1.5"
24
+ }
12500/config.yaml ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ checkpoints:
2
+ checkpoint_interval: 500
3
+ checkpoints_path: checkpoints
4
+ checkpoints_path_is_shared_file_system: false
5
+ load_lr_scheduler: true
6
+ load_optimizer: true
7
+ resume_checkpoint_path: checkpoints
8
+ save_final_state: true
9
+ save_initial_state: false
10
+ data_stages:
11
+ - data:
12
+ dataset:
13
+ dataset_folder:
14
+ - /home/aiops/zhuty/cont_data/mathprosf/tokenized
15
+ dataset_max_tokens: null
16
+ dataset_read_path: null
17
+ dataset_weights:
18
+ - 1
19
+ pad_samples_to_global_batch_size: false
20
+ return_positions: true
21
+ shuffle_files: false
22
+ skip_in_stream: false
23
+ token_size_in_bytes: 4
24
+ tokenizer_name: meta-llama/Llama-3.2-1B
25
+ use_old_brrr_dataloader: false
26
+ vocab_size: 128256
27
+ num_loading_workers: 2
28
+ seed: 42
29
+ name: mathprosf
30
+ sequence_length: 8192
31
+ start_training_step: 1
32
+ tokenizer_max_length: null
33
+ tokenizer_name: null
34
+ tokenizer_revision: null
35
+ vocab_size: null
36
+ general:
37
+ benchmark_csv_path: null
38
+ consumed_train_samples: 6400000
39
+ ignore_sanity_checks: true
40
+ project: nano-cont
41
+ run: llama32-1b-mathprosf
42
+ seed: 42
43
+ step: 12500
44
+ lighteval: null
45
+ logging:
46
+ iteration_step_info_interval: 1
47
+ log_level: info
48
+ log_level_replica: info
49
+ metrics_logging: null
50
+ model:
51
+ ddp_bucket_cap_mb: 25
52
+ dtype: bfloat16
53
+ init_method:
54
+ path: /home/aiops/zhuty/nanotron/checkpoints/llama32-1b-nt
55
+ make_vocab_size_divisible_by: 1
56
+ model_config:
57
+ _attn_implementation: flash_attention_2
58
+ _fused_rms_norm: true
59
+ _fused_rotary_emb: true
60
+ _use_doc_masking: false
61
+ _use_qkv_packed: true
62
+ attention_bias: false
63
+ bos_token_id: 128000
64
+ eos_token_id: 128001
65
+ flex_attention_mask: null
66
+ hidden_act: silu
67
+ hidden_size: 2048
68
+ initializer_range: 0.02
69
+ intermediate_size: 8192
70
+ is_qwen2_config: true
71
+ log_attn_probs: false
72
+ max_position_embeddings: 131072
73
+ moe_config: null
74
+ no_rope_layer: null
75
+ num_attention_heads: 32
76
+ num_hidden_layers: 16
77
+ num_key_value_heads: 8
78
+ pad_token_id: null
79
+ pretraining_tp: 1
80
+ ring_attn_heads_k_stride: null
81
+ rms_norm_eps: 1.0e-05
82
+ rope_interleaved: false
83
+ rope_scaling:
84
+ factor: 32.0
85
+ high_freq_factor: 4.0
86
+ low_freq_factor: 1.0
87
+ original_max_position_embeddings: 8192
88
+ rope_type: llama3
89
+ rope_seq_len_interpolation_factor: null
90
+ rope_theta: 500000.0
91
+ sliding_window_size: null
92
+ tie_word_embeddings: true
93
+ use_cache: true
94
+ vocab_size: 128256
95
+ z_loss_coefficient: 0.0001
96
+ z_loss_enabled: false
97
+ optimizer:
98
+ accumulate_grad_in_fp32: true
99
+ clip_grad: 1.0
100
+ learning_rate_scheduler:
101
+ learning_rate: 5.0e-05
102
+ lr_decay_starting_step: 12500
103
+ lr_decay_steps: 12400
104
+ lr_decay_style: linear
105
+ lr_warmup_steps: 100
106
+ lr_warmup_style: linear
107
+ min_decay_lr: 5.0e-05
108
+ optimizer_factory:
109
+ adam_beta1: 0.9
110
+ adam_beta2: 0.95
111
+ adam_eps: 1.0e-08
112
+ name: adamW
113
+ torch_adam_is_fused: true
114
+ weight_decay: 0.1
115
+ weight_decay_exclude_named_params: []
116
+ zero_stage: 0
117
+ parallelism:
118
+ context_parallel_size: 1
119
+ dp: 8
120
+ expert_parallel_size: 1
121
+ pp: 1
122
+ pp_engine: 1f1b
123
+ recompute_layer: true
124
+ tp: 1
125
+ tp_linear_async_communication: true
126
+ tp_mode: REDUCE_SCATTER
127
+ tp_recompute_allgather: false
128
+ profiler: null
129
+ s3_upload: null
130
+ tokenizer:
131
+ tokenizer_max_length: null
132
+ tokenizer_name_or_path: meta-llama/Llama-3.2-1B
133
+ tokenizer_revision: null
134
+ tokens:
135
+ batch_accumulation_per_replica: 64
136
+ limit_test_batches: 0
137
+ limit_val_batches: 0
138
+ micro_batch_size: 1
139
+ sequence_length: 8192
140
+ train_steps: 12500
141
+ val_check_interval: -1
12500/model_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token_id": 128000, "eos_token_id": 128001, "hidden_act": "silu", "hidden_size": 2048, "initializer_range": 0.02, "intermediate_size": 8192, "is_qwen2_config": true, "max_position_embeddings": 131072, "num_attention_heads": 32, "num_hidden_layers": 16, "num_key_value_heads": 8, "pad_token_id": null, "pretraining_tp": 1, "rms_norm_eps": 1e-05, "rope_scaling": {"factor": 32.0, "high_freq_factor": 4.0, "low_freq_factor": 1.0, "original_max_position_embeddings": 8192, "rope_type": "llama3"}, "rope_theta": 500000.0, "rope_interleaved": false, "rope_seq_len_interpolation_factor": null, "tie_word_embeddings": true, "use_cache": true, "vocab_size": 128256, "_attn_implementation": "flash_attention_2", "flex_attention_mask": null, "attention_bias": false, "sliding_window_size": null, "z_loss_enabled": false, "z_loss_coefficient": 0.0001, "no_rope_layer": null, "_fused_rotary_emb": true, "_fused_rms_norm": true, "_use_qkv_packed": true, "_use_doc_masking": false, "log_attn_probs": false, "ring_attn_heads_k_stride": null, "moe_config": null}
12500/optimizer/optimizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"type": "OptimizerFromGradientAccumulator", "parallelism": {"tp_size": "1", "dp_size": "8", "pp_size": "1", "expert_parallel_size": "1"}, "configs": {}}
12500_hf/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 128001,
10
+ "head_dim": 64,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 2048,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 8192,
15
+ "max_position_embeddings": 131072,
16
+ "mlp_bias": false,
17
+ "model_type": "llama",
18
+ "num_attention_heads": 32,
19
+ "num_hidden_layers": 16,
20
+ "num_key_value_heads": 8,
21
+ "pretraining_tp": 1,
22
+ "rms_norm_eps": 1e-05,
23
+ "rope_scaling": {
24
+ "factor": 32.0,
25
+ "high_freq_factor": 4.0,
26
+ "low_freq_factor": 1.0,
27
+ "original_max_position_embeddings": 8192,
28
+ "rope_type": "llama3"
29
+ },
30
+ "rope_theta": 500000.0,
31
+ "tie_word_embeddings": true,
32
+ "transformers_version": "4.57.3",
33
+ "use_cache": true,
34
+ "vocab_size": 128256
35
+ }
12500_hf/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "eos_token_id": 128001,
5
+ "transformers_version": "4.57.3"
6
+ }
12500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/results_2026-01-07T03-51-35.016035.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "alias": "arc_challenge",
5
+ "acc,none": 0.34215017064846415,
6
+ "acc_stderr,none": 0.013864152159177278,
7
+ "acc_norm,none": 0.3651877133105802,
8
+ "acc_norm_stderr,none": 0.014070265519268802
9
+ },
10
+ "arc_easy": {
11
+ "alias": "arc_easy",
12
+ "acc,none": 0.6325757575757576,
13
+ "acc_stderr,none": 0.009892552616211551,
14
+ "acc_norm,none": 0.6094276094276094,
15
+ "acc_norm_stderr,none": 0.010011059112064236
16
+ },
17
+ "hellaswag": {
18
+ "alias": "hellaswag",
19
+ "acc,none": 0.3997211710814579,
20
+ "acc_stderr,none": 0.004888398535520497,
21
+ "acc_norm,none": 0.5164309898426608,
22
+ "acc_norm_stderr,none": 0.004987086426968581
23
+ }
24
+ },
25
+ "group_subtasks": {
26
+ "arc_challenge": [],
27
+ "arc_easy": [],
28
+ "hellaswag": []
29
+ },
30
+ "configs": {
31
+ "arc_challenge": {
32
+ "task": "arc_challenge",
33
+ "tag": [
34
+ "ai2_arc"
35
+ ],
36
+ "dataset_path": "allenai/ai2_arc",
37
+ "dataset_name": "ARC-Challenge",
38
+ "training_split": "train",
39
+ "validation_split": "validation",
40
+ "test_split": "test",
41
+ "doc_to_text": "Question: {{question}}\nAnswer:",
42
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
43
+ "unsafe_code": false,
44
+ "doc_to_choice": "{{choices.text}}",
45
+ "description": "",
46
+ "target_delimiter": " ",
47
+ "fewshot_delimiter": "\n\n",
48
+ "num_fewshot": 0,
49
+ "metric_list": [
50
+ {
51
+ "metric": "acc",
52
+ "aggregation": "mean",
53
+ "higher_is_better": true
54
+ },
55
+ {
56
+ "metric": "acc_norm",
57
+ "aggregation": "mean",
58
+ "higher_is_better": true
59
+ }
60
+ ],
61
+ "output_type": "multiple_choice",
62
+ "repeats": 1,
63
+ "should_decontaminate": true,
64
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
65
+ "metadata": {
66
+ "version": 1.0,
67
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
68
+ }
69
+ },
70
+ "arc_easy": {
71
+ "task": "arc_easy",
72
+ "tag": [
73
+ "ai2_arc"
74
+ ],
75
+ "dataset_path": "allenai/ai2_arc",
76
+ "dataset_name": "ARC-Easy",
77
+ "training_split": "train",
78
+ "validation_split": "validation",
79
+ "test_split": "test",
80
+ "doc_to_text": "Question: {{question}}\nAnswer:",
81
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
82
+ "unsafe_code": false,
83
+ "doc_to_choice": "{{choices.text}}",
84
+ "description": "",
85
+ "target_delimiter": " ",
86
+ "fewshot_delimiter": "\n\n",
87
+ "num_fewshot": 0,
88
+ "metric_list": [
89
+ {
90
+ "metric": "acc",
91
+ "aggregation": "mean",
92
+ "higher_is_better": true
93
+ },
94
+ {
95
+ "metric": "acc_norm",
96
+ "aggregation": "mean",
97
+ "higher_is_better": true
98
+ }
99
+ ],
100
+ "output_type": "multiple_choice",
101
+ "repeats": 1,
102
+ "should_decontaminate": true,
103
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
104
+ "metadata": {
105
+ "version": 1.0,
106
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
107
+ }
108
+ },
109
+ "hellaswag": {
110
+ "task": "hellaswag",
111
+ "tag": [
112
+ "multiple_choice"
113
+ ],
114
+ "dataset_path": "Rowan/hellaswag",
115
+ "training_split": "train",
116
+ "validation_split": "validation",
117
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc):\n ctx = doc[\"ctx_a\"] + \" \" + doc[\"ctx_b\"].capitalize()\n out_doc = {\n \"query\": preprocess(doc[\"activity_label\"] + \": \" + ctx),\n \"choices\": [preprocess(ending) for ending in doc[\"endings\"]],\n \"gold\": int(doc[\"label\"]),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
118
+ "doc_to_text": "{{query}}",
119
+ "doc_to_target": "{{label}}",
120
+ "unsafe_code": false,
121
+ "doc_to_choice": "choices",
122
+ "description": "",
123
+ "target_delimiter": " ",
124
+ "fewshot_delimiter": "\n\n",
125
+ "num_fewshot": 0,
126
+ "metric_list": [
127
+ {
128
+ "metric": "acc",
129
+ "aggregation": "mean",
130
+ "higher_is_better": true
131
+ },
132
+ {
133
+ "metric": "acc_norm",
134
+ "aggregation": "mean",
135
+ "higher_is_better": true
136
+ }
137
+ ],
138
+ "output_type": "multiple_choice",
139
+ "repeats": 1,
140
+ "should_decontaminate": false,
141
+ "metadata": {
142
+ "version": 1.0,
143
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
144
+ }
145
+ }
146
+ },
147
+ "versions": {
148
+ "arc_challenge": 1.0,
149
+ "arc_easy": 1.0,
150
+ "hellaswag": 1.0
151
+ },
152
+ "n-shot": {
153
+ "arc_challenge": 0,
154
+ "arc_easy": 0,
155
+ "hellaswag": 0
156
+ },
157
+ "higher_is_better": {
158
+ "arc_challenge": {
159
+ "acc": true,
160
+ "acc_norm": true
161
+ },
162
+ "arc_easy": {
163
+ "acc": true,
164
+ "acc_norm": true
165
+ },
166
+ "hellaswag": {
167
+ "acc": true,
168
+ "acc_norm": true
169
+ }
170
+ },
171
+ "n-samples": {
172
+ "hellaswag": {
173
+ "original": 10042,
174
+ "effective": 10042
175
+ },
176
+ "arc_easy": {
177
+ "original": 2376,
178
+ "effective": 2376
179
+ },
180
+ "arc_challenge": {
181
+ "original": 1172,
182
+ "effective": 1172
183
+ }
184
+ },
185
+ "config": {
186
+ "model": "hf",
187
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/12500_hf",
188
+ "model_num_parameters": 1235814400,
189
+ "model_dtype": "torch.bfloat16",
190
+ "model_revision": "main",
191
+ "model_sha": "",
192
+ "batch_size": "32",
193
+ "batch_sizes": [],
194
+ "device": null,
195
+ "use_cache": null,
196
+ "limit": null,
197
+ "bootstrap_iters": 100000,
198
+ "gen_kwargs": null,
199
+ "random_seed": 0,
200
+ "numpy_seed": 1234,
201
+ "torch_seed": 1234,
202
+ "fewshot_seed": 1234
203
+ },
204
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
205
+ "date": 1767757751.1847422,
206
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-193-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.183.06\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 256\nOn-line CPU(s) list: 0-255\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7742 64-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 64\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2250.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4491.81\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca sme sev sev_es\nVirtualization: AMD-V\nL1d cache: 4 MiB (128 instances)\nL1i cache: 4 MiB (128 instances)\nL2 cache: 64 MiB (128 instances)\nL3 cache: 512 MiB (32 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-15,128-143\nNUMA node1 CPU(s): 16-31,144-159\nNUMA node2 CPU(s): 32-47,160-175\nNUMA node3 CPU(s): 48-63,176-191\nNUMA node4 CPU(s): 64-79,192-207\nNUMA node5 CPU(s): 80-95,208-223\nNUMA node6 CPU(s): 96-111,224-239\nNUMA node7 CPU(s): 112-127,240-255\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Retbleed: Vulnerable\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; IBRS_FW; STIBP conditional; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
207
+ "transformers_version": "4.57.3",
208
+ "lm_eval_version": "0.4.9.1",
209
+ "upper_git_hash": null,
210
+ "tokenizer_pad_token": [
211
+ "<|end_of_text|>",
212
+ "128001"
213
+ ],
214
+ "tokenizer_eos_token": [
215
+ "<|end_of_text|>",
216
+ "128001"
217
+ ],
218
+ "tokenizer_bos_token": [
219
+ "<|begin_of_text|>",
220
+ "128000"
221
+ ],
222
+ "eot_token_id": 128001,
223
+ "max_length": 131072,
224
+ "task_hashes": {
225
+ "hellaswag": "edcc7edd27a555d3f7cbca0641152b2c5e4eb6eb79c5e62d7fe5887f47814323",
226
+ "arc_easy": "dce0d9b0f0cecd55bf2ac264042c5e45487df708d13123af3ae9e67bbbefdeb1",
227
+ "arc_challenge": "09f9ae87a0905d63512cffc4aa91a55e44258fc35160e40fa1eb66fb75473e34"
228
+ },
229
+ "model_source": "hf",
230
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf",
231
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__12500_hf",
232
+ "system_instruction": null,
233
+ "system_instruction_sha": null,
234
+ "fewshot_as_multiturn": false,
235
+ "chat_template": null,
236
+ "chat_template_sha": null,
237
+ "start_time": 13820945.793376068,
238
+ "end_time": 13821149.222774314,
239
+ "total_evaluation_time_seconds": "203.42939824610949"
240
+ }
12500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_arc_challenge_2026-01-07T03-51-35.016035.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_arc_easy_2026-01-07T03-51-35.016035.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/results_2026-01-07T03-56-06.178713.json ADDED
@@ -0,0 +1,478 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hendrycks_math": {
4
+ "exact_match,none": 0.1022,
5
+ "exact_match_stderr,none": 0.004240317731917374,
6
+ "alias": "hendrycks_math"
7
+ },
8
+ "hendrycks_math_algebra": {
9
+ "alias": " - hendrycks_math_algebra",
10
+ "exact_match,none": 0.09856781802864364,
11
+ "exact_match_stderr,none": 0.008655491939014121
12
+ },
13
+ "hendrycks_math_counting_and_prob": {
14
+ "alias": " - hendrycks_math_counting_and_prob",
15
+ "exact_match,none": 0.10337552742616034,
16
+ "exact_match_stderr,none": 0.013998567410508813
17
+ },
18
+ "hendrycks_math_geometry": {
19
+ "alias": " - hendrycks_math_geometry",
20
+ "exact_match,none": 0.1022964509394572,
21
+ "exact_match_stderr,none": 0.013860624753065338
22
+ },
23
+ "hendrycks_math_intermediate_algebra": {
24
+ "alias": " - hendrycks_math_intermediate_algebra",
25
+ "exact_match,none": 0.06090808416389812,
26
+ "exact_match_stderr,none": 0.007963211522152501
27
+ },
28
+ "hendrycks_math_num_theory": {
29
+ "alias": " - hendrycks_math_num_theory",
30
+ "exact_match,none": 0.07777777777777778,
31
+ "exact_match_stderr,none": 0.011535891805294773
32
+ },
33
+ "hendrycks_math_prealgebra": {
34
+ "alias": " - hendrycks_math_prealgebra",
35
+ "exact_match,none": 0.19173363949483352,
36
+ "exact_match_stderr,none": 0.013346481347001473
37
+ },
38
+ "hendrycks_math_precalc": {
39
+ "alias": " - hendrycks_math_precalc",
40
+ "exact_match,none": 0.05860805860805861,
41
+ "exact_match_stderr,none": 0.010061567725278801
42
+ }
43
+ },
44
+ "groups": {
45
+ "hendrycks_math": {
46
+ "exact_match,none": 0.1022,
47
+ "exact_match_stderr,none": 0.004240317731917374,
48
+ "alias": "hendrycks_math"
49
+ }
50
+ },
51
+ "group_subtasks": {
52
+ "hendrycks_math": [
53
+ "hendrycks_math_algebra",
54
+ "hendrycks_math_counting_and_prob",
55
+ "hendrycks_math_geometry",
56
+ "hendrycks_math_intermediate_algebra",
57
+ "hendrycks_math_num_theory",
58
+ "hendrycks_math_prealgebra",
59
+ "hendrycks_math_precalc"
60
+ ]
61
+ },
62
+ "configs": {
63
+ "hendrycks_math_algebra": {
64
+ "task": "hendrycks_math_algebra",
65
+ "tag": [
66
+ "math_word_problems"
67
+ ],
68
+ "dataset_path": "EleutherAI/hendrycks_math",
69
+ "dataset_name": "algebra",
70
+ "training_split": "train",
71
+ "test_split": "test",
72
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
73
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
74
+ "doc_to_target": "{{answer}}",
75
+ "unsafe_code": false,
76
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
77
+ "description": "",
78
+ "target_delimiter": " ",
79
+ "fewshot_delimiter": "\n\n",
80
+ "num_fewshot": 4,
81
+ "metric_list": [
82
+ {
83
+ "metric": "exact_match",
84
+ "aggregation": "mean",
85
+ "higher_is_better": true
86
+ }
87
+ ],
88
+ "output_type": "generate_until",
89
+ "generation_kwargs": {
90
+ "until": [
91
+ "Problem:"
92
+ ],
93
+ "do_sample": false,
94
+ "temperature": 0.0
95
+ },
96
+ "repeats": 1,
97
+ "should_decontaminate": false,
98
+ "metadata": {
99
+ "version": 1.0,
100
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
101
+ }
102
+ },
103
+ "hendrycks_math_counting_and_prob": {
104
+ "task": "hendrycks_math_counting_and_prob",
105
+ "tag": [
106
+ "math_word_problems"
107
+ ],
108
+ "dataset_path": "EleutherAI/hendrycks_math",
109
+ "dataset_name": "counting_and_probability",
110
+ "training_split": "train",
111
+ "test_split": "test",
112
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
113
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
114
+ "doc_to_target": "{{answer}}",
115
+ "unsafe_code": false,
116
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
117
+ "description": "",
118
+ "target_delimiter": " ",
119
+ "fewshot_delimiter": "\n\n",
120
+ "num_fewshot": 4,
121
+ "metric_list": [
122
+ {
123
+ "metric": "exact_match",
124
+ "aggregation": "mean",
125
+ "higher_is_better": true
126
+ }
127
+ ],
128
+ "output_type": "generate_until",
129
+ "generation_kwargs": {
130
+ "until": [
131
+ "Problem:"
132
+ ],
133
+ "do_sample": false,
134
+ "temperature": 0.0
135
+ },
136
+ "repeats": 1,
137
+ "should_decontaminate": false,
138
+ "metadata": {
139
+ "version": 1.0,
140
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
141
+ }
142
+ },
143
+ "hendrycks_math_geometry": {
144
+ "task": "hendrycks_math_geometry",
145
+ "tag": [
146
+ "math_word_problems"
147
+ ],
148
+ "dataset_path": "EleutherAI/hendrycks_math",
149
+ "dataset_name": "geometry",
150
+ "training_split": "train",
151
+ "test_split": "test",
152
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
153
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
154
+ "doc_to_target": "{{answer}}",
155
+ "unsafe_code": false,
156
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
157
+ "description": "",
158
+ "target_delimiter": " ",
159
+ "fewshot_delimiter": "\n\n",
160
+ "num_fewshot": 4,
161
+ "metric_list": [
162
+ {
163
+ "metric": "exact_match",
164
+ "aggregation": "mean",
165
+ "higher_is_better": true
166
+ }
167
+ ],
168
+ "output_type": "generate_until",
169
+ "generation_kwargs": {
170
+ "until": [
171
+ "Problem:"
172
+ ],
173
+ "do_sample": false,
174
+ "temperature": 0.0
175
+ },
176
+ "repeats": 1,
177
+ "should_decontaminate": false,
178
+ "metadata": {
179
+ "version": 1.0,
180
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
181
+ }
182
+ },
183
+ "hendrycks_math_intermediate_algebra": {
184
+ "task": "hendrycks_math_intermediate_algebra",
185
+ "tag": [
186
+ "math_word_problems"
187
+ ],
188
+ "dataset_path": "EleutherAI/hendrycks_math",
189
+ "dataset_name": "intermediate_algebra",
190
+ "training_split": "train",
191
+ "test_split": "test",
192
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
193
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
194
+ "doc_to_target": "{{answer}}",
195
+ "unsafe_code": false,
196
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
197
+ "description": "",
198
+ "target_delimiter": " ",
199
+ "fewshot_delimiter": "\n\n",
200
+ "num_fewshot": 4,
201
+ "metric_list": [
202
+ {
203
+ "metric": "exact_match",
204
+ "aggregation": "mean",
205
+ "higher_is_better": true
206
+ }
207
+ ],
208
+ "output_type": "generate_until",
209
+ "generation_kwargs": {
210
+ "until": [
211
+ "Problem:"
212
+ ],
213
+ "do_sample": false,
214
+ "temperature": 0.0
215
+ },
216
+ "repeats": 1,
217
+ "should_decontaminate": false,
218
+ "metadata": {
219
+ "version": 1.0,
220
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
221
+ }
222
+ },
223
+ "hendrycks_math_num_theory": {
224
+ "task": "hendrycks_math_num_theory",
225
+ "tag": [
226
+ "math_word_problems"
227
+ ],
228
+ "dataset_path": "EleutherAI/hendrycks_math",
229
+ "dataset_name": "number_theory",
230
+ "training_split": "train",
231
+ "test_split": "test",
232
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
233
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
234
+ "doc_to_target": "{{answer}}",
235
+ "unsafe_code": false,
236
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
237
+ "description": "",
238
+ "target_delimiter": " ",
239
+ "fewshot_delimiter": "\n\n",
240
+ "num_fewshot": 4,
241
+ "metric_list": [
242
+ {
243
+ "metric": "exact_match",
244
+ "aggregation": "mean",
245
+ "higher_is_better": true
246
+ }
247
+ ],
248
+ "output_type": "generate_until",
249
+ "generation_kwargs": {
250
+ "until": [
251
+ "Problem:"
252
+ ],
253
+ "do_sample": false,
254
+ "temperature": 0.0
255
+ },
256
+ "repeats": 1,
257
+ "should_decontaminate": false,
258
+ "metadata": {
259
+ "version": 1.0,
260
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
261
+ }
262
+ },
263
+ "hendrycks_math_prealgebra": {
264
+ "task": "hendrycks_math_prealgebra",
265
+ "tag": [
266
+ "math_word_problems"
267
+ ],
268
+ "dataset_path": "EleutherAI/hendrycks_math",
269
+ "dataset_name": "prealgebra",
270
+ "training_split": "train",
271
+ "test_split": "test",
272
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
273
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
274
+ "doc_to_target": "{{answer}}",
275
+ "unsafe_code": false,
276
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
277
+ "description": "",
278
+ "target_delimiter": " ",
279
+ "fewshot_delimiter": "\n\n",
280
+ "num_fewshot": 4,
281
+ "metric_list": [
282
+ {
283
+ "metric": "exact_match",
284
+ "aggregation": "mean",
285
+ "higher_is_better": true
286
+ }
287
+ ],
288
+ "output_type": "generate_until",
289
+ "generation_kwargs": {
290
+ "until": [
291
+ "Problem:"
292
+ ],
293
+ "do_sample": false,
294
+ "temperature": 0.0
295
+ },
296
+ "repeats": 1,
297
+ "should_decontaminate": false,
298
+ "metadata": {
299
+ "version": 1.0,
300
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
301
+ }
302
+ },
303
+ "hendrycks_math_precalc": {
304
+ "task": "hendrycks_math_precalc",
305
+ "tag": [
306
+ "math_word_problems"
307
+ ],
308
+ "dataset_path": "EleutherAI/hendrycks_math",
309
+ "dataset_name": "precalculus",
310
+ "training_split": "train",
311
+ "test_split": "test",
312
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
313
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
314
+ "doc_to_target": "{{answer}}",
315
+ "unsafe_code": false,
316
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
317
+ "description": "",
318
+ "target_delimiter": " ",
319
+ "fewshot_delimiter": "\n\n",
320
+ "num_fewshot": 4,
321
+ "metric_list": [
322
+ {
323
+ "metric": "exact_match",
324
+ "aggregation": "mean",
325
+ "higher_is_better": true
326
+ }
327
+ ],
328
+ "output_type": "generate_until",
329
+ "generation_kwargs": {
330
+ "until": [
331
+ "Problem:"
332
+ ],
333
+ "do_sample": false,
334
+ "temperature": 0.0
335
+ },
336
+ "repeats": 1,
337
+ "should_decontaminate": false,
338
+ "metadata": {
339
+ "version": 1.0,
340
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
341
+ }
342
+ }
343
+ },
344
+ "versions": {
345
+ "hendrycks_math": 1.0,
346
+ "hendrycks_math_algebra": 1.0,
347
+ "hendrycks_math_counting_and_prob": 1.0,
348
+ "hendrycks_math_geometry": 1.0,
349
+ "hendrycks_math_intermediate_algebra": 1.0,
350
+ "hendrycks_math_num_theory": 1.0,
351
+ "hendrycks_math_prealgebra": 1.0,
352
+ "hendrycks_math_precalc": 1.0
353
+ },
354
+ "n-shot": {
355
+ "hendrycks_math_algebra": 4,
356
+ "hendrycks_math_counting_and_prob": 4,
357
+ "hendrycks_math_geometry": 4,
358
+ "hendrycks_math_intermediate_algebra": 4,
359
+ "hendrycks_math_num_theory": 4,
360
+ "hendrycks_math_prealgebra": 4,
361
+ "hendrycks_math_precalc": 4
362
+ },
363
+ "higher_is_better": {
364
+ "hendrycks_math": {
365
+ "exact_match": true
366
+ },
367
+ "hendrycks_math_algebra": {
368
+ "exact_match": true
369
+ },
370
+ "hendrycks_math_counting_and_prob": {
371
+ "exact_match": true
372
+ },
373
+ "hendrycks_math_geometry": {
374
+ "exact_match": true
375
+ },
376
+ "hendrycks_math_intermediate_algebra": {
377
+ "exact_match": true
378
+ },
379
+ "hendrycks_math_num_theory": {
380
+ "exact_match": true
381
+ },
382
+ "hendrycks_math_prealgebra": {
383
+ "exact_match": true
384
+ },
385
+ "hendrycks_math_precalc": {
386
+ "exact_match": true
387
+ }
388
+ },
389
+ "n-samples": {
390
+ "hendrycks_math_algebra": {
391
+ "original": 1187,
392
+ "effective": 1187
393
+ },
394
+ "hendrycks_math_counting_and_prob": {
395
+ "original": 474,
396
+ "effective": 474
397
+ },
398
+ "hendrycks_math_geometry": {
399
+ "original": 479,
400
+ "effective": 479
401
+ },
402
+ "hendrycks_math_intermediate_algebra": {
403
+ "original": 903,
404
+ "effective": 903
405
+ },
406
+ "hendrycks_math_num_theory": {
407
+ "original": 540,
408
+ "effective": 540
409
+ },
410
+ "hendrycks_math_prealgebra": {
411
+ "original": 871,
412
+ "effective": 871
413
+ },
414
+ "hendrycks_math_precalc": {
415
+ "original": 546,
416
+ "effective": 546
417
+ }
418
+ },
419
+ "config": {
420
+ "model": "hf",
421
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/12500_hf,trust_remote_code=True",
422
+ "model_num_parameters": 1235814400,
423
+ "model_dtype": "torch.bfloat16",
424
+ "model_revision": "main",
425
+ "model_sha": "",
426
+ "batch_size": "64",
427
+ "batch_sizes": [],
428
+ "device": null,
429
+ "use_cache": null,
430
+ "limit": null,
431
+ "bootstrap_iters": 100000,
432
+ "gen_kwargs": null,
433
+ "random_seed": 0,
434
+ "numpy_seed": 1234,
435
+ "torch_seed": 1234,
436
+ "fewshot_seed": 1234
437
+ },
438
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
439
+ "date": 1767757704.3666651,
440
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-77-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.129.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 48 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 96\nOn-line CPU(s) list: 0-95\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7352 24-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 24\nSocket(s): 2\nStepping: 0\nBogoMIPS: 4591.57\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca\nVirtualization: AMD-V\nL1d cache: 1.5 MiB (48 instances)\nL1i cache: 1.5 MiB (48 instances)\nL2 cache: 24 MiB (48 instances)\nL3 cache: 256 MiB (16 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-5,48-53\nNUMA node1 CPU(s): 6-11,54-59\nNUMA node2 CPU(s): 12-17,60-65\nNUMA node3 CPU(s): 18-23,66-71\nNUMA node4 CPU(s): 24-29,72-77\nNUMA node5 CPU(s): 30-35,78-83\nNUMA node6 CPU(s): 36-41,84-89\nNUMA node7 CPU(s): 42-47,90-95\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Full AMD retpoline, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
441
+ "transformers_version": "4.57.3",
442
+ "lm_eval_version": "0.4.9.1",
443
+ "upper_git_hash": null,
444
+ "tokenizer_pad_token": [
445
+ "<|end_of_text|>",
446
+ "128001"
447
+ ],
448
+ "tokenizer_eos_token": [
449
+ "<|end_of_text|>",
450
+ "128001"
451
+ ],
452
+ "tokenizer_bos_token": [
453
+ "<|begin_of_text|>",
454
+ "128000"
455
+ ],
456
+ "eot_token_id": 128001,
457
+ "max_length": 131072,
458
+ "task_hashes": {
459
+ "hendrycks_math_algebra": "7b50860e4db85f9d690d9679b543412837815a3b9854f0d0d166cfefc428cc23",
460
+ "hendrycks_math_counting_and_prob": "826f9744663d8e2e04e0050aa87b6edc5640c41ca1aa53fee8ef747f7c93b100",
461
+ "hendrycks_math_geometry": "4b67d6f45ba7f89a0c8a0a01fc9b1885bf942346a45b6a4fd03afd3d0e23ab6e",
462
+ "hendrycks_math_intermediate_algebra": "d7bc95882c4f58ef0b7252e6026e98b8bc77bbfbe36de6da6d97127e9744141e",
463
+ "hendrycks_math_num_theory": "ff82fd9f09e0a7c370270b3e7778773f7f1dec4aa40564e1a1318417af819030",
464
+ "hendrycks_math_prealgebra": "e4296905fe5f21ca452a0596371350ffedec630103ac7cce4b76cb806c55be05",
465
+ "hendrycks_math_precalc": "065216e339084dad7a3260b23a5151c8bf63771c7f0d23e88b6f3df4a9f0175a"
466
+ },
467
+ "model_source": "hf",
468
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf",
469
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__12500_hf",
470
+ "system_instruction": null,
471
+ "system_instruction_sha": null,
472
+ "fewshot_as_multiturn": false,
473
+ "chat_template": null,
474
+ "chat_template_sha": null,
475
+ "start_time": 607241.458764998,
476
+ "end_time": 607762.806025351,
477
+ "total_evaluation_time_seconds": "521.347260353039"
478
+ }
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_algebra_2026-01-07T03-56-06.178713.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_counting_and_prob_2026-01-07T03-56-06.178713.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_geometry_2026-01-07T03-56-06.178713.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_intermediate_algebra_2026-01-07T03-56-06.178713.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_num_theory_2026-01-07T03-56-06.178713.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_prealgebra_2026-01-07T03-56-06.178713.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_hendrycks_math_precalc_2026-01-07T03-56-06.178713.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/results_2026-01-07T07-12-17.705915.json ADDED
@@ -0,0 +1,290 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "gsm8k": {
4
+ "alias": "gsm8k",
5
+ "exact_match,strict-match": 0.20394238059135708,
6
+ "exact_match_stderr,strict-match": 0.011098602284899178,
7
+ "exact_match,flexible-extract": 0.20394238059135708,
8
+ "exact_match_stderr,flexible-extract": 0.011098602284899178
9
+ },
10
+ "gsm8k_cot": {
11
+ "alias": "gsm8k_cot",
12
+ "exact_match,strict-match": 0.20621683093252463,
13
+ "exact_match_stderr,strict-match": 0.011144364089781441,
14
+ "exact_match,flexible-extract": 0.2350265352539803,
15
+ "exact_match_stderr,flexible-extract": 0.011679491349994874
16
+ }
17
+ },
18
+ "group_subtasks": {
19
+ "gsm8k": [],
20
+ "gsm8k_cot": []
21
+ },
22
+ "configs": {
23
+ "gsm8k": {
24
+ "task": "gsm8k",
25
+ "tag": [
26
+ "math_word_problems"
27
+ ],
28
+ "dataset_path": "gsm8k",
29
+ "dataset_name": "main",
30
+ "training_split": "train",
31
+ "test_split": "test",
32
+ "fewshot_split": "train",
33
+ "doc_to_text": "Question: {{question}}\nAnswer:",
34
+ "doc_to_target": "{{answer}}",
35
+ "unsafe_code": false,
36
+ "description": "",
37
+ "target_delimiter": " ",
38
+ "fewshot_delimiter": "\n\n",
39
+ "num_fewshot": 8,
40
+ "metric_list": [
41
+ {
42
+ "metric": "exact_match",
43
+ "aggregation": "mean",
44
+ "higher_is_better": true,
45
+ "ignore_case": true,
46
+ "ignore_punctuation": false,
47
+ "regexes_to_ignore": [
48
+ ",",
49
+ "\\$",
50
+ "(?s).*#### ",
51
+ "\\.$"
52
+ ]
53
+ }
54
+ ],
55
+ "output_type": "generate_until",
56
+ "generation_kwargs": {
57
+ "until": [
58
+ "Question:",
59
+ "</s>",
60
+ "<|im_end|>"
61
+ ],
62
+ "do_sample": false,
63
+ "temperature": 0.0
64
+ },
65
+ "repeats": 1,
66
+ "filter_list": [
67
+ {
68
+ "name": "strict-match",
69
+ "filter": [
70
+ {
71
+ "function": "regex",
72
+ "regex_pattern": "#### (\\-?[0-9\\.\\,]+)"
73
+ },
74
+ {
75
+ "function": "take_first"
76
+ }
77
+ ]
78
+ },
79
+ {
80
+ "name": "flexible-extract",
81
+ "filter": [
82
+ {
83
+ "function": "regex",
84
+ "group_select": -1,
85
+ "regex_pattern": "(-?[$0-9.,]{2,})|(-?[0-9]+)"
86
+ },
87
+ {
88
+ "function": "take_first"
89
+ }
90
+ ]
91
+ }
92
+ ],
93
+ "should_decontaminate": false,
94
+ "metadata": {
95
+ "version": 3.0,
96
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
97
+ }
98
+ },
99
+ "gsm8k_cot": {
100
+ "task": "gsm8k_cot",
101
+ "tag": [
102
+ "chain_of_thought"
103
+ ],
104
+ "dataset_path": "gsm8k",
105
+ "dataset_name": "main",
106
+ "test_split": "test",
107
+ "doc_to_text": "Q: {{question}}\nA:",
108
+ "doc_to_target": "{{answer.split('####')[-1].strip() if answer is defined else target}}",
109
+ "unsafe_code": false,
110
+ "description": "",
111
+ "target_delimiter": " ",
112
+ "fewshot_delimiter": "\n\n",
113
+ "fewshot_config": {
114
+ "sampler": "first_n",
115
+ "samples": [
116
+ {
117
+ "question": "There are 15 trees in the grove. Grove workers will plant trees in the grove today. After they are done, there will be 21 trees. How many trees did the grove workers plant today?",
118
+ "target": "There are 15 trees originally. Then there were 21 trees after some more were planted. So there must have been 21 - 15 = 6. The answer is 6."
119
+ },
120
+ {
121
+ "question": "If there are 3 cars in the parking lot and 2 more cars arrive, how many cars are in the parking lot?",
122
+ "target": "There are originally 3 cars. 2 more cars arrive. 3 + 2 = 5. The answer is 5."
123
+ },
124
+ {
125
+ "question": "Leah had 32 chocolates and her sister had 42. If they ate 35, how many pieces do they have left in total?",
126
+ "target": "Originally, Leah had 32 chocolates. Her sister had 42. So in total they had 32 + 42 = 74. After eating 35, they had 74 - 35 = 39. The answer is 39."
127
+ },
128
+ {
129
+ "question": "Jason had 20 lollipops. He gave Denny some lollipops. Now Jason has 12 lollipops. How many lollipops did Jason give to Denny?",
130
+ "target": "Jason started with 20 lollipops. Then he had 12 after giving some to Denny. So he gave Denny 20 - 12 = 8. The answer is 8."
131
+ },
132
+ {
133
+ "question": "Shawn has five toys. For Christmas, he got two toys each from his mom and dad. How many toys does he have now?",
134
+ "target": "Shawn started with 5 toys. If he got 2 toys each from his mom and dad, then that is 4 more toys. 5 + 4 = 9. The answer is 9."
135
+ },
136
+ {
137
+ "question": "There were nine computers in the server room. Five more computers were installed each day, from monday to thursday. How many computers are now in the server room?",
138
+ "target": "There were originally 9 computers. For each of 4 days, 5 more computers were added. So 5 * 4 = 20 computers were added. 9 + 20 is 29. The answer is 29."
139
+ },
140
+ {
141
+ "question": "Michael had 58 golf balls. On tuesday, he lost 23 golf balls. On wednesday, he lost 2 more. How many golf balls did he have at the end of wednesday?",
142
+ "target": "Michael started with 58 golf balls. After losing 23 on tuesday, he had 58 - 23 = 35. After losing 2 more, he had 35 - 2 = 33 golf balls. The answer is 33."
143
+ },
144
+ {
145
+ "question": "Olivia has $23. She bought five bagels for $3 each. How much money does she have left?",
146
+ "target": "Olivia had 23 dollars. 5 bagels for 3 dollars each will be 5 x 3 = 15 dollars. So she has 23 - 15 dollars left. 23 - 15 is 8. The answer is 8."
147
+ }
148
+ ]
149
+ },
150
+ "num_fewshot": 8,
151
+ "metric_list": [
152
+ {
153
+ "aggregation": "mean",
154
+ "higher_is_better": true,
155
+ "ignore_case": true,
156
+ "ignore_punctuation": false,
157
+ "metric": "exact_match",
158
+ "regexes_to_ignore": [
159
+ ",",
160
+ "\\$",
161
+ "(?s).*#### ",
162
+ "\\.$"
163
+ ]
164
+ }
165
+ ],
166
+ "output_type": "generate_until",
167
+ "generation_kwargs": {
168
+ "do_sample": false,
169
+ "until": [
170
+ "Q:",
171
+ "</s>",
172
+ "<|im_end|>"
173
+ ]
174
+ },
175
+ "repeats": 1,
176
+ "filter_list": [
177
+ {
178
+ "filter": [
179
+ {
180
+ "function": "regex",
181
+ "regex_pattern": "The answer is (\\-?[0-9\\.\\,]+)."
182
+ },
183
+ {
184
+ "function": "take_first"
185
+ }
186
+ ],
187
+ "name": "strict-match"
188
+ },
189
+ {
190
+ "filter": [
191
+ {
192
+ "function": "regex",
193
+ "group_select": -1,
194
+ "regex_pattern": "(-?[$0-9.,]{2,})|(-?[0-9]+)"
195
+ },
196
+ {
197
+ "function": "take_first"
198
+ }
199
+ ],
200
+ "name": "flexible-extract"
201
+ }
202
+ ],
203
+ "should_decontaminate": false,
204
+ "metadata": {
205
+ "version": 3.0,
206
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf"
207
+ }
208
+ }
209
+ },
210
+ "versions": {
211
+ "gsm8k": 3.0,
212
+ "gsm8k_cot": 3.0
213
+ },
214
+ "n-shot": {
215
+ "gsm8k": 8,
216
+ "gsm8k_cot": 8
217
+ },
218
+ "higher_is_better": {
219
+ "gsm8k": {
220
+ "exact_match": true
221
+ },
222
+ "gsm8k_cot": {
223
+ "exact_match": true
224
+ }
225
+ },
226
+ "n-samples": {
227
+ "gsm8k_cot": {
228
+ "original": 1319,
229
+ "effective": 1319
230
+ },
231
+ "gsm8k": {
232
+ "original": 1319,
233
+ "effective": 1319
234
+ }
235
+ },
236
+ "config": {
237
+ "model": "hf",
238
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/12500_hf,trust_remote_code=True",
239
+ "model_num_parameters": 1235814400,
240
+ "model_dtype": "torch.bfloat16",
241
+ "model_revision": "main",
242
+ "model_sha": "",
243
+ "batch_size": "64",
244
+ "batch_sizes": [],
245
+ "device": null,
246
+ "use_cache": null,
247
+ "limit": null,
248
+ "bootstrap_iters": 100000,
249
+ "gen_kwargs": null,
250
+ "random_seed": 0,
251
+ "numpy_seed": 1234,
252
+ "torch_seed": 1234,
253
+ "fewshot_seed": 1234
254
+ },
255
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
256
+ "date": 1767769291.2161376,
257
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-72-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.129.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 48 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 96\nOn-line CPU(s) list: 0-95\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7352 24-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 24\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2300.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4591.26\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca\nVirtualization: AMD-V\nL1d cache: 1.5 MiB (48 instances)\nL1i cache: 1.5 MiB (48 instances)\nL2 cache: 24 MiB (48 instances)\nL3 cache: 256 MiB (16 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-5,48-53\nNUMA node1 CPU(s): 6-11,54-59\nNUMA node2 CPU(s): 12-17,60-65\nNUMA node3 CPU(s): 18-23,66-71\nNUMA node4 CPU(s): 24-29,72-77\nNUMA node5 CPU(s): 30-35,78-83\nNUMA node6 CPU(s): 36-41,84-89\nNUMA node7 CPU(s): 42-47,90-95\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Full AMD retpoline, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
258
+ "transformers_version": "4.57.3",
259
+ "lm_eval_version": "0.4.9.1",
260
+ "upper_git_hash": null,
261
+ "tokenizer_pad_token": [
262
+ "<|end_of_text|>",
263
+ "128001"
264
+ ],
265
+ "tokenizer_eos_token": [
266
+ "<|end_of_text|>",
267
+ "128001"
268
+ ],
269
+ "tokenizer_bos_token": [
270
+ "<|begin_of_text|>",
271
+ "128000"
272
+ ],
273
+ "eot_token_id": 128001,
274
+ "max_length": 131072,
275
+ "task_hashes": {
276
+ "gsm8k_cot": "fc360963b39ee52c26a82795124f9ad7da4d6a8fecf1b77e2502823b1669b3d0",
277
+ "gsm8k": "5edaa24ff4f3d939c3e1c5fd65a53cead84d4a52171818c453ec47099bd2a422"
278
+ },
279
+ "model_source": "hf",
280
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/12500_hf",
281
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__12500_hf",
282
+ "system_instruction": null,
283
+ "system_instruction_sha": null,
284
+ "fewshot_as_multiturn": false,
285
+ "chat_template": null,
286
+ "chat_template_sha": null,
287
+ "start_time": 10367669.561683849,
288
+ "end_time": 10368371.15969362,
289
+ "total_evaluation_time_seconds": "701.5980097707361"
290
+ }
12500_hf/harness_eval_8shot/__home__aiops__zhuty__nanotron__checkpoints__12500_hf/samples_gsm8k_cot_2026-01-07T07-12-17.705915.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
12500_hf/special_tokens_map.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|end_of_text|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ }
16
+ }
12500_hf/tokenizer_config.json ADDED
@@ -0,0 +1,2062 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "128000": {
4
+ "content": "<|begin_of_text|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "128001": {
12
+ "content": "<|end_of_text|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "128002": {
20
+ "content": "<|reserved_special_token_0|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "128003": {
28
+ "content": "<|reserved_special_token_1|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128004": {
36
+ "content": "<|finetune_right_pad_id|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "128005": {
44
+ "content": "<|reserved_special_token_2|>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "128006": {
52
+ "content": "<|start_header_id|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "128007": {
60
+ "content": "<|end_header_id|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "128008": {
68
+ "content": "<|eom_id|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "128009": {
76
+ "content": "<|eot_id|>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "128010": {
84
+ "content": "<|python_tag|>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "128011": {
92
+ "content": "<|reserved_special_token_3|>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "128012": {
100
+ "content": "<|reserved_special_token_4|>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "128013": {
108
+ "content": "<|reserved_special_token_5|>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "128014": {
116
+ "content": "<|reserved_special_token_6|>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "128015": {
124
+ "content": "<|reserved_special_token_7|>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "128016": {
132
+ "content": "<|reserved_special_token_8|>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "128017": {
140
+ "content": "<|reserved_special_token_9|>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "128018": {
148
+ "content": "<|reserved_special_token_10|>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "128019": {
156
+ "content": "<|reserved_special_token_11|>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "128020": {
164
+ "content": "<|reserved_special_token_12|>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "128021": {
172
+ "content": "<|reserved_special_token_13|>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "128022": {
180
+ "content": "<|reserved_special_token_14|>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "128023": {
188
+ "content": "<|reserved_special_token_15|>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "128024": {
196
+ "content": "<|reserved_special_token_16|>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "128025": {
204
+ "content": "<|reserved_special_token_17|>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "128026": {
212
+ "content": "<|reserved_special_token_18|>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "128027": {
220
+ "content": "<|reserved_special_token_19|>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "128028": {
228
+ "content": "<|reserved_special_token_20|>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "128029": {
236
+ "content": "<|reserved_special_token_21|>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "128030": {
244
+ "content": "<|reserved_special_token_22|>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "128031": {
252
+ "content": "<|reserved_special_token_23|>",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "128032": {
260
+ "content": "<|reserved_special_token_24|>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "128033": {
268
+ "content": "<|reserved_special_token_25|>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "128034": {
276
+ "content": "<|reserved_special_token_26|>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "128035": {
284
+ "content": "<|reserved_special_token_27|>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "128036": {
292
+ "content": "<|reserved_special_token_28|>",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "128037": {
300
+ "content": "<|reserved_special_token_29|>",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "128038": {
308
+ "content": "<|reserved_special_token_30|>",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "128039": {
316
+ "content": "<|reserved_special_token_31|>",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "128040": {
324
+ "content": "<|reserved_special_token_32|>",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "128041": {
332
+ "content": "<|reserved_special_token_33|>",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "128042": {
340
+ "content": "<|reserved_special_token_34|>",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "128043": {
348
+ "content": "<|reserved_special_token_35|>",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "128044": {
356
+ "content": "<|reserved_special_token_36|>",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "128045": {
364
+ "content": "<|reserved_special_token_37|>",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "128046": {
372
+ "content": "<|reserved_special_token_38|>",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "128047": {
380
+ "content": "<|reserved_special_token_39|>",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "128048": {
388
+ "content": "<|reserved_special_token_40|>",
389
+ "lstrip": false,
390
+ "normalized": false,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "128049": {
396
+ "content": "<|reserved_special_token_41|>",
397
+ "lstrip": false,
398
+ "normalized": false,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "128050": {
404
+ "content": "<|reserved_special_token_42|>",
405
+ "lstrip": false,
406
+ "normalized": false,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "128051": {
412
+ "content": "<|reserved_special_token_43|>",
413
+ "lstrip": false,
414
+ "normalized": false,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "128052": {
420
+ "content": "<|reserved_special_token_44|>",
421
+ "lstrip": false,
422
+ "normalized": false,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "128053": {
428
+ "content": "<|reserved_special_token_45|>",
429
+ "lstrip": false,
430
+ "normalized": false,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "128054": {
436
+ "content": "<|reserved_special_token_46|>",
437
+ "lstrip": false,
438
+ "normalized": false,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": true
442
+ },
443
+ "128055": {
444
+ "content": "<|reserved_special_token_47|>",
445
+ "lstrip": false,
446
+ "normalized": false,
447
+ "rstrip": false,
448
+ "single_word": false,
449
+ "special": true
450
+ },
451
+ "128056": {
452
+ "content": "<|reserved_special_token_48|>",
453
+ "lstrip": false,
454
+ "normalized": false,
455
+ "rstrip": false,
456
+ "single_word": false,
457
+ "special": true
458
+ },
459
+ "128057": {
460
+ "content": "<|reserved_special_token_49|>",
461
+ "lstrip": false,
462
+ "normalized": false,
463
+ "rstrip": false,
464
+ "single_word": false,
465
+ "special": true
466
+ },
467
+ "128058": {
468
+ "content": "<|reserved_special_token_50|>",
469
+ "lstrip": false,
470
+ "normalized": false,
471
+ "rstrip": false,
472
+ "single_word": false,
473
+ "special": true
474
+ },
475
+ "128059": {
476
+ "content": "<|reserved_special_token_51|>",
477
+ "lstrip": false,
478
+ "normalized": false,
479
+ "rstrip": false,
480
+ "single_word": false,
481
+ "special": true
482
+ },
483
+ "128060": {
484
+ "content": "<|reserved_special_token_52|>",
485
+ "lstrip": false,
486
+ "normalized": false,
487
+ "rstrip": false,
488
+ "single_word": false,
489
+ "special": true
490
+ },
491
+ "128061": {
492
+ "content": "<|reserved_special_token_53|>",
493
+ "lstrip": false,
494
+ "normalized": false,
495
+ "rstrip": false,
496
+ "single_word": false,
497
+ "special": true
498
+ },
499
+ "128062": {
500
+ "content": "<|reserved_special_token_54|>",
501
+ "lstrip": false,
502
+ "normalized": false,
503
+ "rstrip": false,
504
+ "single_word": false,
505
+ "special": true
506
+ },
507
+ "128063": {
508
+ "content": "<|reserved_special_token_55|>",
509
+ "lstrip": false,
510
+ "normalized": false,
511
+ "rstrip": false,
512
+ "single_word": false,
513
+ "special": true
514
+ },
515
+ "128064": {
516
+ "content": "<|reserved_special_token_56|>",
517
+ "lstrip": false,
518
+ "normalized": false,
519
+ "rstrip": false,
520
+ "single_word": false,
521
+ "special": true
522
+ },
523
+ "128065": {
524
+ "content": "<|reserved_special_token_57|>",
525
+ "lstrip": false,
526
+ "normalized": false,
527
+ "rstrip": false,
528
+ "single_word": false,
529
+ "special": true
530
+ },
531
+ "128066": {
532
+ "content": "<|reserved_special_token_58|>",
533
+ "lstrip": false,
534
+ "normalized": false,
535
+ "rstrip": false,
536
+ "single_word": false,
537
+ "special": true
538
+ },
539
+ "128067": {
540
+ "content": "<|reserved_special_token_59|>",
541
+ "lstrip": false,
542
+ "normalized": false,
543
+ "rstrip": false,
544
+ "single_word": false,
545
+ "special": true
546
+ },
547
+ "128068": {
548
+ "content": "<|reserved_special_token_60|>",
549
+ "lstrip": false,
550
+ "normalized": false,
551
+ "rstrip": false,
552
+ "single_word": false,
553
+ "special": true
554
+ },
555
+ "128069": {
556
+ "content": "<|reserved_special_token_61|>",
557
+ "lstrip": false,
558
+ "normalized": false,
559
+ "rstrip": false,
560
+ "single_word": false,
561
+ "special": true
562
+ },
563
+ "128070": {
564
+ "content": "<|reserved_special_token_62|>",
565
+ "lstrip": false,
566
+ "normalized": false,
567
+ "rstrip": false,
568
+ "single_word": false,
569
+ "special": true
570
+ },
571
+ "128071": {
572
+ "content": "<|reserved_special_token_63|>",
573
+ "lstrip": false,
574
+ "normalized": false,
575
+ "rstrip": false,
576
+ "single_word": false,
577
+ "special": true
578
+ },
579
+ "128072": {
580
+ "content": "<|reserved_special_token_64|>",
581
+ "lstrip": false,
582
+ "normalized": false,
583
+ "rstrip": false,
584
+ "single_word": false,
585
+ "special": true
586
+ },
587
+ "128073": {
588
+ "content": "<|reserved_special_token_65|>",
589
+ "lstrip": false,
590
+ "normalized": false,
591
+ "rstrip": false,
592
+ "single_word": false,
593
+ "special": true
594
+ },
595
+ "128074": {
596
+ "content": "<|reserved_special_token_66|>",
597
+ "lstrip": false,
598
+ "normalized": false,
599
+ "rstrip": false,
600
+ "single_word": false,
601
+ "special": true
602
+ },
603
+ "128075": {
604
+ "content": "<|reserved_special_token_67|>",
605
+ "lstrip": false,
606
+ "normalized": false,
607
+ "rstrip": false,
608
+ "single_word": false,
609
+ "special": true
610
+ },
611
+ "128076": {
612
+ "content": "<|reserved_special_token_68|>",
613
+ "lstrip": false,
614
+ "normalized": false,
615
+ "rstrip": false,
616
+ "single_word": false,
617
+ "special": true
618
+ },
619
+ "128077": {
620
+ "content": "<|reserved_special_token_69|>",
621
+ "lstrip": false,
622
+ "normalized": false,
623
+ "rstrip": false,
624
+ "single_word": false,
625
+ "special": true
626
+ },
627
+ "128078": {
628
+ "content": "<|reserved_special_token_70|>",
629
+ "lstrip": false,
630
+ "normalized": false,
631
+ "rstrip": false,
632
+ "single_word": false,
633
+ "special": true
634
+ },
635
+ "128079": {
636
+ "content": "<|reserved_special_token_71|>",
637
+ "lstrip": false,
638
+ "normalized": false,
639
+ "rstrip": false,
640
+ "single_word": false,
641
+ "special": true
642
+ },
643
+ "128080": {
644
+ "content": "<|reserved_special_token_72|>",
645
+ "lstrip": false,
646
+ "normalized": false,
647
+ "rstrip": false,
648
+ "single_word": false,
649
+ "special": true
650
+ },
651
+ "128081": {
652
+ "content": "<|reserved_special_token_73|>",
653
+ "lstrip": false,
654
+ "normalized": false,
655
+ "rstrip": false,
656
+ "single_word": false,
657
+ "special": true
658
+ },
659
+ "128082": {
660
+ "content": "<|reserved_special_token_74|>",
661
+ "lstrip": false,
662
+ "normalized": false,
663
+ "rstrip": false,
664
+ "single_word": false,
665
+ "special": true
666
+ },
667
+ "128083": {
668
+ "content": "<|reserved_special_token_75|>",
669
+ "lstrip": false,
670
+ "normalized": false,
671
+ "rstrip": false,
672
+ "single_word": false,
673
+ "special": true
674
+ },
675
+ "128084": {
676
+ "content": "<|reserved_special_token_76|>",
677
+ "lstrip": false,
678
+ "normalized": false,
679
+ "rstrip": false,
680
+ "single_word": false,
681
+ "special": true
682
+ },
683
+ "128085": {
684
+ "content": "<|reserved_special_token_77|>",
685
+ "lstrip": false,
686
+ "normalized": false,
687
+ "rstrip": false,
688
+ "single_word": false,
689
+ "special": true
690
+ },
691
+ "128086": {
692
+ "content": "<|reserved_special_token_78|>",
693
+ "lstrip": false,
694
+ "normalized": false,
695
+ "rstrip": false,
696
+ "single_word": false,
697
+ "special": true
698
+ },
699
+ "128087": {
700
+ "content": "<|reserved_special_token_79|>",
701
+ "lstrip": false,
702
+ "normalized": false,
703
+ "rstrip": false,
704
+ "single_word": false,
705
+ "special": true
706
+ },
707
+ "128088": {
708
+ "content": "<|reserved_special_token_80|>",
709
+ "lstrip": false,
710
+ "normalized": false,
711
+ "rstrip": false,
712
+ "single_word": false,
713
+ "special": true
714
+ },
715
+ "128089": {
716
+ "content": "<|reserved_special_token_81|>",
717
+ "lstrip": false,
718
+ "normalized": false,
719
+ "rstrip": false,
720
+ "single_word": false,
721
+ "special": true
722
+ },
723
+ "128090": {
724
+ "content": "<|reserved_special_token_82|>",
725
+ "lstrip": false,
726
+ "normalized": false,
727
+ "rstrip": false,
728
+ "single_word": false,
729
+ "special": true
730
+ },
731
+ "128091": {
732
+ "content": "<|reserved_special_token_83|>",
733
+ "lstrip": false,
734
+ "normalized": false,
735
+ "rstrip": false,
736
+ "single_word": false,
737
+ "special": true
738
+ },
739
+ "128092": {
740
+ "content": "<|reserved_special_token_84|>",
741
+ "lstrip": false,
742
+ "normalized": false,
743
+ "rstrip": false,
744
+ "single_word": false,
745
+ "special": true
746
+ },
747
+ "128093": {
748
+ "content": "<|reserved_special_token_85|>",
749
+ "lstrip": false,
750
+ "normalized": false,
751
+ "rstrip": false,
752
+ "single_word": false,
753
+ "special": true
754
+ },
755
+ "128094": {
756
+ "content": "<|reserved_special_token_86|>",
757
+ "lstrip": false,
758
+ "normalized": false,
759
+ "rstrip": false,
760
+ "single_word": false,
761
+ "special": true
762
+ },
763
+ "128095": {
764
+ "content": "<|reserved_special_token_87|>",
765
+ "lstrip": false,
766
+ "normalized": false,
767
+ "rstrip": false,
768
+ "single_word": false,
769
+ "special": true
770
+ },
771
+ "128096": {
772
+ "content": "<|reserved_special_token_88|>",
773
+ "lstrip": false,
774
+ "normalized": false,
775
+ "rstrip": false,
776
+ "single_word": false,
777
+ "special": true
778
+ },
779
+ "128097": {
780
+ "content": "<|reserved_special_token_89|>",
781
+ "lstrip": false,
782
+ "normalized": false,
783
+ "rstrip": false,
784
+ "single_word": false,
785
+ "special": true
786
+ },
787
+ "128098": {
788
+ "content": "<|reserved_special_token_90|>",
789
+ "lstrip": false,
790
+ "normalized": false,
791
+ "rstrip": false,
792
+ "single_word": false,
793
+ "special": true
794
+ },
795
+ "128099": {
796
+ "content": "<|reserved_special_token_91|>",
797
+ "lstrip": false,
798
+ "normalized": false,
799
+ "rstrip": false,
800
+ "single_word": false,
801
+ "special": true
802
+ },
803
+ "128100": {
804
+ "content": "<|reserved_special_token_92|>",
805
+ "lstrip": false,
806
+ "normalized": false,
807
+ "rstrip": false,
808
+ "single_word": false,
809
+ "special": true
810
+ },
811
+ "128101": {
812
+ "content": "<|reserved_special_token_93|>",
813
+ "lstrip": false,
814
+ "normalized": false,
815
+ "rstrip": false,
816
+ "single_word": false,
817
+ "special": true
818
+ },
819
+ "128102": {
820
+ "content": "<|reserved_special_token_94|>",
821
+ "lstrip": false,
822
+ "normalized": false,
823
+ "rstrip": false,
824
+ "single_word": false,
825
+ "special": true
826
+ },
827
+ "128103": {
828
+ "content": "<|reserved_special_token_95|>",
829
+ "lstrip": false,
830
+ "normalized": false,
831
+ "rstrip": false,
832
+ "single_word": false,
833
+ "special": true
834
+ },
835
+ "128104": {
836
+ "content": "<|reserved_special_token_96|>",
837
+ "lstrip": false,
838
+ "normalized": false,
839
+ "rstrip": false,
840
+ "single_word": false,
841
+ "special": true
842
+ },
843
+ "128105": {
844
+ "content": "<|reserved_special_token_97|>",
845
+ "lstrip": false,
846
+ "normalized": false,
847
+ "rstrip": false,
848
+ "single_word": false,
849
+ "special": true
850
+ },
851
+ "128106": {
852
+ "content": "<|reserved_special_token_98|>",
853
+ "lstrip": false,
854
+ "normalized": false,
855
+ "rstrip": false,
856
+ "single_word": false,
857
+ "special": true
858
+ },
859
+ "128107": {
860
+ "content": "<|reserved_special_token_99|>",
861
+ "lstrip": false,
862
+ "normalized": false,
863
+ "rstrip": false,
864
+ "single_word": false,
865
+ "special": true
866
+ },
867
+ "128108": {
868
+ "content": "<|reserved_special_token_100|>",
869
+ "lstrip": false,
870
+ "normalized": false,
871
+ "rstrip": false,
872
+ "single_word": false,
873
+ "special": true
874
+ },
875
+ "128109": {
876
+ "content": "<|reserved_special_token_101|>",
877
+ "lstrip": false,
878
+ "normalized": false,
879
+ "rstrip": false,
880
+ "single_word": false,
881
+ "special": true
882
+ },
883
+ "128110": {
884
+ "content": "<|reserved_special_token_102|>",
885
+ "lstrip": false,
886
+ "normalized": false,
887
+ "rstrip": false,
888
+ "single_word": false,
889
+ "special": true
890
+ },
891
+ "128111": {
892
+ "content": "<|reserved_special_token_103|>",
893
+ "lstrip": false,
894
+ "normalized": false,
895
+ "rstrip": false,
896
+ "single_word": false,
897
+ "special": true
898
+ },
899
+ "128112": {
900
+ "content": "<|reserved_special_token_104|>",
901
+ "lstrip": false,
902
+ "normalized": false,
903
+ "rstrip": false,
904
+ "single_word": false,
905
+ "special": true
906
+ },
907
+ "128113": {
908
+ "content": "<|reserved_special_token_105|>",
909
+ "lstrip": false,
910
+ "normalized": false,
911
+ "rstrip": false,
912
+ "single_word": false,
913
+ "special": true
914
+ },
915
+ "128114": {
916
+ "content": "<|reserved_special_token_106|>",
917
+ "lstrip": false,
918
+ "normalized": false,
919
+ "rstrip": false,
920
+ "single_word": false,
921
+ "special": true
922
+ },
923
+ "128115": {
924
+ "content": "<|reserved_special_token_107|>",
925
+ "lstrip": false,
926
+ "normalized": false,
927
+ "rstrip": false,
928
+ "single_word": false,
929
+ "special": true
930
+ },
931
+ "128116": {
932
+ "content": "<|reserved_special_token_108|>",
933
+ "lstrip": false,
934
+ "normalized": false,
935
+ "rstrip": false,
936
+ "single_word": false,
937
+ "special": true
938
+ },
939
+ "128117": {
940
+ "content": "<|reserved_special_token_109|>",
941
+ "lstrip": false,
942
+ "normalized": false,
943
+ "rstrip": false,
944
+ "single_word": false,
945
+ "special": true
946
+ },
947
+ "128118": {
948
+ "content": "<|reserved_special_token_110|>",
949
+ "lstrip": false,
950
+ "normalized": false,
951
+ "rstrip": false,
952
+ "single_word": false,
953
+ "special": true
954
+ },
955
+ "128119": {
956
+ "content": "<|reserved_special_token_111|>",
957
+ "lstrip": false,
958
+ "normalized": false,
959
+ "rstrip": false,
960
+ "single_word": false,
961
+ "special": true
962
+ },
963
+ "128120": {
964
+ "content": "<|reserved_special_token_112|>",
965
+ "lstrip": false,
966
+ "normalized": false,
967
+ "rstrip": false,
968
+ "single_word": false,
969
+ "special": true
970
+ },
971
+ "128121": {
972
+ "content": "<|reserved_special_token_113|>",
973
+ "lstrip": false,
974
+ "normalized": false,
975
+ "rstrip": false,
976
+ "single_word": false,
977
+ "special": true
978
+ },
979
+ "128122": {
980
+ "content": "<|reserved_special_token_114|>",
981
+ "lstrip": false,
982
+ "normalized": false,
983
+ "rstrip": false,
984
+ "single_word": false,
985
+ "special": true
986
+ },
987
+ "128123": {
988
+ "content": "<|reserved_special_token_115|>",
989
+ "lstrip": false,
990
+ "normalized": false,
991
+ "rstrip": false,
992
+ "single_word": false,
993
+ "special": true
994
+ },
995
+ "128124": {
996
+ "content": "<|reserved_special_token_116|>",
997
+ "lstrip": false,
998
+ "normalized": false,
999
+ "rstrip": false,
1000
+ "single_word": false,
1001
+ "special": true
1002
+ },
1003
+ "128125": {
1004
+ "content": "<|reserved_special_token_117|>",
1005
+ "lstrip": false,
1006
+ "normalized": false,
1007
+ "rstrip": false,
1008
+ "single_word": false,
1009
+ "special": true
1010
+ },
1011
+ "128126": {
1012
+ "content": "<|reserved_special_token_118|>",
1013
+ "lstrip": false,
1014
+ "normalized": false,
1015
+ "rstrip": false,
1016
+ "single_word": false,
1017
+ "special": true
1018
+ },
1019
+ "128127": {
1020
+ "content": "<|reserved_special_token_119|>",
1021
+ "lstrip": false,
1022
+ "normalized": false,
1023
+ "rstrip": false,
1024
+ "single_word": false,
1025
+ "special": true
1026
+ },
1027
+ "128128": {
1028
+ "content": "<|reserved_special_token_120|>",
1029
+ "lstrip": false,
1030
+ "normalized": false,
1031
+ "rstrip": false,
1032
+ "single_word": false,
1033
+ "special": true
1034
+ },
1035
+ "128129": {
1036
+ "content": "<|reserved_special_token_121|>",
1037
+ "lstrip": false,
1038
+ "normalized": false,
1039
+ "rstrip": false,
1040
+ "single_word": false,
1041
+ "special": true
1042
+ },
1043
+ "128130": {
1044
+ "content": "<|reserved_special_token_122|>",
1045
+ "lstrip": false,
1046
+ "normalized": false,
1047
+ "rstrip": false,
1048
+ "single_word": false,
1049
+ "special": true
1050
+ },
1051
+ "128131": {
1052
+ "content": "<|reserved_special_token_123|>",
1053
+ "lstrip": false,
1054
+ "normalized": false,
1055
+ "rstrip": false,
1056
+ "single_word": false,
1057
+ "special": true
1058
+ },
1059
+ "128132": {
1060
+ "content": "<|reserved_special_token_124|>",
1061
+ "lstrip": false,
1062
+ "normalized": false,
1063
+ "rstrip": false,
1064
+ "single_word": false,
1065
+ "special": true
1066
+ },
1067
+ "128133": {
1068
+ "content": "<|reserved_special_token_125|>",
1069
+ "lstrip": false,
1070
+ "normalized": false,
1071
+ "rstrip": false,
1072
+ "single_word": false,
1073
+ "special": true
1074
+ },
1075
+ "128134": {
1076
+ "content": "<|reserved_special_token_126|>",
1077
+ "lstrip": false,
1078
+ "normalized": false,
1079
+ "rstrip": false,
1080
+ "single_word": false,
1081
+ "special": true
1082
+ },
1083
+ "128135": {
1084
+ "content": "<|reserved_special_token_127|>",
1085
+ "lstrip": false,
1086
+ "normalized": false,
1087
+ "rstrip": false,
1088
+ "single_word": false,
1089
+ "special": true
1090
+ },
1091
+ "128136": {
1092
+ "content": "<|reserved_special_token_128|>",
1093
+ "lstrip": false,
1094
+ "normalized": false,
1095
+ "rstrip": false,
1096
+ "single_word": false,
1097
+ "special": true
1098
+ },
1099
+ "128137": {
1100
+ "content": "<|reserved_special_token_129|>",
1101
+ "lstrip": false,
1102
+ "normalized": false,
1103
+ "rstrip": false,
1104
+ "single_word": false,
1105
+ "special": true
1106
+ },
1107
+ "128138": {
1108
+ "content": "<|reserved_special_token_130|>",
1109
+ "lstrip": false,
1110
+ "normalized": false,
1111
+ "rstrip": false,
1112
+ "single_word": false,
1113
+ "special": true
1114
+ },
1115
+ "128139": {
1116
+ "content": "<|reserved_special_token_131|>",
1117
+ "lstrip": false,
1118
+ "normalized": false,
1119
+ "rstrip": false,
1120
+ "single_word": false,
1121
+ "special": true
1122
+ },
1123
+ "128140": {
1124
+ "content": "<|reserved_special_token_132|>",
1125
+ "lstrip": false,
1126
+ "normalized": false,
1127
+ "rstrip": false,
1128
+ "single_word": false,
1129
+ "special": true
1130
+ },
1131
+ "128141": {
1132
+ "content": "<|reserved_special_token_133|>",
1133
+ "lstrip": false,
1134
+ "normalized": false,
1135
+ "rstrip": false,
1136
+ "single_word": false,
1137
+ "special": true
1138
+ },
1139
+ "128142": {
1140
+ "content": "<|reserved_special_token_134|>",
1141
+ "lstrip": false,
1142
+ "normalized": false,
1143
+ "rstrip": false,
1144
+ "single_word": false,
1145
+ "special": true
1146
+ },
1147
+ "128143": {
1148
+ "content": "<|reserved_special_token_135|>",
1149
+ "lstrip": false,
1150
+ "normalized": false,
1151
+ "rstrip": false,
1152
+ "single_word": false,
1153
+ "special": true
1154
+ },
1155
+ "128144": {
1156
+ "content": "<|reserved_special_token_136|>",
1157
+ "lstrip": false,
1158
+ "normalized": false,
1159
+ "rstrip": false,
1160
+ "single_word": false,
1161
+ "special": true
1162
+ },
1163
+ "128145": {
1164
+ "content": "<|reserved_special_token_137|>",
1165
+ "lstrip": false,
1166
+ "normalized": false,
1167
+ "rstrip": false,
1168
+ "single_word": false,
1169
+ "special": true
1170
+ },
1171
+ "128146": {
1172
+ "content": "<|reserved_special_token_138|>",
1173
+ "lstrip": false,
1174
+ "normalized": false,
1175
+ "rstrip": false,
1176
+ "single_word": false,
1177
+ "special": true
1178
+ },
1179
+ "128147": {
1180
+ "content": "<|reserved_special_token_139|>",
1181
+ "lstrip": false,
1182
+ "normalized": false,
1183
+ "rstrip": false,
1184
+ "single_word": false,
1185
+ "special": true
1186
+ },
1187
+ "128148": {
1188
+ "content": "<|reserved_special_token_140|>",
1189
+ "lstrip": false,
1190
+ "normalized": false,
1191
+ "rstrip": false,
1192
+ "single_word": false,
1193
+ "special": true
1194
+ },
1195
+ "128149": {
1196
+ "content": "<|reserved_special_token_141|>",
1197
+ "lstrip": false,
1198
+ "normalized": false,
1199
+ "rstrip": false,
1200
+ "single_word": false,
1201
+ "special": true
1202
+ },
1203
+ "128150": {
1204
+ "content": "<|reserved_special_token_142|>",
1205
+ "lstrip": false,
1206
+ "normalized": false,
1207
+ "rstrip": false,
1208
+ "single_word": false,
1209
+ "special": true
1210
+ },
1211
+ "128151": {
1212
+ "content": "<|reserved_special_token_143|>",
1213
+ "lstrip": false,
1214
+ "normalized": false,
1215
+ "rstrip": false,
1216
+ "single_word": false,
1217
+ "special": true
1218
+ },
1219
+ "128152": {
1220
+ "content": "<|reserved_special_token_144|>",
1221
+ "lstrip": false,
1222
+ "normalized": false,
1223
+ "rstrip": false,
1224
+ "single_word": false,
1225
+ "special": true
1226
+ },
1227
+ "128153": {
1228
+ "content": "<|reserved_special_token_145|>",
1229
+ "lstrip": false,
1230
+ "normalized": false,
1231
+ "rstrip": false,
1232
+ "single_word": false,
1233
+ "special": true
1234
+ },
1235
+ "128154": {
1236
+ "content": "<|reserved_special_token_146|>",
1237
+ "lstrip": false,
1238
+ "normalized": false,
1239
+ "rstrip": false,
1240
+ "single_word": false,
1241
+ "special": true
1242
+ },
1243
+ "128155": {
1244
+ "content": "<|reserved_special_token_147|>",
1245
+ "lstrip": false,
1246
+ "normalized": false,
1247
+ "rstrip": false,
1248
+ "single_word": false,
1249
+ "special": true
1250
+ },
1251
+ "128156": {
1252
+ "content": "<|reserved_special_token_148|>",
1253
+ "lstrip": false,
1254
+ "normalized": false,
1255
+ "rstrip": false,
1256
+ "single_word": false,
1257
+ "special": true
1258
+ },
1259
+ "128157": {
1260
+ "content": "<|reserved_special_token_149|>",
1261
+ "lstrip": false,
1262
+ "normalized": false,
1263
+ "rstrip": false,
1264
+ "single_word": false,
1265
+ "special": true
1266
+ },
1267
+ "128158": {
1268
+ "content": "<|reserved_special_token_150|>",
1269
+ "lstrip": false,
1270
+ "normalized": false,
1271
+ "rstrip": false,
1272
+ "single_word": false,
1273
+ "special": true
1274
+ },
1275
+ "128159": {
1276
+ "content": "<|reserved_special_token_151|>",
1277
+ "lstrip": false,
1278
+ "normalized": false,
1279
+ "rstrip": false,
1280
+ "single_word": false,
1281
+ "special": true
1282
+ },
1283
+ "128160": {
1284
+ "content": "<|reserved_special_token_152|>",
1285
+ "lstrip": false,
1286
+ "normalized": false,
1287
+ "rstrip": false,
1288
+ "single_word": false,
1289
+ "special": true
1290
+ },
1291
+ "128161": {
1292
+ "content": "<|reserved_special_token_153|>",
1293
+ "lstrip": false,
1294
+ "normalized": false,
1295
+ "rstrip": false,
1296
+ "single_word": false,
1297
+ "special": true
1298
+ },
1299
+ "128162": {
1300
+ "content": "<|reserved_special_token_154|>",
1301
+ "lstrip": false,
1302
+ "normalized": false,
1303
+ "rstrip": false,
1304
+ "single_word": false,
1305
+ "special": true
1306
+ },
1307
+ "128163": {
1308
+ "content": "<|reserved_special_token_155|>",
1309
+ "lstrip": false,
1310
+ "normalized": false,
1311
+ "rstrip": false,
1312
+ "single_word": false,
1313
+ "special": true
1314
+ },
1315
+ "128164": {
1316
+ "content": "<|reserved_special_token_156|>",
1317
+ "lstrip": false,
1318
+ "normalized": false,
1319
+ "rstrip": false,
1320
+ "single_word": false,
1321
+ "special": true
1322
+ },
1323
+ "128165": {
1324
+ "content": "<|reserved_special_token_157|>",
1325
+ "lstrip": false,
1326
+ "normalized": false,
1327
+ "rstrip": false,
1328
+ "single_word": false,
1329
+ "special": true
1330
+ },
1331
+ "128166": {
1332
+ "content": "<|reserved_special_token_158|>",
1333
+ "lstrip": false,
1334
+ "normalized": false,
1335
+ "rstrip": false,
1336
+ "single_word": false,
1337
+ "special": true
1338
+ },
1339
+ "128167": {
1340
+ "content": "<|reserved_special_token_159|>",
1341
+ "lstrip": false,
1342
+ "normalized": false,
1343
+ "rstrip": false,
1344
+ "single_word": false,
1345
+ "special": true
1346
+ },
1347
+ "128168": {
1348
+ "content": "<|reserved_special_token_160|>",
1349
+ "lstrip": false,
1350
+ "normalized": false,
1351
+ "rstrip": false,
1352
+ "single_word": false,
1353
+ "special": true
1354
+ },
1355
+ "128169": {
1356
+ "content": "<|reserved_special_token_161|>",
1357
+ "lstrip": false,
1358
+ "normalized": false,
1359
+ "rstrip": false,
1360
+ "single_word": false,
1361
+ "special": true
1362
+ },
1363
+ "128170": {
1364
+ "content": "<|reserved_special_token_162|>",
1365
+ "lstrip": false,
1366
+ "normalized": false,
1367
+ "rstrip": false,
1368
+ "single_word": false,
1369
+ "special": true
1370
+ },
1371
+ "128171": {
1372
+ "content": "<|reserved_special_token_163|>",
1373
+ "lstrip": false,
1374
+ "normalized": false,
1375
+ "rstrip": false,
1376
+ "single_word": false,
1377
+ "special": true
1378
+ },
1379
+ "128172": {
1380
+ "content": "<|reserved_special_token_164|>",
1381
+ "lstrip": false,
1382
+ "normalized": false,
1383
+ "rstrip": false,
1384
+ "single_word": false,
1385
+ "special": true
1386
+ },
1387
+ "128173": {
1388
+ "content": "<|reserved_special_token_165|>",
1389
+ "lstrip": false,
1390
+ "normalized": false,
1391
+ "rstrip": false,
1392
+ "single_word": false,
1393
+ "special": true
1394
+ },
1395
+ "128174": {
1396
+ "content": "<|reserved_special_token_166|>",
1397
+ "lstrip": false,
1398
+ "normalized": false,
1399
+ "rstrip": false,
1400
+ "single_word": false,
1401
+ "special": true
1402
+ },
1403
+ "128175": {
1404
+ "content": "<|reserved_special_token_167|>",
1405
+ "lstrip": false,
1406
+ "normalized": false,
1407
+ "rstrip": false,
1408
+ "single_word": false,
1409
+ "special": true
1410
+ },
1411
+ "128176": {
1412
+ "content": "<|reserved_special_token_168|>",
1413
+ "lstrip": false,
1414
+ "normalized": false,
1415
+ "rstrip": false,
1416
+ "single_word": false,
1417
+ "special": true
1418
+ },
1419
+ "128177": {
1420
+ "content": "<|reserved_special_token_169|>",
1421
+ "lstrip": false,
1422
+ "normalized": false,
1423
+ "rstrip": false,
1424
+ "single_word": false,
1425
+ "special": true
1426
+ },
1427
+ "128178": {
1428
+ "content": "<|reserved_special_token_170|>",
1429
+ "lstrip": false,
1430
+ "normalized": false,
1431
+ "rstrip": false,
1432
+ "single_word": false,
1433
+ "special": true
1434
+ },
1435
+ "128179": {
1436
+ "content": "<|reserved_special_token_171|>",
1437
+ "lstrip": false,
1438
+ "normalized": false,
1439
+ "rstrip": false,
1440
+ "single_word": false,
1441
+ "special": true
1442
+ },
1443
+ "128180": {
1444
+ "content": "<|reserved_special_token_172|>",
1445
+ "lstrip": false,
1446
+ "normalized": false,
1447
+ "rstrip": false,
1448
+ "single_word": false,
1449
+ "special": true
1450
+ },
1451
+ "128181": {
1452
+ "content": "<|reserved_special_token_173|>",
1453
+ "lstrip": false,
1454
+ "normalized": false,
1455
+ "rstrip": false,
1456
+ "single_word": false,
1457
+ "special": true
1458
+ },
1459
+ "128182": {
1460
+ "content": "<|reserved_special_token_174|>",
1461
+ "lstrip": false,
1462
+ "normalized": false,
1463
+ "rstrip": false,
1464
+ "single_word": false,
1465
+ "special": true
1466
+ },
1467
+ "128183": {
1468
+ "content": "<|reserved_special_token_175|>",
1469
+ "lstrip": false,
1470
+ "normalized": false,
1471
+ "rstrip": false,
1472
+ "single_word": false,
1473
+ "special": true
1474
+ },
1475
+ "128184": {
1476
+ "content": "<|reserved_special_token_176|>",
1477
+ "lstrip": false,
1478
+ "normalized": false,
1479
+ "rstrip": false,
1480
+ "single_word": false,
1481
+ "special": true
1482
+ },
1483
+ "128185": {
1484
+ "content": "<|reserved_special_token_177|>",
1485
+ "lstrip": false,
1486
+ "normalized": false,
1487
+ "rstrip": false,
1488
+ "single_word": false,
1489
+ "special": true
1490
+ },
1491
+ "128186": {
1492
+ "content": "<|reserved_special_token_178|>",
1493
+ "lstrip": false,
1494
+ "normalized": false,
1495
+ "rstrip": false,
1496
+ "single_word": false,
1497
+ "special": true
1498
+ },
1499
+ "128187": {
1500
+ "content": "<|reserved_special_token_179|>",
1501
+ "lstrip": false,
1502
+ "normalized": false,
1503
+ "rstrip": false,
1504
+ "single_word": false,
1505
+ "special": true
1506
+ },
1507
+ "128188": {
1508
+ "content": "<|reserved_special_token_180|>",
1509
+ "lstrip": false,
1510
+ "normalized": false,
1511
+ "rstrip": false,
1512
+ "single_word": false,
1513
+ "special": true
1514
+ },
1515
+ "128189": {
1516
+ "content": "<|reserved_special_token_181|>",
1517
+ "lstrip": false,
1518
+ "normalized": false,
1519
+ "rstrip": false,
1520
+ "single_word": false,
1521
+ "special": true
1522
+ },
1523
+ "128190": {
1524
+ "content": "<|reserved_special_token_182|>",
1525
+ "lstrip": false,
1526
+ "normalized": false,
1527
+ "rstrip": false,
1528
+ "single_word": false,
1529
+ "special": true
1530
+ },
1531
+ "128191": {
1532
+ "content": "<|reserved_special_token_183|>",
1533
+ "lstrip": false,
1534
+ "normalized": false,
1535
+ "rstrip": false,
1536
+ "single_word": false,
1537
+ "special": true
1538
+ },
1539
+ "128192": {
1540
+ "content": "<|reserved_special_token_184|>",
1541
+ "lstrip": false,
1542
+ "normalized": false,
1543
+ "rstrip": false,
1544
+ "single_word": false,
1545
+ "special": true
1546
+ },
1547
+ "128193": {
1548
+ "content": "<|reserved_special_token_185|>",
1549
+ "lstrip": false,
1550
+ "normalized": false,
1551
+ "rstrip": false,
1552
+ "single_word": false,
1553
+ "special": true
1554
+ },
1555
+ "128194": {
1556
+ "content": "<|reserved_special_token_186|>",
1557
+ "lstrip": false,
1558
+ "normalized": false,
1559
+ "rstrip": false,
1560
+ "single_word": false,
1561
+ "special": true
1562
+ },
1563
+ "128195": {
1564
+ "content": "<|reserved_special_token_187|>",
1565
+ "lstrip": false,
1566
+ "normalized": false,
1567
+ "rstrip": false,
1568
+ "single_word": false,
1569
+ "special": true
1570
+ },
1571
+ "128196": {
1572
+ "content": "<|reserved_special_token_188|>",
1573
+ "lstrip": false,
1574
+ "normalized": false,
1575
+ "rstrip": false,
1576
+ "single_word": false,
1577
+ "special": true
1578
+ },
1579
+ "128197": {
1580
+ "content": "<|reserved_special_token_189|>",
1581
+ "lstrip": false,
1582
+ "normalized": false,
1583
+ "rstrip": false,
1584
+ "single_word": false,
1585
+ "special": true
1586
+ },
1587
+ "128198": {
1588
+ "content": "<|reserved_special_token_190|>",
1589
+ "lstrip": false,
1590
+ "normalized": false,
1591
+ "rstrip": false,
1592
+ "single_word": false,
1593
+ "special": true
1594
+ },
1595
+ "128199": {
1596
+ "content": "<|reserved_special_token_191|>",
1597
+ "lstrip": false,
1598
+ "normalized": false,
1599
+ "rstrip": false,
1600
+ "single_word": false,
1601
+ "special": true
1602
+ },
1603
+ "128200": {
1604
+ "content": "<|reserved_special_token_192|>",
1605
+ "lstrip": false,
1606
+ "normalized": false,
1607
+ "rstrip": false,
1608
+ "single_word": false,
1609
+ "special": true
1610
+ },
1611
+ "128201": {
1612
+ "content": "<|reserved_special_token_193|>",
1613
+ "lstrip": false,
1614
+ "normalized": false,
1615
+ "rstrip": false,
1616
+ "single_word": false,
1617
+ "special": true
1618
+ },
1619
+ "128202": {
1620
+ "content": "<|reserved_special_token_194|>",
1621
+ "lstrip": false,
1622
+ "normalized": false,
1623
+ "rstrip": false,
1624
+ "single_word": false,
1625
+ "special": true
1626
+ },
1627
+ "128203": {
1628
+ "content": "<|reserved_special_token_195|>",
1629
+ "lstrip": false,
1630
+ "normalized": false,
1631
+ "rstrip": false,
1632
+ "single_word": false,
1633
+ "special": true
1634
+ },
1635
+ "128204": {
1636
+ "content": "<|reserved_special_token_196|>",
1637
+ "lstrip": false,
1638
+ "normalized": false,
1639
+ "rstrip": false,
1640
+ "single_word": false,
1641
+ "special": true
1642
+ },
1643
+ "128205": {
1644
+ "content": "<|reserved_special_token_197|>",
1645
+ "lstrip": false,
1646
+ "normalized": false,
1647
+ "rstrip": false,
1648
+ "single_word": false,
1649
+ "special": true
1650
+ },
1651
+ "128206": {
1652
+ "content": "<|reserved_special_token_198|>",
1653
+ "lstrip": false,
1654
+ "normalized": false,
1655
+ "rstrip": false,
1656
+ "single_word": false,
1657
+ "special": true
1658
+ },
1659
+ "128207": {
1660
+ "content": "<|reserved_special_token_199|>",
1661
+ "lstrip": false,
1662
+ "normalized": false,
1663
+ "rstrip": false,
1664
+ "single_word": false,
1665
+ "special": true
1666
+ },
1667
+ "128208": {
1668
+ "content": "<|reserved_special_token_200|>",
1669
+ "lstrip": false,
1670
+ "normalized": false,
1671
+ "rstrip": false,
1672
+ "single_word": false,
1673
+ "special": true
1674
+ },
1675
+ "128209": {
1676
+ "content": "<|reserved_special_token_201|>",
1677
+ "lstrip": false,
1678
+ "normalized": false,
1679
+ "rstrip": false,
1680
+ "single_word": false,
1681
+ "special": true
1682
+ },
1683
+ "128210": {
1684
+ "content": "<|reserved_special_token_202|>",
1685
+ "lstrip": false,
1686
+ "normalized": false,
1687
+ "rstrip": false,
1688
+ "single_word": false,
1689
+ "special": true
1690
+ },
1691
+ "128211": {
1692
+ "content": "<|reserved_special_token_203|>",
1693
+ "lstrip": false,
1694
+ "normalized": false,
1695
+ "rstrip": false,
1696
+ "single_word": false,
1697
+ "special": true
1698
+ },
1699
+ "128212": {
1700
+ "content": "<|reserved_special_token_204|>",
1701
+ "lstrip": false,
1702
+ "normalized": false,
1703
+ "rstrip": false,
1704
+ "single_word": false,
1705
+ "special": true
1706
+ },
1707
+ "128213": {
1708
+ "content": "<|reserved_special_token_205|>",
1709
+ "lstrip": false,
1710
+ "normalized": false,
1711
+ "rstrip": false,
1712
+ "single_word": false,
1713
+ "special": true
1714
+ },
1715
+ "128214": {
1716
+ "content": "<|reserved_special_token_206|>",
1717
+ "lstrip": false,
1718
+ "normalized": false,
1719
+ "rstrip": false,
1720
+ "single_word": false,
1721
+ "special": true
1722
+ },
1723
+ "128215": {
1724
+ "content": "<|reserved_special_token_207|>",
1725
+ "lstrip": false,
1726
+ "normalized": false,
1727
+ "rstrip": false,
1728
+ "single_word": false,
1729
+ "special": true
1730
+ },
1731
+ "128216": {
1732
+ "content": "<|reserved_special_token_208|>",
1733
+ "lstrip": false,
1734
+ "normalized": false,
1735
+ "rstrip": false,
1736
+ "single_word": false,
1737
+ "special": true
1738
+ },
1739
+ "128217": {
1740
+ "content": "<|reserved_special_token_209|>",
1741
+ "lstrip": false,
1742
+ "normalized": false,
1743
+ "rstrip": false,
1744
+ "single_word": false,
1745
+ "special": true
1746
+ },
1747
+ "128218": {
1748
+ "content": "<|reserved_special_token_210|>",
1749
+ "lstrip": false,
1750
+ "normalized": false,
1751
+ "rstrip": false,
1752
+ "single_word": false,
1753
+ "special": true
1754
+ },
1755
+ "128219": {
1756
+ "content": "<|reserved_special_token_211|>",
1757
+ "lstrip": false,
1758
+ "normalized": false,
1759
+ "rstrip": false,
1760
+ "single_word": false,
1761
+ "special": true
1762
+ },
1763
+ "128220": {
1764
+ "content": "<|reserved_special_token_212|>",
1765
+ "lstrip": false,
1766
+ "normalized": false,
1767
+ "rstrip": false,
1768
+ "single_word": false,
1769
+ "special": true
1770
+ },
1771
+ "128221": {
1772
+ "content": "<|reserved_special_token_213|>",
1773
+ "lstrip": false,
1774
+ "normalized": false,
1775
+ "rstrip": false,
1776
+ "single_word": false,
1777
+ "special": true
1778
+ },
1779
+ "128222": {
1780
+ "content": "<|reserved_special_token_214|>",
1781
+ "lstrip": false,
1782
+ "normalized": false,
1783
+ "rstrip": false,
1784
+ "single_word": false,
1785
+ "special": true
1786
+ },
1787
+ "128223": {
1788
+ "content": "<|reserved_special_token_215|>",
1789
+ "lstrip": false,
1790
+ "normalized": false,
1791
+ "rstrip": false,
1792
+ "single_word": false,
1793
+ "special": true
1794
+ },
1795
+ "128224": {
1796
+ "content": "<|reserved_special_token_216|>",
1797
+ "lstrip": false,
1798
+ "normalized": false,
1799
+ "rstrip": false,
1800
+ "single_word": false,
1801
+ "special": true
1802
+ },
1803
+ "128225": {
1804
+ "content": "<|reserved_special_token_217|>",
1805
+ "lstrip": false,
1806
+ "normalized": false,
1807
+ "rstrip": false,
1808
+ "single_word": false,
1809
+ "special": true
1810
+ },
1811
+ "128226": {
1812
+ "content": "<|reserved_special_token_218|>",
1813
+ "lstrip": false,
1814
+ "normalized": false,
1815
+ "rstrip": false,
1816
+ "single_word": false,
1817
+ "special": true
1818
+ },
1819
+ "128227": {
1820
+ "content": "<|reserved_special_token_219|>",
1821
+ "lstrip": false,
1822
+ "normalized": false,
1823
+ "rstrip": false,
1824
+ "single_word": false,
1825
+ "special": true
1826
+ },
1827
+ "128228": {
1828
+ "content": "<|reserved_special_token_220|>",
1829
+ "lstrip": false,
1830
+ "normalized": false,
1831
+ "rstrip": false,
1832
+ "single_word": false,
1833
+ "special": true
1834
+ },
1835
+ "128229": {
1836
+ "content": "<|reserved_special_token_221|>",
1837
+ "lstrip": false,
1838
+ "normalized": false,
1839
+ "rstrip": false,
1840
+ "single_word": false,
1841
+ "special": true
1842
+ },
1843
+ "128230": {
1844
+ "content": "<|reserved_special_token_222|>",
1845
+ "lstrip": false,
1846
+ "normalized": false,
1847
+ "rstrip": false,
1848
+ "single_word": false,
1849
+ "special": true
1850
+ },
1851
+ "128231": {
1852
+ "content": "<|reserved_special_token_223|>",
1853
+ "lstrip": false,
1854
+ "normalized": false,
1855
+ "rstrip": false,
1856
+ "single_word": false,
1857
+ "special": true
1858
+ },
1859
+ "128232": {
1860
+ "content": "<|reserved_special_token_224|>",
1861
+ "lstrip": false,
1862
+ "normalized": false,
1863
+ "rstrip": false,
1864
+ "single_word": false,
1865
+ "special": true
1866
+ },
1867
+ "128233": {
1868
+ "content": "<|reserved_special_token_225|>",
1869
+ "lstrip": false,
1870
+ "normalized": false,
1871
+ "rstrip": false,
1872
+ "single_word": false,
1873
+ "special": true
1874
+ },
1875
+ "128234": {
1876
+ "content": "<|reserved_special_token_226|>",
1877
+ "lstrip": false,
1878
+ "normalized": false,
1879
+ "rstrip": false,
1880
+ "single_word": false,
1881
+ "special": true
1882
+ },
1883
+ "128235": {
1884
+ "content": "<|reserved_special_token_227|>",
1885
+ "lstrip": false,
1886
+ "normalized": false,
1887
+ "rstrip": false,
1888
+ "single_word": false,
1889
+ "special": true
1890
+ },
1891
+ "128236": {
1892
+ "content": "<|reserved_special_token_228|>",
1893
+ "lstrip": false,
1894
+ "normalized": false,
1895
+ "rstrip": false,
1896
+ "single_word": false,
1897
+ "special": true
1898
+ },
1899
+ "128237": {
1900
+ "content": "<|reserved_special_token_229|>",
1901
+ "lstrip": false,
1902
+ "normalized": false,
1903
+ "rstrip": false,
1904
+ "single_word": false,
1905
+ "special": true
1906
+ },
1907
+ "128238": {
1908
+ "content": "<|reserved_special_token_230|>",
1909
+ "lstrip": false,
1910
+ "normalized": false,
1911
+ "rstrip": false,
1912
+ "single_word": false,
1913
+ "special": true
1914
+ },
1915
+ "128239": {
1916
+ "content": "<|reserved_special_token_231|>",
1917
+ "lstrip": false,
1918
+ "normalized": false,
1919
+ "rstrip": false,
1920
+ "single_word": false,
1921
+ "special": true
1922
+ },
1923
+ "128240": {
1924
+ "content": "<|reserved_special_token_232|>",
1925
+ "lstrip": false,
1926
+ "normalized": false,
1927
+ "rstrip": false,
1928
+ "single_word": false,
1929
+ "special": true
1930
+ },
1931
+ "128241": {
1932
+ "content": "<|reserved_special_token_233|>",
1933
+ "lstrip": false,
1934
+ "normalized": false,
1935
+ "rstrip": false,
1936
+ "single_word": false,
1937
+ "special": true
1938
+ },
1939
+ "128242": {
1940
+ "content": "<|reserved_special_token_234|>",
1941
+ "lstrip": false,
1942
+ "normalized": false,
1943
+ "rstrip": false,
1944
+ "single_word": false,
1945
+ "special": true
1946
+ },
1947
+ "128243": {
1948
+ "content": "<|reserved_special_token_235|>",
1949
+ "lstrip": false,
1950
+ "normalized": false,
1951
+ "rstrip": false,
1952
+ "single_word": false,
1953
+ "special": true
1954
+ },
1955
+ "128244": {
1956
+ "content": "<|reserved_special_token_236|>",
1957
+ "lstrip": false,
1958
+ "normalized": false,
1959
+ "rstrip": false,
1960
+ "single_word": false,
1961
+ "special": true
1962
+ },
1963
+ "128245": {
1964
+ "content": "<|reserved_special_token_237|>",
1965
+ "lstrip": false,
1966
+ "normalized": false,
1967
+ "rstrip": false,
1968
+ "single_word": false,
1969
+ "special": true
1970
+ },
1971
+ "128246": {
1972
+ "content": "<|reserved_special_token_238|>",
1973
+ "lstrip": false,
1974
+ "normalized": false,
1975
+ "rstrip": false,
1976
+ "single_word": false,
1977
+ "special": true
1978
+ },
1979
+ "128247": {
1980
+ "content": "<|reserved_special_token_239|>",
1981
+ "lstrip": false,
1982
+ "normalized": false,
1983
+ "rstrip": false,
1984
+ "single_word": false,
1985
+ "special": true
1986
+ },
1987
+ "128248": {
1988
+ "content": "<|reserved_special_token_240|>",
1989
+ "lstrip": false,
1990
+ "normalized": false,
1991
+ "rstrip": false,
1992
+ "single_word": false,
1993
+ "special": true
1994
+ },
1995
+ "128249": {
1996
+ "content": "<|reserved_special_token_241|>",
1997
+ "lstrip": false,
1998
+ "normalized": false,
1999
+ "rstrip": false,
2000
+ "single_word": false,
2001
+ "special": true
2002
+ },
2003
+ "128250": {
2004
+ "content": "<|reserved_special_token_242|>",
2005
+ "lstrip": false,
2006
+ "normalized": false,
2007
+ "rstrip": false,
2008
+ "single_word": false,
2009
+ "special": true
2010
+ },
2011
+ "128251": {
2012
+ "content": "<|reserved_special_token_243|>",
2013
+ "lstrip": false,
2014
+ "normalized": false,
2015
+ "rstrip": false,
2016
+ "single_word": false,
2017
+ "special": true
2018
+ },
2019
+ "128252": {
2020
+ "content": "<|reserved_special_token_244|>",
2021
+ "lstrip": false,
2022
+ "normalized": false,
2023
+ "rstrip": false,
2024
+ "single_word": false,
2025
+ "special": true
2026
+ },
2027
+ "128253": {
2028
+ "content": "<|reserved_special_token_245|>",
2029
+ "lstrip": false,
2030
+ "normalized": false,
2031
+ "rstrip": false,
2032
+ "single_word": false,
2033
+ "special": true
2034
+ },
2035
+ "128254": {
2036
+ "content": "<|reserved_special_token_246|>",
2037
+ "lstrip": false,
2038
+ "normalized": false,
2039
+ "rstrip": false,
2040
+ "single_word": false,
2041
+ "special": true
2042
+ },
2043
+ "128255": {
2044
+ "content": "<|reserved_special_token_247|>",
2045
+ "lstrip": false,
2046
+ "normalized": false,
2047
+ "rstrip": false,
2048
+ "single_word": false,
2049
+ "special": true
2050
+ }
2051
+ },
2052
+ "bos_token": "<|begin_of_text|>",
2053
+ "clean_up_tokenization_spaces": true,
2054
+ "eos_token": "<|end_of_text|>",
2055
+ "extra_special_tokens": {},
2056
+ "model_input_names": [
2057
+ "input_ids",
2058
+ "attention_mask"
2059
+ ],
2060
+ "model_max_length": 131072,
2061
+ "tokenizer_class": "PreTrainedTokenizerFast"
2062
+ }
2500_hf/config.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 128000,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 128001,
10
+ "head_dim": 64,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 2048,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 8192,
15
+ "max_position_embeddings": 131072,
16
+ "mlp_bias": false,
17
+ "model_type": "llama",
18
+ "num_attention_heads": 32,
19
+ "num_hidden_layers": 16,
20
+ "num_key_value_heads": 8,
21
+ "pretraining_tp": 1,
22
+ "rms_norm_eps": 1e-05,
23
+ "rope_scaling": {
24
+ "factor": 32.0,
25
+ "high_freq_factor": 4.0,
26
+ "low_freq_factor": 1.0,
27
+ "original_max_position_embeddings": 8192,
28
+ "rope_type": "llama3"
29
+ },
30
+ "rope_theta": 500000.0,
31
+ "tie_word_embeddings": true,
32
+ "transformers_version": "4.57.3",
33
+ "use_cache": true,
34
+ "vocab_size": 128256
35
+ }
2500_hf/generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 128000,
4
+ "eos_token_id": 128001,
5
+ "transformers_version": "4.57.3"
6
+ }
2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2025-12-29T05-34-34.570482.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "alias": "arc_challenge",
5
+ "acc,none": 0.32849829351535836,
6
+ "acc_stderr,none": 0.013724978465537373,
7
+ "acc_norm,none": 0.36945392491467577,
8
+ "acc_norm_stderr,none": 0.014104578366491894
9
+ },
10
+ "arc_easy": {
11
+ "alias": "arc_easy",
12
+ "acc,none": 0.6616161616161617,
13
+ "acc_stderr,none": 0.009709034670525096,
14
+ "acc_norm,none": 0.627104377104377,
15
+ "acc_norm_stderr,none": 0.00992274319712925
16
+ },
17
+ "hellaswag": {
18
+ "alias": "hellaswag",
19
+ "acc,none": 0.4314877514439355,
20
+ "acc_stderr,none": 0.004942716091996074,
21
+ "acc_norm,none": 0.5677155945030871,
22
+ "acc_norm_stderr,none": 0.004943809330692698
23
+ }
24
+ },
25
+ "group_subtasks": {
26
+ "arc_challenge": [],
27
+ "arc_easy": [],
28
+ "hellaswag": []
29
+ },
30
+ "configs": {
31
+ "arc_challenge": {
32
+ "task": "arc_challenge",
33
+ "tag": [
34
+ "ai2_arc"
35
+ ],
36
+ "dataset_path": "allenai/ai2_arc",
37
+ "dataset_name": "ARC-Challenge",
38
+ "training_split": "train",
39
+ "validation_split": "validation",
40
+ "test_split": "test",
41
+ "doc_to_text": "Question: {{question}}\nAnswer:",
42
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
43
+ "unsafe_code": false,
44
+ "doc_to_choice": "{{choices.text}}",
45
+ "description": "",
46
+ "target_delimiter": " ",
47
+ "fewshot_delimiter": "\n\n",
48
+ "num_fewshot": 0,
49
+ "metric_list": [
50
+ {
51
+ "metric": "acc",
52
+ "aggregation": "mean",
53
+ "higher_is_better": true
54
+ },
55
+ {
56
+ "metric": "acc_norm",
57
+ "aggregation": "mean",
58
+ "higher_is_better": true
59
+ }
60
+ ],
61
+ "output_type": "multiple_choice",
62
+ "repeats": 1,
63
+ "should_decontaminate": true,
64
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
65
+ "metadata": {
66
+ "version": 1.0,
67
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
68
+ }
69
+ },
70
+ "arc_easy": {
71
+ "task": "arc_easy",
72
+ "tag": [
73
+ "ai2_arc"
74
+ ],
75
+ "dataset_path": "allenai/ai2_arc",
76
+ "dataset_name": "ARC-Easy",
77
+ "training_split": "train",
78
+ "validation_split": "validation",
79
+ "test_split": "test",
80
+ "doc_to_text": "Question: {{question}}\nAnswer:",
81
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
82
+ "unsafe_code": false,
83
+ "doc_to_choice": "{{choices.text}}",
84
+ "description": "",
85
+ "target_delimiter": " ",
86
+ "fewshot_delimiter": "\n\n",
87
+ "num_fewshot": 0,
88
+ "metric_list": [
89
+ {
90
+ "metric": "acc",
91
+ "aggregation": "mean",
92
+ "higher_is_better": true
93
+ },
94
+ {
95
+ "metric": "acc_norm",
96
+ "aggregation": "mean",
97
+ "higher_is_better": true
98
+ }
99
+ ],
100
+ "output_type": "multiple_choice",
101
+ "repeats": 1,
102
+ "should_decontaminate": true,
103
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
104
+ "metadata": {
105
+ "version": 1.0,
106
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
107
+ }
108
+ },
109
+ "hellaswag": {
110
+ "task": "hellaswag",
111
+ "tag": [
112
+ "multiple_choice"
113
+ ],
114
+ "dataset_path": "Rowan/hellaswag",
115
+ "training_split": "train",
116
+ "validation_split": "validation",
117
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc):\n ctx = doc[\"ctx_a\"] + \" \" + doc[\"ctx_b\"].capitalize()\n out_doc = {\n \"query\": preprocess(doc[\"activity_label\"] + \": \" + ctx),\n \"choices\": [preprocess(ending) for ending in doc[\"endings\"]],\n \"gold\": int(doc[\"label\"]),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
118
+ "doc_to_text": "{{query}}",
119
+ "doc_to_target": "{{label}}",
120
+ "unsafe_code": false,
121
+ "doc_to_choice": "choices",
122
+ "description": "",
123
+ "target_delimiter": " ",
124
+ "fewshot_delimiter": "\n\n",
125
+ "num_fewshot": 0,
126
+ "metric_list": [
127
+ {
128
+ "metric": "acc",
129
+ "aggregation": "mean",
130
+ "higher_is_better": true
131
+ },
132
+ {
133
+ "metric": "acc_norm",
134
+ "aggregation": "mean",
135
+ "higher_is_better": true
136
+ }
137
+ ],
138
+ "output_type": "multiple_choice",
139
+ "repeats": 1,
140
+ "should_decontaminate": false,
141
+ "metadata": {
142
+ "version": 1.0,
143
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
144
+ }
145
+ }
146
+ },
147
+ "versions": {
148
+ "arc_challenge": 1.0,
149
+ "arc_easy": 1.0,
150
+ "hellaswag": 1.0
151
+ },
152
+ "n-shot": {
153
+ "arc_challenge": 0,
154
+ "arc_easy": 0,
155
+ "hellaswag": 0
156
+ },
157
+ "higher_is_better": {
158
+ "arc_challenge": {
159
+ "acc": true,
160
+ "acc_norm": true
161
+ },
162
+ "arc_easy": {
163
+ "acc": true,
164
+ "acc_norm": true
165
+ },
166
+ "hellaswag": {
167
+ "acc": true,
168
+ "acc_norm": true
169
+ }
170
+ },
171
+ "n-samples": {
172
+ "hellaswag": {
173
+ "original": 10042,
174
+ "effective": 10042
175
+ },
176
+ "arc_easy": {
177
+ "original": 2376,
178
+ "effective": 2376
179
+ },
180
+ "arc_challenge": {
181
+ "original": 1172,
182
+ "effective": 1172
183
+ }
184
+ },
185
+ "config": {
186
+ "model": "hf",
187
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/2500_hf",
188
+ "model_num_parameters": 1235814400,
189
+ "model_dtype": "torch.bfloat16",
190
+ "model_revision": "main",
191
+ "model_sha": "",
192
+ "batch_size": "32",
193
+ "batch_sizes": [],
194
+ "device": null,
195
+ "use_cache": null,
196
+ "limit": null,
197
+ "bootstrap_iters": 100000,
198
+ "gen_kwargs": null,
199
+ "random_seed": 0,
200
+ "numpy_seed": 1234,
201
+ "torch_seed": 1234,
202
+ "fewshot_seed": 1234
203
+ },
204
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
205
+ "date": 1766986308.044293,
206
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-77-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.129.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 48 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 96\nOn-line CPU(s) list: 0-95\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7352 24-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 24\nSocket(s): 2\nStepping: 0\nBogoMIPS: 4591.66\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca\nVirtualization: AMD-V\nL1d cache: 1.5 MiB (48 instances)\nL1i cache: 1.5 MiB (48 instances)\nL2 cache: 24 MiB (48 instances)\nL3 cache: 256 MiB (16 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-5,48-53\nNUMA node1 CPU(s): 6-11,54-59\nNUMA node2 CPU(s): 12-17,60-65\nNUMA node3 CPU(s): 18-23,66-71\nNUMA node4 CPU(s): 24-29,72-77\nNUMA node5 CPU(s): 30-35,78-83\nNUMA node6 CPU(s): 36-41,84-89\nNUMA node7 CPU(s): 42-47,90-95\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Full AMD retpoline, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
207
+ "transformers_version": "4.57.3",
208
+ "lm_eval_version": "0.4.9.1",
209
+ "upper_git_hash": null,
210
+ "tokenizer_pad_token": [
211
+ "<|end_of_text|>",
212
+ "128001"
213
+ ],
214
+ "tokenizer_eos_token": [
215
+ "<|end_of_text|>",
216
+ "128001"
217
+ ],
218
+ "tokenizer_bos_token": [
219
+ "<|begin_of_text|>",
220
+ "128000"
221
+ ],
222
+ "eot_token_id": 128001,
223
+ "max_length": 131072,
224
+ "task_hashes": {
225
+ "hellaswag": "edcc7edd27a555d3f7cbca0641152b2c5e4eb6eb79c5e62d7fe5887f47814323",
226
+ "arc_easy": "dce0d9b0f0cecd55bf2ac264042c5e45487df708d13123af3ae9e67bbbefdeb1",
227
+ "arc_challenge": "09f9ae87a0905d63512cffc4aa91a55e44258fc35160e40fa1eb66fb75473e34"
228
+ },
229
+ "model_source": "hf",
230
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf",
231
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__2500_hf",
232
+ "system_instruction": null,
233
+ "system_instruction_sha": null,
234
+ "fewshot_as_multiturn": false,
235
+ "chat_template": null,
236
+ "chat_template_sha": null,
237
+ "start_time": 18578978.788538285,
238
+ "end_time": 18579202.067502547,
239
+ "total_evaluation_time_seconds": "223.2789642624557"
240
+ }
2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2026-01-07T03-51-34.430609.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "arc_challenge": {
4
+ "alias": "arc_challenge",
5
+ "acc,none": 0.32849829351535836,
6
+ "acc_stderr,none": 0.013724978465537373,
7
+ "acc_norm,none": 0.36945392491467577,
8
+ "acc_norm_stderr,none": 0.014104578366491894
9
+ },
10
+ "arc_easy": {
11
+ "alias": "arc_easy",
12
+ "acc,none": 0.6616161616161617,
13
+ "acc_stderr,none": 0.009709034670525096,
14
+ "acc_norm,none": 0.627104377104377,
15
+ "acc_norm_stderr,none": 0.00992274319712925
16
+ },
17
+ "hellaswag": {
18
+ "alias": "hellaswag",
19
+ "acc,none": 0.4314877514439355,
20
+ "acc_stderr,none": 0.004942716091996074,
21
+ "acc_norm,none": 0.5677155945030871,
22
+ "acc_norm_stderr,none": 0.004943809330692698
23
+ }
24
+ },
25
+ "group_subtasks": {
26
+ "arc_challenge": [],
27
+ "arc_easy": [],
28
+ "hellaswag": []
29
+ },
30
+ "configs": {
31
+ "arc_challenge": {
32
+ "task": "arc_challenge",
33
+ "tag": [
34
+ "ai2_arc"
35
+ ],
36
+ "dataset_path": "allenai/ai2_arc",
37
+ "dataset_name": "ARC-Challenge",
38
+ "training_split": "train",
39
+ "validation_split": "validation",
40
+ "test_split": "test",
41
+ "doc_to_text": "Question: {{question}}\nAnswer:",
42
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
43
+ "unsafe_code": false,
44
+ "doc_to_choice": "{{choices.text}}",
45
+ "description": "",
46
+ "target_delimiter": " ",
47
+ "fewshot_delimiter": "\n\n",
48
+ "num_fewshot": 0,
49
+ "metric_list": [
50
+ {
51
+ "metric": "acc",
52
+ "aggregation": "mean",
53
+ "higher_is_better": true
54
+ },
55
+ {
56
+ "metric": "acc_norm",
57
+ "aggregation": "mean",
58
+ "higher_is_better": true
59
+ }
60
+ ],
61
+ "output_type": "multiple_choice",
62
+ "repeats": 1,
63
+ "should_decontaminate": true,
64
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
65
+ "metadata": {
66
+ "version": 1.0,
67
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
68
+ }
69
+ },
70
+ "arc_easy": {
71
+ "task": "arc_easy",
72
+ "tag": [
73
+ "ai2_arc"
74
+ ],
75
+ "dataset_path": "allenai/ai2_arc",
76
+ "dataset_name": "ARC-Easy",
77
+ "training_split": "train",
78
+ "validation_split": "validation",
79
+ "test_split": "test",
80
+ "doc_to_text": "Question: {{question}}\nAnswer:",
81
+ "doc_to_target": "{{choices.label.index(answerKey)}}",
82
+ "unsafe_code": false,
83
+ "doc_to_choice": "{{choices.text}}",
84
+ "description": "",
85
+ "target_delimiter": " ",
86
+ "fewshot_delimiter": "\n\n",
87
+ "num_fewshot": 0,
88
+ "metric_list": [
89
+ {
90
+ "metric": "acc",
91
+ "aggregation": "mean",
92
+ "higher_is_better": true
93
+ },
94
+ {
95
+ "metric": "acc_norm",
96
+ "aggregation": "mean",
97
+ "higher_is_better": true
98
+ }
99
+ ],
100
+ "output_type": "multiple_choice",
101
+ "repeats": 1,
102
+ "should_decontaminate": true,
103
+ "doc_to_decontamination_query": "Question: {{question}}\nAnswer:",
104
+ "metadata": {
105
+ "version": 1.0,
106
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
107
+ }
108
+ },
109
+ "hellaswag": {
110
+ "task": "hellaswag",
111
+ "tag": [
112
+ "multiple_choice"
113
+ ],
114
+ "dataset_path": "Rowan/hellaswag",
115
+ "training_split": "train",
116
+ "validation_split": "validation",
117
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc):\n ctx = doc[\"ctx_a\"] + \" \" + doc[\"ctx_b\"].capitalize()\n out_doc = {\n \"query\": preprocess(doc[\"activity_label\"] + \": \" + ctx),\n \"choices\": [preprocess(ending) for ending in doc[\"endings\"]],\n \"gold\": int(doc[\"label\"]),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
118
+ "doc_to_text": "{{query}}",
119
+ "doc_to_target": "{{label}}",
120
+ "unsafe_code": false,
121
+ "doc_to_choice": "choices",
122
+ "description": "",
123
+ "target_delimiter": " ",
124
+ "fewshot_delimiter": "\n\n",
125
+ "num_fewshot": 0,
126
+ "metric_list": [
127
+ {
128
+ "metric": "acc",
129
+ "aggregation": "mean",
130
+ "higher_is_better": true
131
+ },
132
+ {
133
+ "metric": "acc_norm",
134
+ "aggregation": "mean",
135
+ "higher_is_better": true
136
+ }
137
+ ],
138
+ "output_type": "multiple_choice",
139
+ "repeats": 1,
140
+ "should_decontaminate": false,
141
+ "metadata": {
142
+ "version": 1.0,
143
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
144
+ }
145
+ }
146
+ },
147
+ "versions": {
148
+ "arc_challenge": 1.0,
149
+ "arc_easy": 1.0,
150
+ "hellaswag": 1.0
151
+ },
152
+ "n-shot": {
153
+ "arc_challenge": 0,
154
+ "arc_easy": 0,
155
+ "hellaswag": 0
156
+ },
157
+ "higher_is_better": {
158
+ "arc_challenge": {
159
+ "acc": true,
160
+ "acc_norm": true
161
+ },
162
+ "arc_easy": {
163
+ "acc": true,
164
+ "acc_norm": true
165
+ },
166
+ "hellaswag": {
167
+ "acc": true,
168
+ "acc_norm": true
169
+ }
170
+ },
171
+ "n-samples": {
172
+ "hellaswag": {
173
+ "original": 10042,
174
+ "effective": 10042
175
+ },
176
+ "arc_easy": {
177
+ "original": 2376,
178
+ "effective": 2376
179
+ },
180
+ "arc_challenge": {
181
+ "original": 1172,
182
+ "effective": 1172
183
+ }
184
+ },
185
+ "config": {
186
+ "model": "hf",
187
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/2500_hf",
188
+ "model_num_parameters": 1235814400,
189
+ "model_dtype": "torch.bfloat16",
190
+ "model_revision": "main",
191
+ "model_sha": "",
192
+ "batch_size": "32",
193
+ "batch_sizes": [],
194
+ "device": null,
195
+ "use_cache": null,
196
+ "limit": null,
197
+ "bootstrap_iters": 100000,
198
+ "gen_kwargs": null,
199
+ "random_seed": 0,
200
+ "numpy_seed": 1234,
201
+ "torch_seed": 1234,
202
+ "fewshot_seed": 1234
203
+ },
204
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
205
+ "date": 1767757750.437144,
206
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-193-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.183.06\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 256\nOn-line CPU(s) list: 0-255\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7742 64-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 64\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2250.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4491.81\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca sme sev sev_es\nVirtualization: AMD-V\nL1d cache: 4 MiB (128 instances)\nL1i cache: 4 MiB (128 instances)\nL2 cache: 64 MiB (128 instances)\nL3 cache: 512 MiB (32 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-15,128-143\nNUMA node1 CPU(s): 16-31,144-159\nNUMA node2 CPU(s): 32-47,160-175\nNUMA node3 CPU(s): 48-63,176-191\nNUMA node4 CPU(s): 64-79,192-207\nNUMA node5 CPU(s): 80-95,208-223\nNUMA node6 CPU(s): 96-111,224-239\nNUMA node7 CPU(s): 112-127,240-255\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Retbleed: Vulnerable\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Retpolines; IBPB conditional; IBRS_FW; STIBP conditional; RSB filling; PBRSB-eIBRS Not affected; BHI Not affected\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
207
+ "transformers_version": "4.57.3",
208
+ "lm_eval_version": "0.4.9.1",
209
+ "upper_git_hash": null,
210
+ "tokenizer_pad_token": [
211
+ "<|end_of_text|>",
212
+ "128001"
213
+ ],
214
+ "tokenizer_eos_token": [
215
+ "<|end_of_text|>",
216
+ "128001"
217
+ ],
218
+ "tokenizer_bos_token": [
219
+ "<|begin_of_text|>",
220
+ "128000"
221
+ ],
222
+ "eot_token_id": 128001,
223
+ "max_length": 131072,
224
+ "task_hashes": {
225
+ "hellaswag": "edcc7edd27a555d3f7cbca0641152b2c5e4eb6eb79c5e62d7fe5887f47814323",
226
+ "arc_easy": "dce0d9b0f0cecd55bf2ac264042c5e45487df708d13123af3ae9e67bbbefdeb1",
227
+ "arc_challenge": "09f9ae87a0905d63512cffc4aa91a55e44258fc35160e40fa1eb66fb75473e34"
228
+ },
229
+ "model_source": "hf",
230
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf",
231
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__2500_hf",
232
+ "system_instruction": null,
233
+ "system_instruction_sha": null,
234
+ "fewshot_as_multiturn": false,
235
+ "chat_template": null,
236
+ "chat_template_sha": null,
237
+ "start_time": 13820945.364448603,
238
+ "end_time": 13821148.638297528,
239
+ "total_evaluation_time_seconds": "203.27384892478585"
240
+ }
2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_challenge_2025-12-29T05-34-34.570482.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_challenge_2026-01-07T03-51-34.430609.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_easy_2025-12-29T05-34-34.570482.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
2500_hf/harness_eval_0shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_arc_easy_2026-01-07T03-51-34.430609.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2025-12-29T06-24-24.342056.json ADDED
@@ -0,0 +1,478 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hendrycks_math": {
4
+ "exact_match,none": 0.0922,
5
+ "exact_match_stderr,none": 0.004063954564129501,
6
+ "alias": "hendrycks_math"
7
+ },
8
+ "hendrycks_math_algebra": {
9
+ "alias": " - hendrycks_math_algebra",
10
+ "exact_match,none": 0.09604043807919124,
11
+ "exact_match_stderr,none": 0.008555772403896333
12
+ },
13
+ "hendrycks_math_counting_and_prob": {
14
+ "alias": " - hendrycks_math_counting_and_prob",
15
+ "exact_match,none": 0.11392405063291139,
16
+ "exact_match_stderr,none": 0.014608735065434468
17
+ },
18
+ "hendrycks_math_geometry": {
19
+ "alias": " - hendrycks_math_geometry",
20
+ "exact_match,none": 0.07724425887265135,
21
+ "exact_match_stderr,none": 0.01221131783003633
22
+ },
23
+ "hendrycks_math_intermediate_algebra": {
24
+ "alias": " - hendrycks_math_intermediate_algebra",
25
+ "exact_match,none": 0.05094130675526024,
26
+ "exact_match_stderr,none": 0.007321133187865446
27
+ },
28
+ "hendrycks_math_num_theory": {
29
+ "alias": " - hendrycks_math_num_theory",
30
+ "exact_match,none": 0.08518518518518518,
31
+ "exact_match_stderr,none": 0.012024146192975114
32
+ },
33
+ "hendrycks_math_prealgebra": {
34
+ "alias": " - hendrycks_math_prealgebra",
35
+ "exact_match,none": 0.1549942594718714,
36
+ "exact_match_stderr,none": 0.012269532352170633
37
+ },
38
+ "hendrycks_math_precalc": {
39
+ "alias": " - hendrycks_math_precalc",
40
+ "exact_match,none": 0.05311355311355311,
41
+ "exact_match_stderr,none": 0.009606238759236748
42
+ }
43
+ },
44
+ "groups": {
45
+ "hendrycks_math": {
46
+ "exact_match,none": 0.0922,
47
+ "exact_match_stderr,none": 0.004063954564129501,
48
+ "alias": "hendrycks_math"
49
+ }
50
+ },
51
+ "group_subtasks": {
52
+ "hendrycks_math": [
53
+ "hendrycks_math_algebra",
54
+ "hendrycks_math_counting_and_prob",
55
+ "hendrycks_math_geometry",
56
+ "hendrycks_math_intermediate_algebra",
57
+ "hendrycks_math_num_theory",
58
+ "hendrycks_math_prealgebra",
59
+ "hendrycks_math_precalc"
60
+ ]
61
+ },
62
+ "configs": {
63
+ "hendrycks_math_algebra": {
64
+ "task": "hendrycks_math_algebra",
65
+ "tag": [
66
+ "math_word_problems"
67
+ ],
68
+ "dataset_path": "EleutherAI/hendrycks_math",
69
+ "dataset_name": "algebra",
70
+ "training_split": "train",
71
+ "test_split": "test",
72
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
73
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
74
+ "doc_to_target": "{{answer}}",
75
+ "unsafe_code": false,
76
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
77
+ "description": "",
78
+ "target_delimiter": " ",
79
+ "fewshot_delimiter": "\n\n",
80
+ "num_fewshot": 4,
81
+ "metric_list": [
82
+ {
83
+ "metric": "exact_match",
84
+ "aggregation": "mean",
85
+ "higher_is_better": true
86
+ }
87
+ ],
88
+ "output_type": "generate_until",
89
+ "generation_kwargs": {
90
+ "until": [
91
+ "Problem:"
92
+ ],
93
+ "do_sample": false,
94
+ "temperature": 0.0
95
+ },
96
+ "repeats": 1,
97
+ "should_decontaminate": false,
98
+ "metadata": {
99
+ "version": 1.0,
100
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
101
+ }
102
+ },
103
+ "hendrycks_math_counting_and_prob": {
104
+ "task": "hendrycks_math_counting_and_prob",
105
+ "tag": [
106
+ "math_word_problems"
107
+ ],
108
+ "dataset_path": "EleutherAI/hendrycks_math",
109
+ "dataset_name": "counting_and_probability",
110
+ "training_split": "train",
111
+ "test_split": "test",
112
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
113
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
114
+ "doc_to_target": "{{answer}}",
115
+ "unsafe_code": false,
116
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
117
+ "description": "",
118
+ "target_delimiter": " ",
119
+ "fewshot_delimiter": "\n\n",
120
+ "num_fewshot": 4,
121
+ "metric_list": [
122
+ {
123
+ "metric": "exact_match",
124
+ "aggregation": "mean",
125
+ "higher_is_better": true
126
+ }
127
+ ],
128
+ "output_type": "generate_until",
129
+ "generation_kwargs": {
130
+ "until": [
131
+ "Problem:"
132
+ ],
133
+ "do_sample": false,
134
+ "temperature": 0.0
135
+ },
136
+ "repeats": 1,
137
+ "should_decontaminate": false,
138
+ "metadata": {
139
+ "version": 1.0,
140
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
141
+ }
142
+ },
143
+ "hendrycks_math_geometry": {
144
+ "task": "hendrycks_math_geometry",
145
+ "tag": [
146
+ "math_word_problems"
147
+ ],
148
+ "dataset_path": "EleutherAI/hendrycks_math",
149
+ "dataset_name": "geometry",
150
+ "training_split": "train",
151
+ "test_split": "test",
152
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
153
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
154
+ "doc_to_target": "{{answer}}",
155
+ "unsafe_code": false,
156
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
157
+ "description": "",
158
+ "target_delimiter": " ",
159
+ "fewshot_delimiter": "\n\n",
160
+ "num_fewshot": 4,
161
+ "metric_list": [
162
+ {
163
+ "metric": "exact_match",
164
+ "aggregation": "mean",
165
+ "higher_is_better": true
166
+ }
167
+ ],
168
+ "output_type": "generate_until",
169
+ "generation_kwargs": {
170
+ "until": [
171
+ "Problem:"
172
+ ],
173
+ "do_sample": false,
174
+ "temperature": 0.0
175
+ },
176
+ "repeats": 1,
177
+ "should_decontaminate": false,
178
+ "metadata": {
179
+ "version": 1.0,
180
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
181
+ }
182
+ },
183
+ "hendrycks_math_intermediate_algebra": {
184
+ "task": "hendrycks_math_intermediate_algebra",
185
+ "tag": [
186
+ "math_word_problems"
187
+ ],
188
+ "dataset_path": "EleutherAI/hendrycks_math",
189
+ "dataset_name": "intermediate_algebra",
190
+ "training_split": "train",
191
+ "test_split": "test",
192
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
193
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
194
+ "doc_to_target": "{{answer}}",
195
+ "unsafe_code": false,
196
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
197
+ "description": "",
198
+ "target_delimiter": " ",
199
+ "fewshot_delimiter": "\n\n",
200
+ "num_fewshot": 4,
201
+ "metric_list": [
202
+ {
203
+ "metric": "exact_match",
204
+ "aggregation": "mean",
205
+ "higher_is_better": true
206
+ }
207
+ ],
208
+ "output_type": "generate_until",
209
+ "generation_kwargs": {
210
+ "until": [
211
+ "Problem:"
212
+ ],
213
+ "do_sample": false,
214
+ "temperature": 0.0
215
+ },
216
+ "repeats": 1,
217
+ "should_decontaminate": false,
218
+ "metadata": {
219
+ "version": 1.0,
220
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
221
+ }
222
+ },
223
+ "hendrycks_math_num_theory": {
224
+ "task": "hendrycks_math_num_theory",
225
+ "tag": [
226
+ "math_word_problems"
227
+ ],
228
+ "dataset_path": "EleutherAI/hendrycks_math",
229
+ "dataset_name": "number_theory",
230
+ "training_split": "train",
231
+ "test_split": "test",
232
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
233
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
234
+ "doc_to_target": "{{answer}}",
235
+ "unsafe_code": false,
236
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
237
+ "description": "",
238
+ "target_delimiter": " ",
239
+ "fewshot_delimiter": "\n\n",
240
+ "num_fewshot": 4,
241
+ "metric_list": [
242
+ {
243
+ "metric": "exact_match",
244
+ "aggregation": "mean",
245
+ "higher_is_better": true
246
+ }
247
+ ],
248
+ "output_type": "generate_until",
249
+ "generation_kwargs": {
250
+ "until": [
251
+ "Problem:"
252
+ ],
253
+ "do_sample": false,
254
+ "temperature": 0.0
255
+ },
256
+ "repeats": 1,
257
+ "should_decontaminate": false,
258
+ "metadata": {
259
+ "version": 1.0,
260
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
261
+ }
262
+ },
263
+ "hendrycks_math_prealgebra": {
264
+ "task": "hendrycks_math_prealgebra",
265
+ "tag": [
266
+ "math_word_problems"
267
+ ],
268
+ "dataset_path": "EleutherAI/hendrycks_math",
269
+ "dataset_name": "prealgebra",
270
+ "training_split": "train",
271
+ "test_split": "test",
272
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
273
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
274
+ "doc_to_target": "{{answer}}",
275
+ "unsafe_code": false,
276
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
277
+ "description": "",
278
+ "target_delimiter": " ",
279
+ "fewshot_delimiter": "\n\n",
280
+ "num_fewshot": 4,
281
+ "metric_list": [
282
+ {
283
+ "metric": "exact_match",
284
+ "aggregation": "mean",
285
+ "higher_is_better": true
286
+ }
287
+ ],
288
+ "output_type": "generate_until",
289
+ "generation_kwargs": {
290
+ "until": [
291
+ "Problem:"
292
+ ],
293
+ "do_sample": false,
294
+ "temperature": 0.0
295
+ },
296
+ "repeats": 1,
297
+ "should_decontaminate": false,
298
+ "metadata": {
299
+ "version": 1.0,
300
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
301
+ }
302
+ },
303
+ "hendrycks_math_precalc": {
304
+ "task": "hendrycks_math_precalc",
305
+ "tag": [
306
+ "math_word_problems"
307
+ ],
308
+ "dataset_path": "EleutherAI/hendrycks_math",
309
+ "dataset_name": "precalculus",
310
+ "training_split": "train",
311
+ "test_split": "test",
312
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
313
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
314
+ "doc_to_target": "{{answer}}",
315
+ "unsafe_code": false,
316
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
317
+ "description": "",
318
+ "target_delimiter": " ",
319
+ "fewshot_delimiter": "\n\n",
320
+ "num_fewshot": 4,
321
+ "metric_list": [
322
+ {
323
+ "metric": "exact_match",
324
+ "aggregation": "mean",
325
+ "higher_is_better": true
326
+ }
327
+ ],
328
+ "output_type": "generate_until",
329
+ "generation_kwargs": {
330
+ "until": [
331
+ "Problem:"
332
+ ],
333
+ "do_sample": false,
334
+ "temperature": 0.0
335
+ },
336
+ "repeats": 1,
337
+ "should_decontaminate": false,
338
+ "metadata": {
339
+ "version": 1.0,
340
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
341
+ }
342
+ }
343
+ },
344
+ "versions": {
345
+ "hendrycks_math": 1.0,
346
+ "hendrycks_math_algebra": 1.0,
347
+ "hendrycks_math_counting_and_prob": 1.0,
348
+ "hendrycks_math_geometry": 1.0,
349
+ "hendrycks_math_intermediate_algebra": 1.0,
350
+ "hendrycks_math_num_theory": 1.0,
351
+ "hendrycks_math_prealgebra": 1.0,
352
+ "hendrycks_math_precalc": 1.0
353
+ },
354
+ "n-shot": {
355
+ "hendrycks_math_algebra": 4,
356
+ "hendrycks_math_counting_and_prob": 4,
357
+ "hendrycks_math_geometry": 4,
358
+ "hendrycks_math_intermediate_algebra": 4,
359
+ "hendrycks_math_num_theory": 4,
360
+ "hendrycks_math_prealgebra": 4,
361
+ "hendrycks_math_precalc": 4
362
+ },
363
+ "higher_is_better": {
364
+ "hendrycks_math": {
365
+ "exact_match": true
366
+ },
367
+ "hendrycks_math_algebra": {
368
+ "exact_match": true
369
+ },
370
+ "hendrycks_math_counting_and_prob": {
371
+ "exact_match": true
372
+ },
373
+ "hendrycks_math_geometry": {
374
+ "exact_match": true
375
+ },
376
+ "hendrycks_math_intermediate_algebra": {
377
+ "exact_match": true
378
+ },
379
+ "hendrycks_math_num_theory": {
380
+ "exact_match": true
381
+ },
382
+ "hendrycks_math_prealgebra": {
383
+ "exact_match": true
384
+ },
385
+ "hendrycks_math_precalc": {
386
+ "exact_match": true
387
+ }
388
+ },
389
+ "n-samples": {
390
+ "hendrycks_math_algebra": {
391
+ "original": 1187,
392
+ "effective": 1187
393
+ },
394
+ "hendrycks_math_counting_and_prob": {
395
+ "original": 474,
396
+ "effective": 474
397
+ },
398
+ "hendrycks_math_geometry": {
399
+ "original": 479,
400
+ "effective": 479
401
+ },
402
+ "hendrycks_math_intermediate_algebra": {
403
+ "original": 903,
404
+ "effective": 903
405
+ },
406
+ "hendrycks_math_num_theory": {
407
+ "original": 540,
408
+ "effective": 540
409
+ },
410
+ "hendrycks_math_prealgebra": {
411
+ "original": 871,
412
+ "effective": 871
413
+ },
414
+ "hendrycks_math_precalc": {
415
+ "original": 546,
416
+ "effective": 546
417
+ }
418
+ },
419
+ "config": {
420
+ "model": "hf",
421
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/2500_hf,trust_remote_code=True",
422
+ "model_num_parameters": 1235814400,
423
+ "model_dtype": "torch.bfloat16",
424
+ "model_revision": "main",
425
+ "model_sha": "",
426
+ "batch_size": "64",
427
+ "batch_sizes": [],
428
+ "device": null,
429
+ "use_cache": null,
430
+ "limit": null,
431
+ "bootstrap_iters": 100000,
432
+ "gen_kwargs": null,
433
+ "random_seed": 0,
434
+ "numpy_seed": 1234,
435
+ "torch_seed": 1234,
436
+ "fewshot_seed": 1234
437
+ },
438
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
439
+ "date": 1766989037.6777337,
440
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-77-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.129.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 48 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 96\nOn-line CPU(s) list: 0-95\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7352 24-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 24\nSocket(s): 2\nStepping: 0\nBogoMIPS: 4591.66\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca\nVirtualization: AMD-V\nL1d cache: 1.5 MiB (48 instances)\nL1i cache: 1.5 MiB (48 instances)\nL2 cache: 24 MiB (48 instances)\nL3 cache: 256 MiB (16 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-5,48-53\nNUMA node1 CPU(s): 6-11,54-59\nNUMA node2 CPU(s): 12-17,60-65\nNUMA node3 CPU(s): 18-23,66-71\nNUMA node4 CPU(s): 24-29,72-77\nNUMA node5 CPU(s): 30-35,78-83\nNUMA node6 CPU(s): 36-41,84-89\nNUMA node7 CPU(s): 42-47,90-95\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Full AMD retpoline, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
441
+ "transformers_version": "4.57.3",
442
+ "lm_eval_version": "0.4.9.1",
443
+ "upper_git_hash": null,
444
+ "tokenizer_pad_token": [
445
+ "<|end_of_text|>",
446
+ "128001"
447
+ ],
448
+ "tokenizer_eos_token": [
449
+ "<|end_of_text|>",
450
+ "128001"
451
+ ],
452
+ "tokenizer_bos_token": [
453
+ "<|begin_of_text|>",
454
+ "128000"
455
+ ],
456
+ "eot_token_id": 128001,
457
+ "max_length": 131072,
458
+ "task_hashes": {
459
+ "hendrycks_math_algebra": "7b50860e4db85f9d690d9679b543412837815a3b9854f0d0d166cfefc428cc23",
460
+ "hendrycks_math_counting_and_prob": "826f9744663d8e2e04e0050aa87b6edc5640c41ca1aa53fee8ef747f7c93b100",
461
+ "hendrycks_math_geometry": "4b67d6f45ba7f89a0c8a0a01fc9b1885bf942346a45b6a4fd03afd3d0e23ab6e",
462
+ "hendrycks_math_intermediate_algebra": "d7bc95882c4f58ef0b7252e6026e98b8bc77bbfbe36de6da6d97127e9744141e",
463
+ "hendrycks_math_num_theory": "ff82fd9f09e0a7c370270b3e7778773f7f1dec4aa40564e1a1318417af819030",
464
+ "hendrycks_math_prealgebra": "e4296905fe5f21ca452a0596371350ffedec630103ac7cce4b76cb806c55be05",
465
+ "hendrycks_math_precalc": "065216e339084dad7a3260b23a5151c8bf63771c7f0d23e88b6f3df4a9f0175a"
466
+ },
467
+ "model_source": "hf",
468
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf",
469
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__2500_hf",
470
+ "system_instruction": null,
471
+ "system_instruction_sha": null,
472
+ "fewshot_as_multiturn": false,
473
+ "chat_template": null,
474
+ "chat_template_sha": null,
475
+ "start_time": 18581707.45205234,
476
+ "end_time": 18582191.849963754,
477
+ "total_evaluation_time_seconds": "484.39791141450405"
478
+ }
2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/results_2026-01-07T03-55-15.968865.json ADDED
@@ -0,0 +1,478 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "results": {
3
+ "hendrycks_math": {
4
+ "exact_match,none": 0.0922,
5
+ "exact_match_stderr,none": 0.004063954564129501,
6
+ "alias": "hendrycks_math"
7
+ },
8
+ "hendrycks_math_algebra": {
9
+ "alias": " - hendrycks_math_algebra",
10
+ "exact_match,none": 0.09604043807919124,
11
+ "exact_match_stderr,none": 0.008555772403896333
12
+ },
13
+ "hendrycks_math_counting_and_prob": {
14
+ "alias": " - hendrycks_math_counting_and_prob",
15
+ "exact_match,none": 0.11392405063291139,
16
+ "exact_match_stderr,none": 0.014608735065434468
17
+ },
18
+ "hendrycks_math_geometry": {
19
+ "alias": " - hendrycks_math_geometry",
20
+ "exact_match,none": 0.07724425887265135,
21
+ "exact_match_stderr,none": 0.01221131783003633
22
+ },
23
+ "hendrycks_math_intermediate_algebra": {
24
+ "alias": " - hendrycks_math_intermediate_algebra",
25
+ "exact_match,none": 0.05094130675526024,
26
+ "exact_match_stderr,none": 0.007321133187865446
27
+ },
28
+ "hendrycks_math_num_theory": {
29
+ "alias": " - hendrycks_math_num_theory",
30
+ "exact_match,none": 0.08518518518518518,
31
+ "exact_match_stderr,none": 0.012024146192975114
32
+ },
33
+ "hendrycks_math_prealgebra": {
34
+ "alias": " - hendrycks_math_prealgebra",
35
+ "exact_match,none": 0.1549942594718714,
36
+ "exact_match_stderr,none": 0.012269532352170633
37
+ },
38
+ "hendrycks_math_precalc": {
39
+ "alias": " - hendrycks_math_precalc",
40
+ "exact_match,none": 0.05311355311355311,
41
+ "exact_match_stderr,none": 0.009606238759236748
42
+ }
43
+ },
44
+ "groups": {
45
+ "hendrycks_math": {
46
+ "exact_match,none": 0.0922,
47
+ "exact_match_stderr,none": 0.004063954564129501,
48
+ "alias": "hendrycks_math"
49
+ }
50
+ },
51
+ "group_subtasks": {
52
+ "hendrycks_math": [
53
+ "hendrycks_math_algebra",
54
+ "hendrycks_math_counting_and_prob",
55
+ "hendrycks_math_geometry",
56
+ "hendrycks_math_intermediate_algebra",
57
+ "hendrycks_math_num_theory",
58
+ "hendrycks_math_prealgebra",
59
+ "hendrycks_math_precalc"
60
+ ]
61
+ },
62
+ "configs": {
63
+ "hendrycks_math_algebra": {
64
+ "task": "hendrycks_math_algebra",
65
+ "tag": [
66
+ "math_word_problems"
67
+ ],
68
+ "dataset_path": "EleutherAI/hendrycks_math",
69
+ "dataset_name": "algebra",
70
+ "training_split": "train",
71
+ "test_split": "test",
72
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
73
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
74
+ "doc_to_target": "{{answer}}",
75
+ "unsafe_code": false,
76
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
77
+ "description": "",
78
+ "target_delimiter": " ",
79
+ "fewshot_delimiter": "\n\n",
80
+ "num_fewshot": 4,
81
+ "metric_list": [
82
+ {
83
+ "metric": "exact_match",
84
+ "aggregation": "mean",
85
+ "higher_is_better": true
86
+ }
87
+ ],
88
+ "output_type": "generate_until",
89
+ "generation_kwargs": {
90
+ "until": [
91
+ "Problem:"
92
+ ],
93
+ "do_sample": false,
94
+ "temperature": 0.0
95
+ },
96
+ "repeats": 1,
97
+ "should_decontaminate": false,
98
+ "metadata": {
99
+ "version": 1.0,
100
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
101
+ }
102
+ },
103
+ "hendrycks_math_counting_and_prob": {
104
+ "task": "hendrycks_math_counting_and_prob",
105
+ "tag": [
106
+ "math_word_problems"
107
+ ],
108
+ "dataset_path": "EleutherAI/hendrycks_math",
109
+ "dataset_name": "counting_and_probability",
110
+ "training_split": "train",
111
+ "test_split": "test",
112
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
113
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
114
+ "doc_to_target": "{{answer}}",
115
+ "unsafe_code": false,
116
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
117
+ "description": "",
118
+ "target_delimiter": " ",
119
+ "fewshot_delimiter": "\n\n",
120
+ "num_fewshot": 4,
121
+ "metric_list": [
122
+ {
123
+ "metric": "exact_match",
124
+ "aggregation": "mean",
125
+ "higher_is_better": true
126
+ }
127
+ ],
128
+ "output_type": "generate_until",
129
+ "generation_kwargs": {
130
+ "until": [
131
+ "Problem:"
132
+ ],
133
+ "do_sample": false,
134
+ "temperature": 0.0
135
+ },
136
+ "repeats": 1,
137
+ "should_decontaminate": false,
138
+ "metadata": {
139
+ "version": 1.0,
140
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
141
+ }
142
+ },
143
+ "hendrycks_math_geometry": {
144
+ "task": "hendrycks_math_geometry",
145
+ "tag": [
146
+ "math_word_problems"
147
+ ],
148
+ "dataset_path": "EleutherAI/hendrycks_math",
149
+ "dataset_name": "geometry",
150
+ "training_split": "train",
151
+ "test_split": "test",
152
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
153
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
154
+ "doc_to_target": "{{answer}}",
155
+ "unsafe_code": false,
156
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
157
+ "description": "",
158
+ "target_delimiter": " ",
159
+ "fewshot_delimiter": "\n\n",
160
+ "num_fewshot": 4,
161
+ "metric_list": [
162
+ {
163
+ "metric": "exact_match",
164
+ "aggregation": "mean",
165
+ "higher_is_better": true
166
+ }
167
+ ],
168
+ "output_type": "generate_until",
169
+ "generation_kwargs": {
170
+ "until": [
171
+ "Problem:"
172
+ ],
173
+ "do_sample": false,
174
+ "temperature": 0.0
175
+ },
176
+ "repeats": 1,
177
+ "should_decontaminate": false,
178
+ "metadata": {
179
+ "version": 1.0,
180
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
181
+ }
182
+ },
183
+ "hendrycks_math_intermediate_algebra": {
184
+ "task": "hendrycks_math_intermediate_algebra",
185
+ "tag": [
186
+ "math_word_problems"
187
+ ],
188
+ "dataset_path": "EleutherAI/hendrycks_math",
189
+ "dataset_name": "intermediate_algebra",
190
+ "training_split": "train",
191
+ "test_split": "test",
192
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
193
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
194
+ "doc_to_target": "{{answer}}",
195
+ "unsafe_code": false,
196
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
197
+ "description": "",
198
+ "target_delimiter": " ",
199
+ "fewshot_delimiter": "\n\n",
200
+ "num_fewshot": 4,
201
+ "metric_list": [
202
+ {
203
+ "metric": "exact_match",
204
+ "aggregation": "mean",
205
+ "higher_is_better": true
206
+ }
207
+ ],
208
+ "output_type": "generate_until",
209
+ "generation_kwargs": {
210
+ "until": [
211
+ "Problem:"
212
+ ],
213
+ "do_sample": false,
214
+ "temperature": 0.0
215
+ },
216
+ "repeats": 1,
217
+ "should_decontaminate": false,
218
+ "metadata": {
219
+ "version": 1.0,
220
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
221
+ }
222
+ },
223
+ "hendrycks_math_num_theory": {
224
+ "task": "hendrycks_math_num_theory",
225
+ "tag": [
226
+ "math_word_problems"
227
+ ],
228
+ "dataset_path": "EleutherAI/hendrycks_math",
229
+ "dataset_name": "number_theory",
230
+ "training_split": "train",
231
+ "test_split": "test",
232
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
233
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
234
+ "doc_to_target": "{{answer}}",
235
+ "unsafe_code": false,
236
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
237
+ "description": "",
238
+ "target_delimiter": " ",
239
+ "fewshot_delimiter": "\n\n",
240
+ "num_fewshot": 4,
241
+ "metric_list": [
242
+ {
243
+ "metric": "exact_match",
244
+ "aggregation": "mean",
245
+ "higher_is_better": true
246
+ }
247
+ ],
248
+ "output_type": "generate_until",
249
+ "generation_kwargs": {
250
+ "until": [
251
+ "Problem:"
252
+ ],
253
+ "do_sample": false,
254
+ "temperature": 0.0
255
+ },
256
+ "repeats": 1,
257
+ "should_decontaminate": false,
258
+ "metadata": {
259
+ "version": 1.0,
260
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
261
+ }
262
+ },
263
+ "hendrycks_math_prealgebra": {
264
+ "task": "hendrycks_math_prealgebra",
265
+ "tag": [
266
+ "math_word_problems"
267
+ ],
268
+ "dataset_path": "EleutherAI/hendrycks_math",
269
+ "dataset_name": "prealgebra",
270
+ "training_split": "train",
271
+ "test_split": "test",
272
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
273
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
274
+ "doc_to_target": "{{answer}}",
275
+ "unsafe_code": false,
276
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
277
+ "description": "",
278
+ "target_delimiter": " ",
279
+ "fewshot_delimiter": "\n\n",
280
+ "num_fewshot": 4,
281
+ "metric_list": [
282
+ {
283
+ "metric": "exact_match",
284
+ "aggregation": "mean",
285
+ "higher_is_better": true
286
+ }
287
+ ],
288
+ "output_type": "generate_until",
289
+ "generation_kwargs": {
290
+ "until": [
291
+ "Problem:"
292
+ ],
293
+ "do_sample": false,
294
+ "temperature": 0.0
295
+ },
296
+ "repeats": 1,
297
+ "should_decontaminate": false,
298
+ "metadata": {
299
+ "version": 1.0,
300
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
301
+ }
302
+ },
303
+ "hendrycks_math_precalc": {
304
+ "task": "hendrycks_math_precalc",
305
+ "tag": [
306
+ "math_word_problems"
307
+ ],
308
+ "dataset_path": "EleutherAI/hendrycks_math",
309
+ "dataset_name": "precalculus",
310
+ "training_split": "train",
311
+ "test_split": "test",
312
+ "process_docs": "def process_docs(dataset: datasets.Dataset) -> datasets.Dataset:\n def _process_doc(doc: dict) -> dict:\n out_doc = {\n \"problem\": doc[\"problem\"],\n \"solution\": doc[\"solution\"],\n \"answer\": remove_boxed(last_boxed_only_string(doc[\"solution\"])),\n }\n return out_doc\n\n return dataset.map(_process_doc)\n",
313
+ "doc_to_text": "Problem: {{problem}}\nAnswer:",
314
+ "doc_to_target": "{{answer}}",
315
+ "unsafe_code": false,
316
+ "process_results": "def process_results(doc: dict, results: List[str]) -> Dict[str, int]:\n retval = 0\n indices = [pos for pos, char in enumerate(results[0]) if char == \"$\"]\n if len(indices) <= 1:\n answer = results[0]\n else:\n answer = results[0][indices[0] + 1 : indices[-1]]\n\n if is_equiv(answer, remove_boxed(last_boxed_only_string(doc[\"solution\"]))):\n retval = 1\n\n results = {\n \"exact_match\": retval,\n }\n return results\n",
317
+ "description": "",
318
+ "target_delimiter": " ",
319
+ "fewshot_delimiter": "\n\n",
320
+ "num_fewshot": 4,
321
+ "metric_list": [
322
+ {
323
+ "metric": "exact_match",
324
+ "aggregation": "mean",
325
+ "higher_is_better": true
326
+ }
327
+ ],
328
+ "output_type": "generate_until",
329
+ "generation_kwargs": {
330
+ "until": [
331
+ "Problem:"
332
+ ],
333
+ "do_sample": false,
334
+ "temperature": 0.0
335
+ },
336
+ "repeats": 1,
337
+ "should_decontaminate": false,
338
+ "metadata": {
339
+ "version": 1.0,
340
+ "pretrained": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf"
341
+ }
342
+ }
343
+ },
344
+ "versions": {
345
+ "hendrycks_math": 1.0,
346
+ "hendrycks_math_algebra": 1.0,
347
+ "hendrycks_math_counting_and_prob": 1.0,
348
+ "hendrycks_math_geometry": 1.0,
349
+ "hendrycks_math_intermediate_algebra": 1.0,
350
+ "hendrycks_math_num_theory": 1.0,
351
+ "hendrycks_math_prealgebra": 1.0,
352
+ "hendrycks_math_precalc": 1.0
353
+ },
354
+ "n-shot": {
355
+ "hendrycks_math_algebra": 4,
356
+ "hendrycks_math_counting_and_prob": 4,
357
+ "hendrycks_math_geometry": 4,
358
+ "hendrycks_math_intermediate_algebra": 4,
359
+ "hendrycks_math_num_theory": 4,
360
+ "hendrycks_math_prealgebra": 4,
361
+ "hendrycks_math_precalc": 4
362
+ },
363
+ "higher_is_better": {
364
+ "hendrycks_math": {
365
+ "exact_match": true
366
+ },
367
+ "hendrycks_math_algebra": {
368
+ "exact_match": true
369
+ },
370
+ "hendrycks_math_counting_and_prob": {
371
+ "exact_match": true
372
+ },
373
+ "hendrycks_math_geometry": {
374
+ "exact_match": true
375
+ },
376
+ "hendrycks_math_intermediate_algebra": {
377
+ "exact_match": true
378
+ },
379
+ "hendrycks_math_num_theory": {
380
+ "exact_match": true
381
+ },
382
+ "hendrycks_math_prealgebra": {
383
+ "exact_match": true
384
+ },
385
+ "hendrycks_math_precalc": {
386
+ "exact_match": true
387
+ }
388
+ },
389
+ "n-samples": {
390
+ "hendrycks_math_algebra": {
391
+ "original": 1187,
392
+ "effective": 1187
393
+ },
394
+ "hendrycks_math_counting_and_prob": {
395
+ "original": 474,
396
+ "effective": 474
397
+ },
398
+ "hendrycks_math_geometry": {
399
+ "original": 479,
400
+ "effective": 479
401
+ },
402
+ "hendrycks_math_intermediate_algebra": {
403
+ "original": 903,
404
+ "effective": 903
405
+ },
406
+ "hendrycks_math_num_theory": {
407
+ "original": 540,
408
+ "effective": 540
409
+ },
410
+ "hendrycks_math_prealgebra": {
411
+ "original": 871,
412
+ "effective": 871
413
+ },
414
+ "hendrycks_math_precalc": {
415
+ "original": 546,
416
+ "effective": 546
417
+ }
418
+ },
419
+ "config": {
420
+ "model": "hf",
421
+ "model_args": "pretrained=/home/aiops/zhuty/nanotron/checkpoints/2500_hf,trust_remote_code=True",
422
+ "model_num_parameters": 1235814400,
423
+ "model_dtype": "torch.bfloat16",
424
+ "model_revision": "main",
425
+ "model_sha": "",
426
+ "batch_size": "64",
427
+ "batch_sizes": [],
428
+ "device": null,
429
+ "use_cache": null,
430
+ "limit": null,
431
+ "bootstrap_iters": 100000,
432
+ "gen_kwargs": null,
433
+ "random_seed": 0,
434
+ "numpy_seed": 1234,
435
+ "torch_seed": 1234,
436
+ "fewshot_seed": 1234
437
+ },
438
+ "git_hash": "d19df7aa0d9031b97c04db84db410bcb459de665",
439
+ "date": 1767757704.703941,
440
+ "pretty_env_info": "PyTorch version: 2.7.1+cu126\nIs debug build: False\nCUDA used to build PyTorch: 12.6\nROCM used to build PyTorch: N/A\n\nOS: Ubuntu 22.04.4 LTS (x86_64)\nGCC version: (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0\nClang version: 13.0.0 (https://github.com/apple/llvm-project.git 3dade082a9b1989207a7fa7f3975868485d16a49)\nCMake version: Could not collect\nLibc version: glibc-2.35\n\nPython version: 3.10.18 (main, Jun 5 2025, 13:14:17) [GCC 11.2.0] (64-bit runtime)\nPython platform: Linux-5.4.0-173-generic-x86_64-with-glibc2.35\nIs CUDA available: True\nCUDA runtime version: Could not collect\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB\nNvidia driver version: 535.216.03\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nAddress sizes: 43 bits physical, 48 bits virtual\nByte Order: Little Endian\nCPU(s): 256\nOn-line CPU(s) list: 0-255\nVendor ID: AuthenticAMD\nModel name: AMD EPYC 7742 64-Core Processor\nCPU family: 23\nModel: 49\nThread(s) per core: 2\nCore(s) per socket: 64\nSocket(s): 2\nStepping: 0\nFrequency boost: enabled\nCPU max MHz: 2250.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 4491.80\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc rep_good nopl nonstop_tsc cpuid extd_apicid aperfmperf pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_llc mwaitx cpb cat_l3 cdp_l3 hw_pstate ssbd mba ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 xsaves cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr wbnoinvd arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip rdpid overflow_recov succor smca sme sev sev_es\nVirtualization: AMD-V\nL1d cache: 4 MiB (128 instances)\nL1i cache: 4 MiB (128 instances)\nL2 cache: 64 MiB (128 instances)\nL3 cache: 512 MiB (32 instances)\nNUMA node(s): 8\nNUMA node0 CPU(s): 0-15,128-143\nNUMA node1 CPU(s): 16-31,144-159\nNUMA node2 CPU(s): 32-47,160-175\nNUMA node3 CPU(s): 48-63,176-191\nNUMA node4 CPU(s): 64-79,192-207\nNUMA node5 CPU(s): 80-95,208-223\nNUMA node6 CPU(s): 96-111,224-239\nNUMA node7 CPU(s): 112-127,240-255\nVulnerability Gather data sampling: Not affected\nVulnerability Itlb multihit: Not affected\nVulnerability L1tf: Not affected\nVulnerability Mds: Not affected\nVulnerability Meltdown: Not affected\nVulnerability Mmio stale data: Not affected\nVulnerability Retbleed: Vulnerable\nVulnerability Spec store bypass: Mitigation; Speculative Store Bypass disabled via prctl and seccomp\nVulnerability Spectre v1: Mitigation; usercopy/swapgs barriers and __user pointer sanitization\nVulnerability Spectre v2: Mitigation; Retpolines, IBPB conditional, IBRS_FW, STIBP conditional, RSB filling, PBRSB-eIBRS Not affected\nVulnerability Srbds: Not affected\nVulnerability Tsx async abort: Not affected\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] nvidia-cublas-cu12==12.6.4.1\n[pip3] nvidia-cuda-cupti-cu12==12.6.80\n[pip3] nvidia-cuda-nvrtc-cu12==12.6.77\n[pip3] nvidia-cuda-runtime-cu12==12.6.77\n[pip3] nvidia-cudnn-cu12==9.5.1.17\n[pip3] nvidia-cufft-cu12==11.3.0.4\n[pip3] nvidia-curand-cu12==10.3.7.77\n[pip3] nvidia-cusolver-cu12==11.7.1.2\n[pip3] nvidia-cusparse-cu12==12.5.4.2\n[pip3] nvidia-cusparselt-cu12==0.6.3\n[pip3] nvidia-nccl-cu12==2.26.2\n[pip3] nvidia-nvjitlink-cu12==12.6.85\n[pip3] nvidia-nvtx-cu12==12.6.77\n[pip3] optree==0.16.0\n[pip3] pytorch-lightning==2.5.1.post0\n[pip3] torch==2.7.1\n[pip3] torchmetrics==1.7.2\n[pip3] torchvision==0.22.1\n[pip3] triton==3.3.1\n[conda] numpy 1.26.4 pypi_0 pypi\n[conda] nvidia-cublas-cu12 12.6.4.1 pypi_0 pypi\n[conda] nvidia-cuda-cupti-cu12 12.6.80 pypi_0 pypi\n[conda] nvidia-cuda-nvrtc-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cuda-runtime-cu12 12.6.77 pypi_0 pypi\n[conda] nvidia-cudnn-cu12 9.5.1.17 pypi_0 pypi\n[conda] nvidia-cufft-cu12 11.3.0.4 pypi_0 pypi\n[conda] nvidia-curand-cu12 10.3.7.77 pypi_0 pypi\n[conda] nvidia-cusolver-cu12 11.7.1.2 pypi_0 pypi\n[conda] nvidia-cusparse-cu12 12.5.4.2 pypi_0 pypi\n[conda] nvidia-cusparselt-cu12 0.6.3 pypi_0 pypi\n[conda] nvidia-nccl-cu12 2.26.2 pypi_0 pypi\n[conda] nvidia-nvjitlink-cu12 12.6.85 pypi_0 pypi\n[conda] nvidia-nvtx-cu12 12.6.77 pypi_0 pypi\n[conda] optree 0.16.0 pypi_0 pypi\n[conda] pytorch-lightning 2.5.1.post0 pypi_0 pypi\n[conda] torch 2.7.1 pypi_0 pypi\n[conda] torchmetrics 1.7.2 pypi_0 pypi\n[conda] torchvision 0.22.1 pypi_0 pypi\n[conda] triton 3.3.1 pypi_0 pypi",
441
+ "transformers_version": "4.57.3",
442
+ "lm_eval_version": "0.4.9.1",
443
+ "upper_git_hash": null,
444
+ "tokenizer_pad_token": [
445
+ "<|end_of_text|>",
446
+ "128001"
447
+ ],
448
+ "tokenizer_eos_token": [
449
+ "<|end_of_text|>",
450
+ "128001"
451
+ ],
452
+ "tokenizer_bos_token": [
453
+ "<|begin_of_text|>",
454
+ "128000"
455
+ ],
456
+ "eot_token_id": 128001,
457
+ "max_length": 131072,
458
+ "task_hashes": {
459
+ "hendrycks_math_algebra": "7b50860e4db85f9d690d9679b543412837815a3b9854f0d0d166cfefc428cc23",
460
+ "hendrycks_math_counting_and_prob": "826f9744663d8e2e04e0050aa87b6edc5640c41ca1aa53fee8ef747f7c93b100",
461
+ "hendrycks_math_geometry": "4b67d6f45ba7f89a0c8a0a01fc9b1885bf942346a45b6a4fd03afd3d0e23ab6e",
462
+ "hendrycks_math_intermediate_algebra": "d7bc95882c4f58ef0b7252e6026e98b8bc77bbfbe36de6da6d97127e9744141e",
463
+ "hendrycks_math_num_theory": "ff82fd9f09e0a7c370270b3e7778773f7f1dec4aa40564e1a1318417af819030",
464
+ "hendrycks_math_prealgebra": "e4296905fe5f21ca452a0596371350ffedec630103ac7cce4b76cb806c55be05",
465
+ "hendrycks_math_precalc": "065216e339084dad7a3260b23a5151c8bf63771c7f0d23e88b6f3df4a9f0175a"
466
+ },
467
+ "model_source": "hf",
468
+ "model_name": "/home/aiops/zhuty/nanotron/checkpoints/2500_hf",
469
+ "model_name_sanitized": "__home__aiops__zhuty__nanotron__checkpoints__2500_hf",
470
+ "system_instruction": null,
471
+ "system_instruction_sha": null,
472
+ "fewshot_as_multiturn": false,
473
+ "chat_template": null,
474
+ "chat_template_sha": null,
475
+ "start_time": 77078.278406873,
476
+ "end_time": 77549.351266943,
477
+ "total_evaluation_time_seconds": "471.0728600700095"
478
+ }
2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_hendrycks_math_algebra_2025-12-29T06-24-24.342056.jsonl ADDED
The diff for this file is too large to render. See raw diff
 
2500_hf/harness_eval_4shot/__home__aiops__zhuty__nanotron__checkpoints__2500_hf/samples_hendrycks_math_algebra_2026-01-07T03-55-15.968865.jsonl ADDED
The diff for this file is too large to render. See raw diff