HAissa commited on
Commit
8642c89
·
verified ·
1 Parent(s): 83efbbd

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1024,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 28,
16
+ "model_type": "qwen3",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 28,
19
+ "num_key_value_heads": 8,
20
+ "rms_norm_eps": 1e-06,
21
+ "rope_scaling": null,
22
+ "rope_theta": 1000000,
23
+ "sliding_window": null,
24
+ "tie_word_embeddings": true,
25
+ "torch_dtype": "float16",
26
+ "transformers_version": "4.51.3",
27
+ "use_cache": true,
28
+ "use_sliding_window": false,
29
+ "vocab_size": 151936
30
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_new_tokens": 2048,
5
+ "transformers_version": "4.51.3"
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34daee22f3da79f04a110d67751c885964c9e42cea74eb16b11e1445d65f31e8
3
+ size 1192134784
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea006d31679633f09f87669e36037e0958ed216eafbe6f86af1823fac9c2175b
3
+ size 1211416522
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4c917636c7a58af68a29056522a757e9f9b99005b776641aa157c536967817d
3
+ size 14244
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9b8f34b96b9d90cfe89bdd9aafba61ba227bc92faead103f96edaa3b9fd8c78
3
+ size 1064
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
tokenizer_config.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}\n{%- for message in messages[::-1] %}\n {%- set index = (messages|length - 1) - loop.index0 %}\n {%- if ns.multi_step_tool and message.role == \"user\" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}\n {%- set ns.multi_step_tool = false %}\n {%- set ns.last_query_index = index %}\n {%- endif %}\n{%- endfor %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {%- set content = message.content %}\n {%- set reasoning_content = '' %}\n {%- if message.reasoning_content is defined and message.reasoning_content is not none %}\n {%- set reasoning_content = message.reasoning_content %}\n {%- else %}\n {%- if '</think>' in message.content %}\n {%- set content = message.content.split('</think>')[-1].lstrip('\\n') %}\n {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\\n').split('<think>')[-1].lstrip('\\n') %}\n {%- endif %}\n {%- endif %}\n {%- if loop.index0 > ns.last_query_index %}\n {%- if loop.last or (not loop.last and reasoning_content) %}\n {{- '<|im_start|>' + message.role + '\\n<think>\\n' + reasoning_content.strip('\\n') + '\\n</think>\\n\\n' + content.lstrip('\\n') }}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- if message.tool_calls %}\n {%- for tool_call in message.tool_calls %}\n {%- if (loop.first and content) or (not loop.first) %}\n {{- '\\n' }}\n {%- endif %}\n {%- if tool_call.function %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {%- if tool_call.arguments is string %}\n {{- tool_call.arguments }}\n {%- else %}\n {{- tool_call.arguments | tojson }}\n {%- endif %}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.first or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n {%- if enable_thinking is defined and enable_thinking is false %}\n {{- '<think>\\n\\n</think>\\n\\n' }}\n {%- endif %}\n{%- endif %}",
231
+ "clean_up_tokenization_spaces": false,
232
+ "eos_token": "<|endoftext|>",
233
+ "errors": "replace",
234
+ "extra_special_tokens": {},
235
+ "model_max_length": 131072,
236
+ "pad_token": "<|endoftext|>",
237
+ "split_special_tokens": false,
238
+ "tokenizer_class": "Qwen2Tokenizer",
239
+ "unk_token": null
240
+ }
trainer_state.json ADDED
@@ -0,0 +1,2734 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.5010271055664112,
6
+ "eval_steps": 500,
7
+ "global_step": 15000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.001670090351888037,
14
+ "grad_norm": 7.16015625,
15
+ "learning_rate": 1.9600000000000002e-05,
16
+ "loss": 0.6304,
17
+ "mean_token_accuracy": 0.811416609287262,
18
+ "num_tokens": 150666.0,
19
+ "step": 50
20
+ },
21
+ {
22
+ "epoch": 0.003340180703776074,
23
+ "grad_norm": 5.58984375,
24
+ "learning_rate": 1.999986736222114e-05,
25
+ "loss": 0.4501,
26
+ "mean_token_accuracy": 0.8562266689538955,
27
+ "num_tokens": 307563.0,
28
+ "step": 100
29
+ },
30
+ {
31
+ "epoch": 0.0050102710556641115,
32
+ "grad_norm": 5.0859375,
33
+ "learning_rate": 1.9999458569756985e-05,
34
+ "loss": 0.5118,
35
+ "mean_token_accuracy": 0.838346039056778,
36
+ "num_tokens": 454754.0,
37
+ "step": 150
38
+ },
39
+ {
40
+ "epoch": 0.006680361407552148,
41
+ "grad_norm": 4.37890625,
42
+ "learning_rate": 1.9998773578656093e-05,
43
+ "loss": 0.5075,
44
+ "mean_token_accuracy": 0.8403138118982315,
45
+ "num_tokens": 604733.0,
46
+ "step": 200
47
+ },
48
+ {
49
+ "epoch": 0.008350451759440186,
50
+ "grad_norm": 4.40234375,
51
+ "learning_rate": 1.999781240783885e-05,
52
+ "loss": 0.5057,
53
+ "mean_token_accuracy": 0.8395608925819397,
54
+ "num_tokens": 751094.0,
55
+ "step": 250
56
+ },
57
+ {
58
+ "epoch": 0.010020542111328223,
59
+ "grad_norm": 4.26953125,
60
+ "learning_rate": 1.9996575083854096e-05,
61
+ "loss": 0.5063,
62
+ "mean_token_accuracy": 0.8390217781066894,
63
+ "num_tokens": 906188.0,
64
+ "step": 300
65
+ },
66
+ {
67
+ "epoch": 0.01169063246321626,
68
+ "grad_norm": 5.03125,
69
+ "learning_rate": 1.9995061640878407e-05,
70
+ "loss": 0.5076,
71
+ "mean_token_accuracy": 0.8388406819105149,
72
+ "num_tokens": 1058839.0,
73
+ "step": 350
74
+ },
75
+ {
76
+ "epoch": 0.013360722815104297,
77
+ "grad_norm": 4.41015625,
78
+ "learning_rate": 1.9993272120715135e-05,
79
+ "loss": 0.5248,
80
+ "mean_token_accuracy": 0.83592469394207,
81
+ "num_tokens": 1210117.0,
82
+ "step": 400
83
+ },
84
+ {
85
+ "epoch": 0.015030813166992334,
86
+ "grad_norm": 4.4921875,
87
+ "learning_rate": 1.9991206572793253e-05,
88
+ "loss": 0.5103,
89
+ "mean_token_accuracy": 0.8381919747591019,
90
+ "num_tokens": 1365295.0,
91
+ "step": 450
92
+ },
93
+ {
94
+ "epoch": 0.016700903518880372,
95
+ "grad_norm": 4.69921875,
96
+ "learning_rate": 1.998886505416601e-05,
97
+ "loss": 0.5177,
98
+ "mean_token_accuracy": 0.838459021449089,
99
+ "num_tokens": 1519295.0,
100
+ "step": 500
101
+ },
102
+ {
103
+ "epoch": 0.018370993870768407,
104
+ "grad_norm": 4.6484375,
105
+ "learning_rate": 1.9986247629509327e-05,
106
+ "loss": 0.5784,
107
+ "mean_token_accuracy": 0.8219448757171631,
108
+ "num_tokens": 1670026.0,
109
+ "step": 550
110
+ },
111
+ {
112
+ "epoch": 0.020041084222656446,
113
+ "grad_norm": 4.25,
114
+ "learning_rate": 1.9983354371120035e-05,
115
+ "loss": 0.7304,
116
+ "mean_token_accuracy": 0.7821861469745636,
117
+ "num_tokens": 1830249.0,
118
+ "step": 600
119
+ },
120
+ {
121
+ "epoch": 0.02171117457454448,
122
+ "grad_norm": 3.994140625,
123
+ "learning_rate": 1.998018535891386e-05,
124
+ "loss": 0.7442,
125
+ "mean_token_accuracy": 0.7787324601411819,
126
+ "num_tokens": 1981040.0,
127
+ "step": 650
128
+ },
129
+ {
130
+ "epoch": 0.02338126492643252,
131
+ "grad_norm": 3.548828125,
132
+ "learning_rate": 1.9976740680423227e-05,
133
+ "loss": 0.7469,
134
+ "mean_token_accuracy": 0.7790761750936508,
135
+ "num_tokens": 2133645.0,
136
+ "step": 700
137
+ },
138
+ {
139
+ "epoch": 0.02505135527832056,
140
+ "grad_norm": 3.595703125,
141
+ "learning_rate": 1.997302043079484e-05,
142
+ "loss": 0.7073,
143
+ "mean_token_accuracy": 0.7876522129774094,
144
+ "num_tokens": 2286147.0,
145
+ "step": 750
146
+ },
147
+ {
148
+ "epoch": 0.026721445630208594,
149
+ "grad_norm": 3.787109375,
150
+ "learning_rate": 1.9969024712787045e-05,
151
+ "loss": 0.7117,
152
+ "mean_token_accuracy": 0.7861441165208817,
153
+ "num_tokens": 2435302.0,
154
+ "step": 800
155
+ },
156
+ {
157
+ "epoch": 0.028391535982096632,
158
+ "grad_norm": 4.07421875,
159
+ "learning_rate": 1.996475363676701e-05,
160
+ "loss": 0.7301,
161
+ "mean_token_accuracy": 0.7839706349372864,
162
+ "num_tokens": 2586271.0,
163
+ "step": 850
164
+ },
165
+ {
166
+ "epoch": 0.030061626333984667,
167
+ "grad_norm": 4.15234375,
168
+ "learning_rate": 1.9960207320707655e-05,
169
+ "loss": 0.7237,
170
+ "mean_token_accuracy": 0.7836953294277191,
171
+ "num_tokens": 2736739.0,
172
+ "step": 900
173
+ },
174
+ {
175
+ "epoch": 0.031731716685872706,
176
+ "grad_norm": 4.83984375,
177
+ "learning_rate": 1.9955385890184415e-05,
178
+ "loss": 0.7354,
179
+ "mean_token_accuracy": 0.7828176522254944,
180
+ "num_tokens": 2882560.0,
181
+ "step": 950
182
+ },
183
+ {
184
+ "epoch": 0.033401807037760745,
185
+ "grad_norm": 3.923828125,
186
+ "learning_rate": 1.995028947837175e-05,
187
+ "loss": 0.7236,
188
+ "mean_token_accuracy": 0.7848378765583038,
189
+ "num_tokens": 3043075.0,
190
+ "step": 1000
191
+ },
192
+ {
193
+ "epoch": 0.03507189738964878,
194
+ "grad_norm": 3.59375,
195
+ "learning_rate": 1.994491822603948e-05,
196
+ "loss": 0.7182,
197
+ "mean_token_accuracy": 0.7871662402153015,
198
+ "num_tokens": 3198694.0,
199
+ "step": 1050
200
+ },
201
+ {
202
+ "epoch": 0.036741987741536815,
203
+ "grad_norm": 4.04296875,
204
+ "learning_rate": 1.9939272281548897e-05,
205
+ "loss": 0.7076,
206
+ "mean_token_accuracy": 0.7869892084598541,
207
+ "num_tokens": 3351166.0,
208
+ "step": 1100
209
+ },
210
+ {
211
+ "epoch": 0.03841207809342485,
212
+ "grad_norm": 4.29296875,
213
+ "learning_rate": 1.9933351800848663e-05,
214
+ "loss": 0.7079,
215
+ "mean_token_accuracy": 0.7874267196655274,
216
+ "num_tokens": 3497030.0,
217
+ "step": 1150
218
+ },
219
+ {
220
+ "epoch": 0.04008216844531289,
221
+ "grad_norm": 3.650390625,
222
+ "learning_rate": 1.99271569474705e-05,
223
+ "loss": 0.6844,
224
+ "mean_token_accuracy": 0.7945140707492828,
225
+ "num_tokens": 3650753.0,
226
+ "step": 1200
227
+ },
228
+ {
229
+ "epoch": 0.04175225879720093,
230
+ "grad_norm": 4.3515625,
231
+ "learning_rate": 1.9920687892524676e-05,
232
+ "loss": 0.714,
233
+ "mean_token_accuracy": 0.7859313750267028,
234
+ "num_tokens": 3803593.0,
235
+ "step": 1250
236
+ },
237
+ {
238
+ "epoch": 0.04342234914908896,
239
+ "grad_norm": 3.603515625,
240
+ "learning_rate": 1.9913944814695285e-05,
241
+ "loss": 0.6818,
242
+ "mean_token_accuracy": 0.7935056626796723,
243
+ "num_tokens": 3954344.0,
244
+ "step": 1300
245
+ },
246
+ {
247
+ "epoch": 0.045092439500977,
248
+ "grad_norm": 3.8828125,
249
+ "learning_rate": 1.99069279002353e-05,
250
+ "loss": 0.6986,
251
+ "mean_token_accuracy": 0.7902963298559189,
252
+ "num_tokens": 4109865.0,
253
+ "step": 1350
254
+ },
255
+ {
256
+ "epoch": 0.04676252985286504,
257
+ "grad_norm": 4.125,
258
+ "learning_rate": 1.9899637342961434e-05,
259
+ "loss": 0.6922,
260
+ "mean_token_accuracy": 0.7937425565719605,
261
+ "num_tokens": 4258123.0,
262
+ "step": 1400
263
+ },
264
+ {
265
+ "epoch": 0.04843262020475308,
266
+ "grad_norm": 4.23046875,
267
+ "learning_rate": 1.9892073344248794e-05,
268
+ "loss": 0.6954,
269
+ "mean_token_accuracy": 0.7912928277254104,
270
+ "num_tokens": 4404817.0,
271
+ "step": 1450
272
+ },
273
+ {
274
+ "epoch": 0.05010271055664112,
275
+ "grad_norm": 3.623046875,
276
+ "learning_rate": 1.98842361130253e-05,
277
+ "loss": 0.6794,
278
+ "mean_token_accuracy": 0.7929424160718918,
279
+ "num_tokens": 4552500.0,
280
+ "step": 1500
281
+ },
282
+ {
283
+ "epoch": 0.05177280090852915,
284
+ "grad_norm": 4.046875,
285
+ "learning_rate": 1.987612586576593e-05,
286
+ "loss": 0.7069,
287
+ "mean_token_accuracy": 0.7885507994890213,
288
+ "num_tokens": 4700529.0,
289
+ "step": 1550
290
+ },
291
+ {
292
+ "epoch": 0.05344289126041719,
293
+ "grad_norm": 3.5078125,
294
+ "learning_rate": 1.986774282648674e-05,
295
+ "loss": 0.6928,
296
+ "mean_token_accuracy": 0.7903753596544266,
297
+ "num_tokens": 4854159.0,
298
+ "step": 1600
299
+ },
300
+ {
301
+ "epoch": 0.055112981612305226,
302
+ "grad_norm": 3.23828125,
303
+ "learning_rate": 1.9859087226738667e-05,
304
+ "loss": 0.6862,
305
+ "mean_token_accuracy": 0.7946210086345673,
306
+ "num_tokens": 5004881.0,
307
+ "step": 1650
308
+ },
309
+ {
310
+ "epoch": 0.056783071964193264,
311
+ "grad_norm": 3.8828125,
312
+ "learning_rate": 1.9850159305601142e-05,
313
+ "loss": 0.681,
314
+ "mean_token_accuracy": 0.7951335823535919,
315
+ "num_tokens": 5155074.0,
316
+ "step": 1700
317
+ },
318
+ {
319
+ "epoch": 0.0584531623160813,
320
+ "grad_norm": 3.435546875,
321
+ "learning_rate": 1.984095930967548e-05,
322
+ "loss": 0.6775,
323
+ "mean_token_accuracy": 0.7948978447914123,
324
+ "num_tokens": 5305327.0,
325
+ "step": 1750
326
+ },
327
+ {
328
+ "epoch": 0.060123252667969335,
329
+ "grad_norm": 4.3125,
330
+ "learning_rate": 1.983148749307807e-05,
331
+ "loss": 0.6815,
332
+ "mean_token_accuracy": 0.7944793283939362,
333
+ "num_tokens": 5453165.0,
334
+ "step": 1800
335
+ },
336
+ {
337
+ "epoch": 0.06179334301985737,
338
+ "grad_norm": 3.498046875,
339
+ "learning_rate": 1.9821744117433367e-05,
340
+ "loss": 0.6737,
341
+ "mean_token_accuracy": 0.7980403685569764,
342
+ "num_tokens": 5605423.0,
343
+ "step": 1850
344
+ },
345
+ {
346
+ "epoch": 0.06346343337174541,
347
+ "grad_norm": 4.46484375,
348
+ "learning_rate": 1.9811729451866647e-05,
349
+ "loss": 0.6914,
350
+ "mean_token_accuracy": 0.7934044080972672,
351
+ "num_tokens": 5751905.0,
352
+ "step": 1900
353
+ },
354
+ {
355
+ "epoch": 0.06513352372363344,
356
+ "grad_norm": 3.724609375,
357
+ "learning_rate": 1.980144377299658e-05,
358
+ "loss": 0.6879,
359
+ "mean_token_accuracy": 0.7944984430074692,
360
+ "num_tokens": 5900554.0,
361
+ "step": 1950
362
+ },
363
+ {
364
+ "epoch": 0.06680361407552149,
365
+ "grad_norm": 3.11328125,
366
+ "learning_rate": 1.9790887364927603e-05,
367
+ "loss": 0.6887,
368
+ "mean_token_accuracy": 0.7929914563894271,
369
+ "num_tokens": 6057264.0,
370
+ "step": 2000
371
+ },
372
+ {
373
+ "epoch": 0.06847370442740952,
374
+ "grad_norm": 3.974609375,
375
+ "learning_rate": 1.9780060519242047e-05,
376
+ "loss": 0.6795,
377
+ "mean_token_accuracy": 0.7947278439998626,
378
+ "num_tokens": 6207563.0,
379
+ "step": 2050
380
+ },
381
+ {
382
+ "epoch": 0.07014379477929757,
383
+ "grad_norm": 3.337890625,
384
+ "learning_rate": 1.9768963534992114e-05,
385
+ "loss": 0.695,
386
+ "mean_token_accuracy": 0.7921637284755707,
387
+ "num_tokens": 6362024.0,
388
+ "step": 2100
389
+ },
390
+ {
391
+ "epoch": 0.0718138851311856,
392
+ "grad_norm": 4.35546875,
393
+ "learning_rate": 1.975759671869159e-05,
394
+ "loss": 0.6813,
395
+ "mean_token_accuracy": 0.7945687365531922,
396
+ "num_tokens": 6511612.0,
397
+ "step": 2150
398
+ },
399
+ {
400
+ "epoch": 0.07348397548307363,
401
+ "grad_norm": 3.443359375,
402
+ "learning_rate": 1.9745960384307384e-05,
403
+ "loss": 0.6664,
404
+ "mean_token_accuracy": 0.7972371423244476,
405
+ "num_tokens": 6664163.0,
406
+ "step": 2200
407
+ },
408
+ {
409
+ "epoch": 0.07515406583496168,
410
+ "grad_norm": 3.927734375,
411
+ "learning_rate": 1.9734054853250876e-05,
412
+ "loss": 0.6745,
413
+ "mean_token_accuracy": 0.7965255242586136,
414
+ "num_tokens": 6818983.0,
415
+ "step": 2250
416
+ },
417
+ {
418
+ "epoch": 0.0768241561868497,
419
+ "grad_norm": 4.40625,
420
+ "learning_rate": 1.9721880454369012e-05,
421
+ "loss": 0.6775,
422
+ "mean_token_accuracy": 0.7949534428119659,
423
+ "num_tokens": 6969178.0,
424
+ "step": 2300
425
+ },
426
+ {
427
+ "epoch": 0.07849424653873775,
428
+ "grad_norm": 3.94140625,
429
+ "learning_rate": 1.9709437523935234e-05,
430
+ "loss": 0.6933,
431
+ "mean_token_accuracy": 0.7906970298290252,
432
+ "num_tokens": 7120862.0,
433
+ "step": 2350
434
+ },
435
+ {
436
+ "epoch": 0.08016433689062578,
437
+ "grad_norm": 4.1484375,
438
+ "learning_rate": 1.9696726405640194e-05,
439
+ "loss": 0.6678,
440
+ "mean_token_accuracy": 0.7977210181951523,
441
+ "num_tokens": 7273755.0,
442
+ "step": 2400
443
+ },
444
+ {
445
+ "epoch": 0.08183442724251382,
446
+ "grad_norm": 3.671875,
447
+ "learning_rate": 1.9683747450582257e-05,
448
+ "loss": 0.7022,
449
+ "mean_token_accuracy": 0.7893240767717361,
450
+ "num_tokens": 7423742.0,
451
+ "step": 2450
452
+ },
453
+ {
454
+ "epoch": 0.08350451759440186,
455
+ "grad_norm": 3.5078125,
456
+ "learning_rate": 1.9670501017257805e-05,
457
+ "loss": 0.6847,
458
+ "mean_token_accuracy": 0.7929366368055344,
459
+ "num_tokens": 7574222.0,
460
+ "step": 2500
461
+ },
462
+ {
463
+ "epoch": 0.0851746079462899,
464
+ "grad_norm": 3.384765625,
465
+ "learning_rate": 1.965698747155133e-05,
466
+ "loss": 0.6571,
467
+ "mean_token_accuracy": 0.7997360718250275,
468
+ "num_tokens": 7730085.0,
469
+ "step": 2550
470
+ },
471
+ {
472
+ "epoch": 0.08684469829817792,
473
+ "grad_norm": 3.466796875,
474
+ "learning_rate": 1.9643207186725324e-05,
475
+ "loss": 0.6672,
476
+ "mean_token_accuracy": 0.7982296937704086,
477
+ "num_tokens": 7881768.0,
478
+ "step": 2600
479
+ },
480
+ {
481
+ "epoch": 0.08851478865006597,
482
+ "grad_norm": 3.521484375,
483
+ "learning_rate": 1.962916054340999e-05,
484
+ "loss": 0.685,
485
+ "mean_token_accuracy": 0.7952978044748307,
486
+ "num_tokens": 8027897.0,
487
+ "step": 2650
488
+ },
489
+ {
490
+ "epoch": 0.090184879001954,
491
+ "grad_norm": 3.79296875,
492
+ "learning_rate": 1.961484792959271e-05,
493
+ "loss": 0.6778,
494
+ "mean_token_accuracy": 0.7966848593950272,
495
+ "num_tokens": 8175954.0,
496
+ "step": 2700
497
+ },
498
+ {
499
+ "epoch": 0.09185496935384205,
500
+ "grad_norm": 3.40234375,
501
+ "learning_rate": 1.9600269740607328e-05,
502
+ "loss": 0.6488,
503
+ "mean_token_accuracy": 0.8027435272932053,
504
+ "num_tokens": 8332915.0,
505
+ "step": 2750
506
+ },
507
+ {
508
+ "epoch": 0.09352505970573008,
509
+ "grad_norm": 3.408203125,
510
+ "learning_rate": 1.958542637912324e-05,
511
+ "loss": 0.673,
512
+ "mean_token_accuracy": 0.7967682188749313,
513
+ "num_tokens": 8485109.0,
514
+ "step": 2800
515
+ },
516
+ {
517
+ "epoch": 0.09519515005761811,
518
+ "grad_norm": 3.244140625,
519
+ "learning_rate": 1.9570318255134263e-05,
520
+ "loss": 0.6544,
521
+ "mean_token_accuracy": 0.7994941908121109,
522
+ "num_tokens": 8638846.0,
523
+ "step": 2850
524
+ },
525
+ {
526
+ "epoch": 0.09686524040950616,
527
+ "grad_norm": 3.62109375,
528
+ "learning_rate": 1.9554945785947314e-05,
529
+ "loss": 0.67,
530
+ "mean_token_accuracy": 0.7985292530059814,
531
+ "num_tokens": 8788759.0,
532
+ "step": 2900
533
+ },
534
+ {
535
+ "epoch": 0.09853533076139419,
536
+ "grad_norm": 3.466796875,
537
+ "learning_rate": 1.953930939617089e-05,
538
+ "loss": 0.6722,
539
+ "mean_token_accuracy": 0.7943358135223388,
540
+ "num_tokens": 8941895.0,
541
+ "step": 2950
542
+ },
543
+ {
544
+ "epoch": 0.10020542111328223,
545
+ "grad_norm": 3.6796875,
546
+ "learning_rate": 1.952340951770333e-05,
547
+ "loss": 0.6488,
548
+ "mean_token_accuracy": 0.8029305410385131,
549
+ "num_tokens": 9093399.0,
550
+ "step": 3000
551
+ },
552
+ {
553
+ "epoch": 0.10187551146517027,
554
+ "grad_norm": 3.673828125,
555
+ "learning_rate": 1.9507246589720887e-05,
556
+ "loss": 0.6591,
557
+ "mean_token_accuracy": 0.7998628175258636,
558
+ "num_tokens": 9246318.0,
559
+ "step": 3050
560
+ },
561
+ {
562
+ "epoch": 0.1035456018170583,
563
+ "grad_norm": 3.796875,
564
+ "learning_rate": 1.9490821058665597e-05,
565
+ "loss": 0.6842,
566
+ "mean_token_accuracy": 0.7952588319778442,
567
+ "num_tokens": 9393063.0,
568
+ "step": 3100
569
+ },
570
+ {
571
+ "epoch": 0.10521569216894634,
572
+ "grad_norm": 3.64453125,
573
+ "learning_rate": 1.9474133378232955e-05,
574
+ "loss": 0.6627,
575
+ "mean_token_accuracy": 0.7986302423477173,
576
+ "num_tokens": 9546557.0,
577
+ "step": 3150
578
+ },
579
+ {
580
+ "epoch": 0.10688578252083437,
581
+ "grad_norm": 3.662109375,
582
+ "learning_rate": 1.945718400935938e-05,
583
+ "loss": 0.6564,
584
+ "mean_token_accuracy": 0.8012412291765213,
585
+ "num_tokens": 9698417.0,
586
+ "step": 3200
587
+ },
588
+ {
589
+ "epoch": 0.10855587287272242,
590
+ "grad_norm": 3.21484375,
591
+ "learning_rate": 1.9439973420209467e-05,
592
+ "loss": 0.67,
593
+ "mean_token_accuracy": 0.7971057319641113,
594
+ "num_tokens": 9846015.0,
595
+ "step": 3250
596
+ },
597
+ {
598
+ "epoch": 0.11022596322461045,
599
+ "grad_norm": 3.294921875,
600
+ "learning_rate": 1.9422502086163093e-05,
601
+ "loss": 0.6672,
602
+ "mean_token_accuracy": 0.797788228392601,
603
+ "num_tokens": 10001053.0,
604
+ "step": 3300
605
+ },
606
+ {
607
+ "epoch": 0.11189605357649848,
608
+ "grad_norm": 3.607421875,
609
+ "learning_rate": 1.9404770489802242e-05,
610
+ "loss": 0.6529,
611
+ "mean_token_accuracy": 0.7996513998508453,
612
+ "num_tokens": 10151845.0,
613
+ "step": 3350
614
+ },
615
+ {
616
+ "epoch": 0.11356614392838653,
617
+ "grad_norm": 3.240234375,
618
+ "learning_rate": 1.9386779120897712e-05,
619
+ "loss": 0.6268,
620
+ "mean_token_accuracy": 0.8078747218847275,
621
+ "num_tokens": 10303872.0,
622
+ "step": 3400
623
+ },
624
+ {
625
+ "epoch": 0.11523623428027456,
626
+ "grad_norm": 3.38671875,
627
+ "learning_rate": 1.936852847639556e-05,
628
+ "loss": 0.6608,
629
+ "mean_token_accuracy": 0.8000670689344406,
630
+ "num_tokens": 10456152.0,
631
+ "step": 3450
632
+ },
633
+ {
634
+ "epoch": 0.1169063246321626,
635
+ "grad_norm": 3.337890625,
636
+ "learning_rate": 1.9350019060403398e-05,
637
+ "loss": 0.6469,
638
+ "mean_token_accuracy": 0.8042119026184082,
639
+ "num_tokens": 10605108.0,
640
+ "step": 3500
641
+ },
642
+ {
643
+ "epoch": 0.11857641498405064,
644
+ "grad_norm": 3.521484375,
645
+ "learning_rate": 1.933125138417645e-05,
646
+ "loss": 0.6506,
647
+ "mean_token_accuracy": 0.8032512718439102,
648
+ "num_tokens": 10749490.0,
649
+ "step": 3550
650
+ },
651
+ {
652
+ "epoch": 0.12024650533593867,
653
+ "grad_norm": 3.251953125,
654
+ "learning_rate": 1.9312225966103445e-05,
655
+ "loss": 0.6753,
656
+ "mean_token_accuracy": 0.7979339599609375,
657
+ "num_tokens": 10899284.0,
658
+ "step": 3600
659
+ },
660
+ {
661
+ "epoch": 0.12191659568782671,
662
+ "grad_norm": 3.548828125,
663
+ "learning_rate": 1.9292943331692285e-05,
664
+ "loss": 0.6663,
665
+ "mean_token_accuracy": 0.7995783388614655,
666
+ "num_tokens": 11046487.0,
667
+ "step": 3650
668
+ },
669
+ {
670
+ "epoch": 0.12358668603971475,
671
+ "grad_norm": 3.27734375,
672
+ "learning_rate": 1.9273404013555544e-05,
673
+ "loss": 0.6481,
674
+ "mean_token_accuracy": 0.8006290888786316,
675
+ "num_tokens": 11197224.0,
676
+ "step": 3700
677
+ },
678
+ {
679
+ "epoch": 0.1252567763916028,
680
+ "grad_norm": 3.7109375,
681
+ "learning_rate": 1.9253608551395745e-05,
682
+ "loss": 0.6645,
683
+ "mean_token_accuracy": 0.7999362868070602,
684
+ "num_tokens": 11349738.0,
685
+ "step": 3750
686
+ },
687
+ {
688
+ "epoch": 0.12692686674349082,
689
+ "grad_norm": 4.05078125,
690
+ "learning_rate": 1.923355749199046e-05,
691
+ "loss": 0.644,
692
+ "mean_token_accuracy": 0.8039193218946457,
693
+ "num_tokens": 11501057.0,
694
+ "step": 3800
695
+ },
696
+ {
697
+ "epoch": 0.12859695709537886,
698
+ "grad_norm": 3.26171875,
699
+ "learning_rate": 1.92132513891772e-05,
700
+ "loss": 0.6398,
701
+ "mean_token_accuracy": 0.8046829622983932,
702
+ "num_tokens": 11656329.0,
703
+ "step": 3850
704
+ },
705
+ {
706
+ "epoch": 0.1302670474472669,
707
+ "grad_norm": 3.41796875,
708
+ "learning_rate": 1.919269080383813e-05,
709
+ "loss": 0.6595,
710
+ "mean_token_accuracy": 0.7993488019704819,
711
+ "num_tokens": 11800257.0,
712
+ "step": 3900
713
+ },
714
+ {
715
+ "epoch": 0.13193713779915495,
716
+ "grad_norm": 3.822265625,
717
+ "learning_rate": 1.9171876303884555e-05,
718
+ "loss": 0.6533,
719
+ "mean_token_accuracy": 0.8011213219165803,
720
+ "num_tokens": 11952176.0,
721
+ "step": 3950
722
+ },
723
+ {
724
+ "epoch": 0.13360722815104298,
725
+ "grad_norm": 3.0859375,
726
+ "learning_rate": 1.9150808464241257e-05,
727
+ "loss": 0.6443,
728
+ "mean_token_accuracy": 0.8050255984067917,
729
+ "num_tokens": 12103443.0,
730
+ "step": 4000
731
+ },
732
+ {
733
+ "epoch": 0.135277318502931,
734
+ "grad_norm": 3.5078125,
735
+ "learning_rate": 1.9129487866830602e-05,
736
+ "loss": 0.6533,
737
+ "mean_token_accuracy": 0.8013356167078018,
738
+ "num_tokens": 12254997.0,
739
+ "step": 4050
740
+ },
741
+ {
742
+ "epoch": 0.13694740885481904,
743
+ "grad_norm": 3.30078125,
744
+ "learning_rate": 1.9107915100556473e-05,
745
+ "loss": 0.6532,
746
+ "mean_token_accuracy": 0.8014611840248108,
747
+ "num_tokens": 12408985.0,
748
+ "step": 4100
749
+ },
750
+ {
751
+ "epoch": 0.13861749920670707,
752
+ "grad_norm": 3.14453125,
753
+ "learning_rate": 1.908609076128799e-05,
754
+ "loss": 0.6575,
755
+ "mean_token_accuracy": 0.8021772617101669,
756
+ "num_tokens": 12559124.0,
757
+ "step": 4150
758
+ },
759
+ {
760
+ "epoch": 0.14028758955859513,
761
+ "grad_norm": 3.515625,
762
+ "learning_rate": 1.906401545184307e-05,
763
+ "loss": 0.6475,
764
+ "mean_token_accuracy": 0.8028253322839737,
765
+ "num_tokens": 12707796.0,
766
+ "step": 4200
767
+ },
768
+ {
769
+ "epoch": 0.14195767991048316,
770
+ "grad_norm": 3.646484375,
771
+ "learning_rate": 1.9041689781971765e-05,
772
+ "loss": 0.6506,
773
+ "mean_token_accuracy": 0.8031806921958924,
774
+ "num_tokens": 12858676.0,
775
+ "step": 4250
776
+ },
777
+ {
778
+ "epoch": 0.1436277702623712,
779
+ "grad_norm": 4.37109375,
780
+ "learning_rate": 1.901911436833942e-05,
781
+ "loss": 0.6558,
782
+ "mean_token_accuracy": 0.8009441828727722,
783
+ "num_tokens": 13007021.0,
784
+ "step": 4300
785
+ },
786
+ {
787
+ "epoch": 0.14529786061425923,
788
+ "grad_norm": 3.494140625,
789
+ "learning_rate": 1.899628983450964e-05,
790
+ "loss": 0.629,
791
+ "mean_token_accuracy": 0.8062937146425248,
792
+ "num_tokens": 13155780.0,
793
+ "step": 4350
794
+ },
795
+ {
796
+ "epoch": 0.14696795096614726,
797
+ "grad_norm": 2.982421875,
798
+ "learning_rate": 1.8973216810927077e-05,
799
+ "loss": 0.6461,
800
+ "mean_token_accuracy": 0.8051383906602859,
801
+ "num_tokens": 13302653.0,
802
+ "step": 4400
803
+ },
804
+ {
805
+ "epoch": 0.14863804131803532,
806
+ "grad_norm": 3.529296875,
807
+ "learning_rate": 1.89498959349e-05,
808
+ "loss": 0.6367,
809
+ "mean_token_accuracy": 0.805459765791893,
810
+ "num_tokens": 13455088.0,
811
+ "step": 4450
812
+ },
813
+ {
814
+ "epoch": 0.15030813166992335,
815
+ "grad_norm": 3.228515625,
816
+ "learning_rate": 1.8926327850582706e-05,
817
+ "loss": 0.6288,
818
+ "mean_token_accuracy": 0.8068876641988755,
819
+ "num_tokens": 13609983.0,
820
+ "step": 4500
821
+ },
822
+ {
823
+ "epoch": 0.15197822202181138,
824
+ "grad_norm": 3.72265625,
825
+ "learning_rate": 1.8902513208957712e-05,
826
+ "loss": 0.6521,
827
+ "mean_token_accuracy": 0.8004961597919464,
828
+ "num_tokens": 13759536.0,
829
+ "step": 4550
830
+ },
831
+ {
832
+ "epoch": 0.1536483123736994,
833
+ "grad_norm": 3.83203125,
834
+ "learning_rate": 1.8878452667817785e-05,
835
+ "loss": 0.6314,
836
+ "mean_token_accuracy": 0.8060730373859406,
837
+ "num_tokens": 13914544.0,
838
+ "step": 4600
839
+ },
840
+ {
841
+ "epoch": 0.15531840272558745,
842
+ "grad_norm": 3.88671875,
843
+ "learning_rate": 1.8854146891747788e-05,
844
+ "loss": 0.6397,
845
+ "mean_token_accuracy": 0.805094587802887,
846
+ "num_tokens": 14064759.0,
847
+ "step": 4650
848
+ },
849
+ {
850
+ "epoch": 0.1569884930774755,
851
+ "grad_norm": 3.373046875,
852
+ "learning_rate": 1.8829596552106276e-05,
853
+ "loss": 0.6477,
854
+ "mean_token_accuracy": 0.8041272169351578,
855
+ "num_tokens": 14213937.0,
856
+ "step": 4700
857
+ },
858
+ {
859
+ "epoch": 0.15865858342936354,
860
+ "grad_norm": 3.1484375,
861
+ "learning_rate": 1.8804802327007005e-05,
862
+ "loss": 0.63,
863
+ "mean_token_accuracy": 0.8083387905359268,
864
+ "num_tokens": 14363168.0,
865
+ "step": 4750
866
+ },
867
+ {
868
+ "epoch": 0.16032867378125157,
869
+ "grad_norm": 3.6328125,
870
+ "learning_rate": 1.8779764901300174e-05,
871
+ "loss": 0.633,
872
+ "mean_token_accuracy": 0.8068809920549392,
873
+ "num_tokens": 14511952.0,
874
+ "step": 4800
875
+ },
876
+ {
877
+ "epoch": 0.1619987641331396,
878
+ "grad_norm": 3.693359375,
879
+ "learning_rate": 1.875448496655351e-05,
880
+ "loss": 0.6529,
881
+ "mean_token_accuracy": 0.8010576856136322,
882
+ "num_tokens": 14659521.0,
883
+ "step": 4850
884
+ },
885
+ {
886
+ "epoch": 0.16366885448502763,
887
+ "grad_norm": 3.970703125,
888
+ "learning_rate": 1.8728963221033168e-05,
889
+ "loss": 0.6299,
890
+ "mean_token_accuracy": 0.8058782732486724,
891
+ "num_tokens": 14812856.0,
892
+ "step": 4900
893
+ },
894
+ {
895
+ "epoch": 0.16533894483691566,
896
+ "grad_norm": 3.056640625,
897
+ "learning_rate": 1.870320036968445e-05,
898
+ "loss": 0.6102,
899
+ "mean_token_accuracy": 0.8117638313770295,
900
+ "num_tokens": 14968517.0,
901
+ "step": 4950
902
+ },
903
+ {
904
+ "epoch": 0.16700903518880372,
905
+ "grad_norm": 3.572265625,
906
+ "learning_rate": 1.867719712411233e-05,
907
+ "loss": 0.6454,
908
+ "mean_token_accuracy": 0.8036060810089112,
909
+ "num_tokens": 15120586.0,
910
+ "step": 5000
911
+ },
912
+ {
913
+ "epoch": 0.16867912554069175,
914
+ "grad_norm": 3.0390625,
915
+ "learning_rate": 1.8650954202561786e-05,
916
+ "loss": 0.6341,
917
+ "mean_token_accuracy": 0.8054838758707047,
918
+ "num_tokens": 15272994.0,
919
+ "step": 5050
920
+ },
921
+ {
922
+ "epoch": 0.1703492158925798,
923
+ "grad_norm": 2.947265625,
924
+ "learning_rate": 1.8624472329897985e-05,
925
+ "loss": 0.6328,
926
+ "mean_token_accuracy": 0.8070685297250748,
927
+ "num_tokens": 15423469.0,
928
+ "step": 5100
929
+ },
930
+ {
931
+ "epoch": 0.17201930624446782,
932
+ "grad_norm": 3.240234375,
933
+ "learning_rate": 1.8597752237586242e-05,
934
+ "loss": 0.6228,
935
+ "mean_token_accuracy": 0.809105772972107,
936
+ "num_tokens": 15568404.0,
937
+ "step": 5150
938
+ },
939
+ {
940
+ "epoch": 0.17368939659635585,
941
+ "grad_norm": 3.26171875,
942
+ "learning_rate": 1.8570794663671824e-05,
943
+ "loss": 0.6128,
944
+ "mean_token_accuracy": 0.811137170791626,
945
+ "num_tokens": 15721906.0,
946
+ "step": 5200
947
+ },
948
+ {
949
+ "epoch": 0.1753594869482439,
950
+ "grad_norm": 3.345703125,
951
+ "learning_rate": 1.8543600352759563e-05,
952
+ "loss": 0.6207,
953
+ "mean_token_accuracy": 0.8088931626081467,
954
+ "num_tokens": 15867611.0,
955
+ "step": 5250
956
+ },
957
+ {
958
+ "epoch": 0.17702957730013194,
959
+ "grad_norm": 3.76171875,
960
+ "learning_rate": 1.851617005599329e-05,
961
+ "loss": 0.6198,
962
+ "mean_token_accuracy": 0.810358560681343,
963
+ "num_tokens": 16017301.0,
964
+ "step": 5300
965
+ },
966
+ {
967
+ "epoch": 0.17869966765201997,
968
+ "grad_norm": 3.484375,
969
+ "learning_rate": 1.8488504531035084e-05,
970
+ "loss": 0.6354,
971
+ "mean_token_accuracy": 0.806412605047226,
972
+ "num_tokens": 16174153.0,
973
+ "step": 5350
974
+ },
975
+ {
976
+ "epoch": 0.180369758003908,
977
+ "grad_norm": 3.6015625,
978
+ "learning_rate": 1.8460604542044344e-05,
979
+ "loss": 0.6206,
980
+ "mean_token_accuracy": 0.809535750746727,
981
+ "num_tokens": 16326166.0,
982
+ "step": 5400
983
+ },
984
+ {
985
+ "epoch": 0.18203984835579604,
986
+ "grad_norm": 4.4765625,
987
+ "learning_rate": 1.8432470859656687e-05,
988
+ "loss": 0.6399,
989
+ "mean_token_accuracy": 0.8042701333761215,
990
+ "num_tokens": 16476025.0,
991
+ "step": 5450
992
+ },
993
+ {
994
+ "epoch": 0.1837099387076841,
995
+ "grad_norm": 3.203125,
996
+ "learning_rate": 1.8404104260962662e-05,
997
+ "loss": 0.6148,
998
+ "mean_token_accuracy": 0.811469207406044,
999
+ "num_tokens": 16622887.0,
1000
+ "step": 5500
1001
+ },
1002
+ {
1003
+ "epoch": 0.18538002905957213,
1004
+ "grad_norm": 3.763671875,
1005
+ "learning_rate": 1.8375505529486283e-05,
1006
+ "loss": 0.6266,
1007
+ "mean_token_accuracy": 0.8079740327596664,
1008
+ "num_tokens": 16772669.0,
1009
+ "step": 5550
1010
+ },
1011
+ {
1012
+ "epoch": 0.18705011941146016,
1013
+ "grad_norm": 3.28515625,
1014
+ "learning_rate": 1.8346675455163376e-05,
1015
+ "loss": 0.6151,
1016
+ "mean_token_accuracy": 0.8107228165864945,
1017
+ "num_tokens": 16920269.0,
1018
+ "step": 5600
1019
+ },
1020
+ {
1021
+ "epoch": 0.1887202097633482,
1022
+ "grad_norm": 4.6015625,
1023
+ "learning_rate": 1.8317614834319783e-05,
1024
+ "loss": 0.6321,
1025
+ "mean_token_accuracy": 0.8072718501091003,
1026
+ "num_tokens": 17067401.0,
1027
+ "step": 5650
1028
+ },
1029
+ {
1030
+ "epoch": 0.19039030011523622,
1031
+ "grad_norm": 4.953125,
1032
+ "learning_rate": 1.828832446964935e-05,
1033
+ "loss": 0.635,
1034
+ "mean_token_accuracy": 0.8083576399087906,
1035
+ "num_tokens": 17210399.0,
1036
+ "step": 5700
1037
+ },
1038
+ {
1039
+ "epoch": 0.19206039046712428,
1040
+ "grad_norm": 3.9453125,
1041
+ "learning_rate": 1.825880517019176e-05,
1042
+ "loss": 0.6101,
1043
+ "mean_token_accuracy": 0.8119984632730484,
1044
+ "num_tokens": 17355215.0,
1045
+ "step": 5750
1046
+ },
1047
+ {
1048
+ "epoch": 0.1937304808190123,
1049
+ "grad_norm": 3.5390625,
1050
+ "learning_rate": 1.8229057751310188e-05,
1051
+ "loss": 0.6179,
1052
+ "mean_token_accuracy": 0.8104551881551743,
1053
+ "num_tokens": 17501111.0,
1054
+ "step": 5800
1055
+ },
1056
+ {
1057
+ "epoch": 0.19540057117090034,
1058
+ "grad_norm": 3.25,
1059
+ "learning_rate": 1.8199083034668774e-05,
1060
+ "loss": 0.6245,
1061
+ "mean_token_accuracy": 0.8110457295179367,
1062
+ "num_tokens": 17655096.0,
1063
+ "step": 5850
1064
+ },
1065
+ {
1066
+ "epoch": 0.19707066152278838,
1067
+ "grad_norm": 3.59375,
1068
+ "learning_rate": 1.816888184820993e-05,
1069
+ "loss": 0.6302,
1070
+ "mean_token_accuracy": 0.8070375627279281,
1071
+ "num_tokens": 17809403.0,
1072
+ "step": 5900
1073
+ },
1074
+ {
1075
+ "epoch": 0.1987407518746764,
1076
+ "grad_norm": 3.259765625,
1077
+ "learning_rate": 1.8138455026131477e-05,
1078
+ "loss": 0.6078,
1079
+ "mean_token_accuracy": 0.8126195806264878,
1080
+ "num_tokens": 17959815.0,
1081
+ "step": 5950
1082
+ },
1083
+ {
1084
+ "epoch": 0.20041084222656447,
1085
+ "grad_norm": 3.291015625,
1086
+ "learning_rate": 1.81078034088636e-05,
1087
+ "loss": 0.6,
1088
+ "mean_token_accuracy": 0.8153526914119721,
1089
+ "num_tokens": 18112482.0,
1090
+ "step": 6000
1091
+ },
1092
+ {
1093
+ "epoch": 0.2020809325784525,
1094
+ "grad_norm": 3.619140625,
1095
+ "learning_rate": 1.807692784304562e-05,
1096
+ "loss": 0.6088,
1097
+ "mean_token_accuracy": 0.8110599368810654,
1098
+ "num_tokens": 18269537.0,
1099
+ "step": 6050
1100
+ },
1101
+ {
1102
+ "epoch": 0.20375102293034053,
1103
+ "grad_norm": 3.349609375,
1104
+ "learning_rate": 1.8045829181502642e-05,
1105
+ "loss": 0.62,
1106
+ "mean_token_accuracy": 0.8105544656515121,
1107
+ "num_tokens": 18418714.0,
1108
+ "step": 6100
1109
+ },
1110
+ {
1111
+ "epoch": 0.20542111328222856,
1112
+ "grad_norm": 3.341796875,
1113
+ "learning_rate": 1.8014508283221953e-05,
1114
+ "loss": 0.6249,
1115
+ "mean_token_accuracy": 0.807090380191803,
1116
+ "num_tokens": 18577084.0,
1117
+ "step": 6150
1118
+ },
1119
+ {
1120
+ "epoch": 0.2070912036341166,
1121
+ "grad_norm": 3.5546875,
1122
+ "learning_rate": 1.798296601332935e-05,
1123
+ "loss": 0.6166,
1124
+ "mean_token_accuracy": 0.8106195080280304,
1125
+ "num_tokens": 18723440.0,
1126
+ "step": 6200
1127
+ },
1128
+ {
1129
+ "epoch": 0.20876129398600465,
1130
+ "grad_norm": 3.4921875,
1131
+ "learning_rate": 1.7951203243065184e-05,
1132
+ "loss": 0.6169,
1133
+ "mean_token_accuracy": 0.8116037601232529,
1134
+ "num_tokens": 18870030.0,
1135
+ "step": 6250
1136
+ },
1137
+ {
1138
+ "epoch": 0.21043138433789269,
1139
+ "grad_norm": 3.80859375,
1140
+ "learning_rate": 1.7919220849760347e-05,
1141
+ "loss": 0.6093,
1142
+ "mean_token_accuracy": 0.8129760199785232,
1143
+ "num_tokens": 19020540.0,
1144
+ "step": 6300
1145
+ },
1146
+ {
1147
+ "epoch": 0.21210147468978072,
1148
+ "grad_norm": 3.212890625,
1149
+ "learning_rate": 1.7887019716812015e-05,
1150
+ "loss": 0.6202,
1151
+ "mean_token_accuracy": 0.8104865264892578,
1152
+ "num_tokens": 19171916.0,
1153
+ "step": 6350
1154
+ },
1155
+ {
1156
+ "epoch": 0.21377156504166875,
1157
+ "grad_norm": 3.998046875,
1158
+ "learning_rate": 1.785460073365924e-05,
1159
+ "loss": 0.6243,
1160
+ "mean_token_accuracy": 0.810001392364502,
1161
+ "num_tokens": 19317653.0,
1162
+ "step": 6400
1163
+ },
1164
+ {
1165
+ "epoch": 0.21544165539355678,
1166
+ "grad_norm": 3.6015625,
1167
+ "learning_rate": 1.782196479575841e-05,
1168
+ "loss": 0.5972,
1169
+ "mean_token_accuracy": 0.815451757311821,
1170
+ "num_tokens": 19469488.0,
1171
+ "step": 6450
1172
+ },
1173
+ {
1174
+ "epoch": 0.21711174574544484,
1175
+ "grad_norm": 3.171875,
1176
+ "learning_rate": 1.778911280455848e-05,
1177
+ "loss": 0.6094,
1178
+ "mean_token_accuracy": 0.8127148807048797,
1179
+ "num_tokens": 19617524.0,
1180
+ "step": 6500
1181
+ },
1182
+ {
1183
+ "epoch": 0.21878183609733287,
1184
+ "grad_norm": 3.478515625,
1185
+ "learning_rate": 1.7756045667476095e-05,
1186
+ "loss": 0.6365,
1187
+ "mean_token_accuracy": 0.8052962213754654,
1188
+ "num_tokens": 19766937.0,
1189
+ "step": 6550
1190
+ },
1191
+ {
1192
+ "epoch": 0.2204519264492209,
1193
+ "grad_norm": 3.943359375,
1194
+ "learning_rate": 1.7722764297870534e-05,
1195
+ "loss": 0.6008,
1196
+ "mean_token_accuracy": 0.8136671632528305,
1197
+ "num_tokens": 19918307.0,
1198
+ "step": 6600
1199
+ },
1200
+ {
1201
+ "epoch": 0.22212201680110893,
1202
+ "grad_norm": 3.169921875,
1203
+ "learning_rate": 1.768926961501846e-05,
1204
+ "loss": 0.6066,
1205
+ "mean_token_accuracy": 0.8141713869571686,
1206
+ "num_tokens": 20066541.0,
1207
+ "step": 6650
1208
+ },
1209
+ {
1210
+ "epoch": 0.22379210715299697,
1211
+ "grad_norm": 3.19921875,
1212
+ "learning_rate": 1.7655562544088534e-05,
1213
+ "loss": 0.6034,
1214
+ "mean_token_accuracy": 0.8118478727340698,
1215
+ "num_tokens": 20221668.0,
1216
+ "step": 6700
1217
+ },
1218
+ {
1219
+ "epoch": 0.22546219750488503,
1220
+ "grad_norm": 3.275390625,
1221
+ "learning_rate": 1.762164401611587e-05,
1222
+ "loss": 0.6189,
1223
+ "mean_token_accuracy": 0.8125155258178711,
1224
+ "num_tokens": 20368732.0,
1225
+ "step": 6750
1226
+ },
1227
+ {
1228
+ "epoch": 0.22713228785677306,
1229
+ "grad_norm": 3.462890625,
1230
+ "learning_rate": 1.7587514967976315e-05,
1231
+ "loss": 0.6095,
1232
+ "mean_token_accuracy": 0.8130715191364288,
1233
+ "num_tokens": 20519573.0,
1234
+ "step": 6800
1235
+ },
1236
+ {
1237
+ "epoch": 0.2288023782086611,
1238
+ "grad_norm": 3.05859375,
1239
+ "learning_rate": 1.7553176342360564e-05,
1240
+ "loss": 0.6381,
1241
+ "mean_token_accuracy": 0.8038551956415176,
1242
+ "num_tokens": 20668027.0,
1243
+ "step": 6850
1244
+ },
1245
+ {
1246
+ "epoch": 0.23047246856054912,
1247
+ "grad_norm": 3.392578125,
1248
+ "learning_rate": 1.7518629087748125e-05,
1249
+ "loss": 0.6147,
1250
+ "mean_token_accuracy": 0.8109205877780914,
1251
+ "num_tokens": 20815967.0,
1252
+ "step": 6900
1253
+ },
1254
+ {
1255
+ "epoch": 0.23214255891243715,
1256
+ "grad_norm": 3.642578125,
1257
+ "learning_rate": 1.7483874158381137e-05,
1258
+ "loss": 0.6093,
1259
+ "mean_token_accuracy": 0.812092969417572,
1260
+ "num_tokens": 20966880.0,
1261
+ "step": 6950
1262
+ },
1263
+ {
1264
+ "epoch": 0.2338126492643252,
1265
+ "grad_norm": 3.2578125,
1266
+ "learning_rate": 1.7448912514237976e-05,
1267
+ "loss": 0.5986,
1268
+ "mean_token_accuracy": 0.8138788425922394,
1269
+ "num_tokens": 21117408.0,
1270
+ "step": 7000
1271
+ },
1272
+ {
1273
+ "epoch": 0.23548273961621324,
1274
+ "grad_norm": 3.28125,
1275
+ "learning_rate": 1.7413745121006777e-05,
1276
+ "loss": 0.6143,
1277
+ "mean_token_accuracy": 0.8106629079580308,
1278
+ "num_tokens": 21258290.0,
1279
+ "step": 7050
1280
+ },
1281
+ {
1282
+ "epoch": 0.23715282996810128,
1283
+ "grad_norm": 3.671875,
1284
+ "learning_rate": 1.737837295005874e-05,
1285
+ "loss": 0.6026,
1286
+ "mean_token_accuracy": 0.8134672123193741,
1287
+ "num_tokens": 21413317.0,
1288
+ "step": 7100
1289
+ },
1290
+ {
1291
+ "epoch": 0.2388229203199893,
1292
+ "grad_norm": 3.42578125,
1293
+ "learning_rate": 1.73427969784213e-05,
1294
+ "loss": 0.5989,
1295
+ "mean_token_accuracy": 0.81451107442379,
1296
+ "num_tokens": 21559517.0,
1297
+ "step": 7150
1298
+ },
1299
+ {
1300
+ "epoch": 0.24049301067187734,
1301
+ "grad_norm": 3.2890625,
1302
+ "learning_rate": 1.7307018188751153e-05,
1303
+ "loss": 0.6134,
1304
+ "mean_token_accuracy": 0.8111221677064896,
1305
+ "num_tokens": 21711355.0,
1306
+ "step": 7200
1307
+ },
1308
+ {
1309
+ "epoch": 0.2421631010237654,
1310
+ "grad_norm": 3.314453125,
1311
+ "learning_rate": 1.72710375693071e-05,
1312
+ "loss": 0.5866,
1313
+ "mean_token_accuracy": 0.8179266756772995,
1314
+ "num_tokens": 21863362.0,
1315
+ "step": 7250
1316
+ },
1317
+ {
1318
+ "epoch": 0.24383319137565343,
1319
+ "grad_norm": 3.07421875,
1320
+ "learning_rate": 1.723485611392275e-05,
1321
+ "loss": 0.5775,
1322
+ "mean_token_accuracy": 0.8185840421915054,
1323
+ "num_tokens": 22015880.0,
1324
+ "step": 7300
1325
+ },
1326
+ {
1327
+ "epoch": 0.24550328172754146,
1328
+ "grad_norm": 3.33203125,
1329
+ "learning_rate": 1.7198474821979075e-05,
1330
+ "loss": 0.6035,
1331
+ "mean_token_accuracy": 0.814561493396759,
1332
+ "num_tokens": 22164719.0,
1333
+ "step": 7350
1334
+ },
1335
+ {
1336
+ "epoch": 0.2471733720794295,
1337
+ "grad_norm": 3.439453125,
1338
+ "learning_rate": 1.7161894698376814e-05,
1339
+ "loss": 0.5944,
1340
+ "mean_token_accuracy": 0.8162802219390869,
1341
+ "num_tokens": 22316748.0,
1342
+ "step": 7400
1343
+ },
1344
+ {
1345
+ "epoch": 0.24884346243131752,
1346
+ "grad_norm": 3.787109375,
1347
+ "learning_rate": 1.7125116753508696e-05,
1348
+ "loss": 0.5999,
1349
+ "mean_token_accuracy": 0.8156676411628723,
1350
+ "num_tokens": 22466344.0,
1351
+ "step": 7450
1352
+ },
1353
+ {
1354
+ "epoch": 0.2505135527832056,
1355
+ "grad_norm": 3.5625,
1356
+ "learning_rate": 1.708814200323155e-05,
1357
+ "loss": 0.5842,
1358
+ "mean_token_accuracy": 0.8202319222688675,
1359
+ "num_tokens": 22619286.0,
1360
+ "step": 7500
1361
+ },
1362
+ {
1363
+ "epoch": 0.2521836431350936,
1364
+ "grad_norm": 3.37890625,
1365
+ "learning_rate": 1.705097146883823e-05,
1366
+ "loss": 0.6108,
1367
+ "mean_token_accuracy": 0.8124721819162368,
1368
+ "num_tokens": 22769773.0,
1369
+ "step": 7550
1370
+ },
1371
+ {
1372
+ "epoch": 0.25385373348698165,
1373
+ "grad_norm": 3.431640625,
1374
+ "learning_rate": 1.7013606177029417e-05,
1375
+ "loss": 0.6097,
1376
+ "mean_token_accuracy": 0.812991201877594,
1377
+ "num_tokens": 22923130.0,
1378
+ "step": 7600
1379
+ },
1380
+ {
1381
+ "epoch": 0.2555238238388697,
1382
+ "grad_norm": 3.142578125,
1383
+ "learning_rate": 1.6976047159885258e-05,
1384
+ "loss": 0.5725,
1385
+ "mean_token_accuracy": 0.8219118654727936,
1386
+ "num_tokens": 23072994.0,
1387
+ "step": 7650
1388
+ },
1389
+ {
1390
+ "epoch": 0.2571939141907577,
1391
+ "grad_norm": 2.869140625,
1392
+ "learning_rate": 1.6938295454836855e-05,
1393
+ "loss": 0.5903,
1394
+ "mean_token_accuracy": 0.8173499464988708,
1395
+ "num_tokens": 23220605.0,
1396
+ "step": 7700
1397
+ },
1398
+ {
1399
+ "epoch": 0.25886400454264574,
1400
+ "grad_norm": 4.078125,
1401
+ "learning_rate": 1.6900352104637616e-05,
1402
+ "loss": 0.6188,
1403
+ "mean_token_accuracy": 0.8102790033817291,
1404
+ "num_tokens": 23370496.0,
1405
+ "step": 7750
1406
+ },
1407
+ {
1408
+ "epoch": 0.2605340948945338,
1409
+ "grad_norm": 3.7421875,
1410
+ "learning_rate": 1.686221815733444e-05,
1411
+ "loss": 0.6076,
1412
+ "mean_token_accuracy": 0.8124845671653748,
1413
+ "num_tokens": 23524575.0,
1414
+ "step": 7800
1415
+ },
1416
+ {
1417
+ "epoch": 0.2622041852464218,
1418
+ "grad_norm": 4.11328125,
1419
+ "learning_rate": 1.682389466623879e-05,
1420
+ "loss": 0.5937,
1421
+ "mean_token_accuracy": 0.8181423616409301,
1422
+ "num_tokens": 23673370.0,
1423
+ "step": 7850
1424
+ },
1425
+ {
1426
+ "epoch": 0.2638742755983099,
1427
+ "grad_norm": 3.591796875,
1428
+ "learning_rate": 1.678538268989757e-05,
1429
+ "loss": 0.5962,
1430
+ "mean_token_accuracy": 0.8151667928695678,
1431
+ "num_tokens": 23824567.0,
1432
+ "step": 7900
1433
+ },
1434
+ {
1435
+ "epoch": 0.2655443659501979,
1436
+ "grad_norm": 3.638671875,
1437
+ "learning_rate": 1.6746683292063913e-05,
1438
+ "loss": 0.595,
1439
+ "mean_token_accuracy": 0.816837123632431,
1440
+ "num_tokens": 23979656.0,
1441
+ "step": 7950
1442
+ },
1443
+ {
1444
+ "epoch": 0.26721445630208596,
1445
+ "grad_norm": 3.806640625,
1446
+ "learning_rate": 1.670779754166779e-05,
1447
+ "loss": 0.571,
1448
+ "mean_token_accuracy": 0.8221717000007629,
1449
+ "num_tokens": 24124997.0,
1450
+ "step": 8000
1451
+ },
1452
+ {
1453
+ "epoch": 0.268884546653974,
1454
+ "grad_norm": 4.38671875,
1455
+ "learning_rate": 1.666872651278648e-05,
1456
+ "loss": 0.5823,
1457
+ "mean_token_accuracy": 0.8184084331989289,
1458
+ "num_tokens": 24274928.0,
1459
+ "step": 8050
1460
+ },
1461
+ {
1462
+ "epoch": 0.270554637005862,
1463
+ "grad_norm": 3.314453125,
1464
+ "learning_rate": 1.66294712846149e-05,
1465
+ "loss": 0.5908,
1466
+ "mean_token_accuracy": 0.8170400822162628,
1467
+ "num_tokens": 24425253.0,
1468
+ "step": 8100
1469
+ },
1470
+ {
1471
+ "epoch": 0.27222472735775005,
1472
+ "grad_norm": 3.328125,
1473
+ "learning_rate": 1.6590032941435812e-05,
1474
+ "loss": 0.587,
1475
+ "mean_token_accuracy": 0.8180436098575592,
1476
+ "num_tokens": 24570789.0,
1477
+ "step": 8150
1478
+ },
1479
+ {
1480
+ "epoch": 0.2738948177096381,
1481
+ "grad_norm": 2.984375,
1482
+ "learning_rate": 1.6550412572589857e-05,
1483
+ "loss": 0.5961,
1484
+ "mean_token_accuracy": 0.8165696883201599,
1485
+ "num_tokens": 24728204.0,
1486
+ "step": 8200
1487
+ },
1488
+ {
1489
+ "epoch": 0.2755649080615261,
1490
+ "grad_norm": 3.09375,
1491
+ "learning_rate": 1.6510611272445472e-05,
1492
+ "loss": 0.5883,
1493
+ "mean_token_accuracy": 0.8188752412796021,
1494
+ "num_tokens": 24884334.0,
1495
+ "step": 8250
1496
+ },
1497
+ {
1498
+ "epoch": 0.27723499841341415,
1499
+ "grad_norm": 3.611328125,
1500
+ "learning_rate": 1.6470630140368663e-05,
1501
+ "loss": 0.591,
1502
+ "mean_token_accuracy": 0.8173883992433548,
1503
+ "num_tokens": 25029094.0,
1504
+ "step": 8300
1505
+ },
1506
+ {
1507
+ "epoch": 0.2789050887653022,
1508
+ "grad_norm": 3.271484375,
1509
+ "learning_rate": 1.6430470280692644e-05,
1510
+ "loss": 0.5848,
1511
+ "mean_token_accuracy": 0.820589507818222,
1512
+ "num_tokens": 25177586.0,
1513
+ "step": 8350
1514
+ },
1515
+ {
1516
+ "epoch": 0.28057517911719027,
1517
+ "grad_norm": 3.2109375,
1518
+ "learning_rate": 1.639013280268732e-05,
1519
+ "loss": 0.5581,
1520
+ "mean_token_accuracy": 0.8250171238183975,
1521
+ "num_tokens": 25332667.0,
1522
+ "step": 8400
1523
+ },
1524
+ {
1525
+ "epoch": 0.2822452694690783,
1526
+ "grad_norm": 3.505859375,
1527
+ "learning_rate": 1.6349618820528662e-05,
1528
+ "loss": 0.5978,
1529
+ "mean_token_accuracy": 0.8156885832548142,
1530
+ "num_tokens": 25482520.0,
1531
+ "step": 8450
1532
+ },
1533
+ {
1534
+ "epoch": 0.28391535982096633,
1535
+ "grad_norm": 3.703125,
1536
+ "learning_rate": 1.6308929453267918e-05,
1537
+ "loss": 0.5633,
1538
+ "mean_token_accuracy": 0.8238223212957382,
1539
+ "num_tokens": 25639672.0,
1540
+ "step": 8500
1541
+ },
1542
+ {
1543
+ "epoch": 0.28558545017285436,
1544
+ "grad_norm": 3.390625,
1545
+ "learning_rate": 1.6268065824800723e-05,
1546
+ "loss": 0.5888,
1547
+ "mean_token_accuracy": 0.8175674903392792,
1548
+ "num_tokens": 25793282.0,
1549
+ "step": 8550
1550
+ },
1551
+ {
1552
+ "epoch": 0.2872555405247424,
1553
+ "grad_norm": 3.59375,
1554
+ "learning_rate": 1.622702906383603e-05,
1555
+ "loss": 0.579,
1556
+ "mean_token_accuracy": 0.8193080085515976,
1557
+ "num_tokens": 25946540.0,
1558
+ "step": 8600
1559
+ },
1560
+ {
1561
+ "epoch": 0.2889256308766304,
1562
+ "grad_norm": 3.0390625,
1563
+ "learning_rate": 1.6185820303864957e-05,
1564
+ "loss": 0.5738,
1565
+ "mean_token_accuracy": 0.8222082054615021,
1566
+ "num_tokens": 26105903.0,
1567
+ "step": 8650
1568
+ },
1569
+ {
1570
+ "epoch": 0.29059572122851846,
1571
+ "grad_norm": 3.185546875,
1572
+ "learning_rate": 1.614444068312946e-05,
1573
+ "loss": 0.595,
1574
+ "mean_token_accuracy": 0.8175247597694397,
1575
+ "num_tokens": 26254429.0,
1576
+ "step": 8700
1577
+ },
1578
+ {
1579
+ "epoch": 0.2922658115804065,
1580
+ "grad_norm": 2.779296875,
1581
+ "learning_rate": 1.610289134459091e-05,
1582
+ "loss": 0.5878,
1583
+ "mean_token_accuracy": 0.818896831870079,
1584
+ "num_tokens": 26410820.0,
1585
+ "step": 8750
1586
+ },
1587
+ {
1588
+ "epoch": 0.2939359019322945,
1589
+ "grad_norm": 3.423828125,
1590
+ "learning_rate": 1.606117343589851e-05,
1591
+ "loss": 0.5953,
1592
+ "mean_token_accuracy": 0.8172416472434998,
1593
+ "num_tokens": 26561505.0,
1594
+ "step": 8800
1595
+ },
1596
+ {
1597
+ "epoch": 0.29560599228418255,
1598
+ "grad_norm": 3.47265625,
1599
+ "learning_rate": 1.60192881093576e-05,
1600
+ "loss": 0.5691,
1601
+ "mean_token_accuracy": 0.8218266826868057,
1602
+ "num_tokens": 26714736.0,
1603
+ "step": 8850
1604
+ },
1605
+ {
1606
+ "epoch": 0.29727608263607064,
1607
+ "grad_norm": 3.501953125,
1608
+ "learning_rate": 1.5977236521897824e-05,
1609
+ "loss": 0.5887,
1610
+ "mean_token_accuracy": 0.8170481222867966,
1611
+ "num_tokens": 26864415.0,
1612
+ "step": 8900
1613
+ },
1614
+ {
1615
+ "epoch": 0.29894617298795867,
1616
+ "grad_norm": 3.005859375,
1617
+ "learning_rate": 1.593501983504119e-05,
1618
+ "loss": 0.556,
1619
+ "mean_token_accuracy": 0.8250821077823639,
1620
+ "num_tokens": 27012496.0,
1621
+ "step": 8950
1622
+ },
1623
+ {
1624
+ "epoch": 0.3006162633398467,
1625
+ "grad_norm": 3.279296875,
1626
+ "learning_rate": 1.589263921486997e-05,
1627
+ "loss": 0.596,
1628
+ "mean_token_accuracy": 0.8154706001281739,
1629
+ "num_tokens": 27165174.0,
1630
+ "step": 9000
1631
+ },
1632
+ {
1633
+ "epoch": 0.30228635369173473,
1634
+ "grad_norm": 4.0078125,
1635
+ "learning_rate": 1.5850095831994486e-05,
1636
+ "loss": 0.5945,
1637
+ "mean_token_accuracy": 0.8172478175163269,
1638
+ "num_tokens": 27322736.0,
1639
+ "step": 9050
1640
+ },
1641
+ {
1642
+ "epoch": 0.30395644404362276,
1643
+ "grad_norm": 3.69921875,
1644
+ "learning_rate": 1.5807390861520805e-05,
1645
+ "loss": 0.5964,
1646
+ "mean_token_accuracy": 0.8172479033470154,
1647
+ "num_tokens": 27470431.0,
1648
+ "step": 9100
1649
+ },
1650
+ {
1651
+ "epoch": 0.3056265343955108,
1652
+ "grad_norm": 3.1796875,
1653
+ "learning_rate": 1.5764525483018254e-05,
1654
+ "loss": 0.5978,
1655
+ "mean_token_accuracy": 0.8172558772563935,
1656
+ "num_tokens": 27616926.0,
1657
+ "step": 9150
1658
+ },
1659
+ {
1660
+ "epoch": 0.3072966247473988,
1661
+ "grad_norm": 3.384765625,
1662
+ "learning_rate": 1.5721500880486847e-05,
1663
+ "loss": 0.5727,
1664
+ "mean_token_accuracy": 0.8231468296051025,
1665
+ "num_tokens": 27766315.0,
1666
+ "step": 9200
1667
+ },
1668
+ {
1669
+ "epoch": 0.30896671509928686,
1670
+ "grad_norm": 3.265625,
1671
+ "learning_rate": 1.5678318242324587e-05,
1672
+ "loss": 0.6048,
1673
+ "mean_token_accuracy": 0.8166863393783569,
1674
+ "num_tokens": 27912309.0,
1675
+ "step": 9250
1676
+ },
1677
+ {
1678
+ "epoch": 0.3106368054511749,
1679
+ "grad_norm": 3.587890625,
1680
+ "learning_rate": 1.5634978761294627e-05,
1681
+ "loss": 0.5783,
1682
+ "mean_token_accuracy": 0.8223829501867295,
1683
+ "num_tokens": 28068124.0,
1684
+ "step": 9300
1685
+ },
1686
+ {
1687
+ "epoch": 0.3123068958030629,
1688
+ "grad_norm": 3.619140625,
1689
+ "learning_rate": 1.559148363449235e-05,
1690
+ "loss": 0.5815,
1691
+ "mean_token_accuracy": 0.8205723863840103,
1692
+ "num_tokens": 28212240.0,
1693
+ "step": 9350
1694
+ },
1695
+ {
1696
+ "epoch": 0.313976986154951,
1697
+ "grad_norm": 3.5078125,
1698
+ "learning_rate": 1.554783406331226e-05,
1699
+ "loss": 0.5708,
1700
+ "mean_token_accuracy": 0.8217070680856705,
1701
+ "num_tokens": 28361807.0,
1702
+ "step": 9400
1703
+ },
1704
+ {
1705
+ "epoch": 0.31564707650683904,
1706
+ "grad_norm": 3.46875,
1707
+ "learning_rate": 1.5504031253414854e-05,
1708
+ "loss": 0.5704,
1709
+ "mean_token_accuracy": 0.8224059015512466,
1710
+ "num_tokens": 28520675.0,
1711
+ "step": 9450
1712
+ },
1713
+ {
1714
+ "epoch": 0.3173171668587271,
1715
+ "grad_norm": 3.56640625,
1716
+ "learning_rate": 1.5460076414693276e-05,
1717
+ "loss": 0.5947,
1718
+ "mean_token_accuracy": 0.8166932672262192,
1719
+ "num_tokens": 28671012.0,
1720
+ "step": 9500
1721
+ },
1722
+ {
1723
+ "epoch": 0.3189872572106151,
1724
+ "grad_norm": 3.40625,
1725
+ "learning_rate": 1.5415970761239908e-05,
1726
+ "loss": 0.5703,
1727
+ "mean_token_accuracy": 0.823592340350151,
1728
+ "num_tokens": 28826413.0,
1729
+ "step": 9550
1730
+ },
1731
+ {
1732
+ "epoch": 0.32065734756250314,
1733
+ "grad_norm": 3.021484375,
1734
+ "learning_rate": 1.5371715511312846e-05,
1735
+ "loss": 0.5524,
1736
+ "mean_token_accuracy": 0.8269542592763901,
1737
+ "num_tokens": 28973595.0,
1738
+ "step": 9600
1739
+ },
1740
+ {
1741
+ "epoch": 0.32232743791439117,
1742
+ "grad_norm": 4.5625,
1743
+ "learning_rate": 1.5327311887302247e-05,
1744
+ "loss": 0.587,
1745
+ "mean_token_accuracy": 0.8197296333312988,
1746
+ "num_tokens": 29118143.0,
1747
+ "step": 9650
1748
+ },
1749
+ {
1750
+ "epoch": 0.3239975282662792,
1751
+ "grad_norm": 4.28125,
1752
+ "learning_rate": 1.528276111569656e-05,
1753
+ "loss": 0.573,
1754
+ "mean_token_accuracy": 0.8212918490171432,
1755
+ "num_tokens": 29271964.0,
1756
+ "step": 9700
1757
+ },
1758
+ {
1759
+ "epoch": 0.32566761861816723,
1760
+ "grad_norm": 3.27734375,
1761
+ "learning_rate": 1.5238064427048644e-05,
1762
+ "loss": 0.5636,
1763
+ "mean_token_accuracy": 0.8254357093572616,
1764
+ "num_tokens": 29423354.0,
1765
+ "step": 9750
1766
+ },
1767
+ {
1768
+ "epoch": 0.32733770897005526,
1769
+ "grad_norm": 3.314453125,
1770
+ "learning_rate": 1.5193223055941793e-05,
1771
+ "loss": 0.5585,
1772
+ "mean_token_accuracy": 0.8258392840623856,
1773
+ "num_tokens": 29573920.0,
1774
+ "step": 9800
1775
+ },
1776
+ {
1777
+ "epoch": 0.3290077993219433,
1778
+ "grad_norm": 3.05078125,
1779
+ "learning_rate": 1.5148238240955628e-05,
1780
+ "loss": 0.5751,
1781
+ "mean_token_accuracy": 0.8223698770999909,
1782
+ "num_tokens": 29723005.0,
1783
+ "step": 9850
1784
+ },
1785
+ {
1786
+ "epoch": 0.3306778896738313,
1787
+ "grad_norm": 3.27734375,
1788
+ "learning_rate": 1.5103111224631875e-05,
1789
+ "loss": 0.5801,
1790
+ "mean_token_accuracy": 0.8200888675451279,
1791
+ "num_tokens": 29869002.0,
1792
+ "step": 9900
1793
+ },
1794
+ {
1795
+ "epoch": 0.3323479800257194,
1796
+ "grad_norm": 3.923828125,
1797
+ "learning_rate": 1.5057843253440063e-05,
1798
+ "loss": 0.5861,
1799
+ "mean_token_accuracy": 0.8186714911460876,
1800
+ "num_tokens": 30015782.0,
1801
+ "step": 9950
1802
+ },
1803
+ {
1804
+ "epoch": 0.33401807037760745,
1805
+ "grad_norm": 3.38671875,
1806
+ "learning_rate": 1.5012435577743085e-05,
1807
+ "loss": 0.5844,
1808
+ "mean_token_accuracy": 0.821356486082077,
1809
+ "num_tokens": 30161096.0,
1810
+ "step": 10000
1811
+ },
1812
+ {
1813
+ "epoch": 0.3356881607294955,
1814
+ "grad_norm": 4.05078125,
1815
+ "learning_rate": 1.4966889451762658e-05,
1816
+ "loss": 0.5633,
1817
+ "mean_token_accuracy": 0.8244505578279495,
1818
+ "num_tokens": 30310339.0,
1819
+ "step": 10050
1820
+ },
1821
+ {
1822
+ "epoch": 0.3373582510813835,
1823
+ "grad_norm": 3.69921875,
1824
+ "learning_rate": 1.492120613354469e-05,
1825
+ "loss": 0.5602,
1826
+ "mean_token_accuracy": 0.8251171362400055,
1827
+ "num_tokens": 30462338.0,
1828
+ "step": 10100
1829
+ },
1830
+ {
1831
+ "epoch": 0.33902834143327154,
1832
+ "grad_norm": 3.177734375,
1833
+ "learning_rate": 1.487538688492452e-05,
1834
+ "loss": 0.5581,
1835
+ "mean_token_accuracy": 0.825383592247963,
1836
+ "num_tokens": 30614667.0,
1837
+ "step": 10150
1838
+ },
1839
+ {
1840
+ "epoch": 0.3406984317851596,
1841
+ "grad_norm": 3.109375,
1842
+ "learning_rate": 1.482943297149207e-05,
1843
+ "loss": 0.5503,
1844
+ "mean_token_accuracy": 0.8295097362995147,
1845
+ "num_tokens": 30763133.0,
1846
+ "step": 10200
1847
+ },
1848
+ {
1849
+ "epoch": 0.3423685221370476,
1850
+ "grad_norm": 3.802734375,
1851
+ "learning_rate": 1.4783345662556894e-05,
1852
+ "loss": 0.573,
1853
+ "mean_token_accuracy": 0.823209963440895,
1854
+ "num_tokens": 30917281.0,
1855
+ "step": 10250
1856
+ },
1857
+ {
1858
+ "epoch": 0.34403861248893564,
1859
+ "grad_norm": 3.8828125,
1860
+ "learning_rate": 1.47371262311131e-05,
1861
+ "loss": 0.5714,
1862
+ "mean_token_accuracy": 0.8230595916509629,
1863
+ "num_tokens": 31063586.0,
1864
+ "step": 10300
1865
+ },
1866
+ {
1867
+ "epoch": 0.34570870284082367,
1868
+ "grad_norm": 3.333984375,
1869
+ "learning_rate": 1.4690775953804206e-05,
1870
+ "loss": 0.5487,
1871
+ "mean_token_accuracy": 0.8284380221366883,
1872
+ "num_tokens": 31219950.0,
1873
+ "step": 10350
1874
+ },
1875
+ {
1876
+ "epoch": 0.3473787931927117,
1877
+ "grad_norm": 3.666015625,
1878
+ "learning_rate": 1.4644296110887869e-05,
1879
+ "loss": 0.5672,
1880
+ "mean_token_accuracy": 0.8228234505653381,
1881
+ "num_tokens": 31370295.0,
1882
+ "step": 10400
1883
+ },
1884
+ {
1885
+ "epoch": 0.3490488835445998,
1886
+ "grad_norm": 2.96875,
1887
+ "learning_rate": 1.4597687986200529e-05,
1888
+ "loss": 0.5744,
1889
+ "mean_token_accuracy": 0.8230278396606445,
1890
+ "num_tokens": 31517590.0,
1891
+ "step": 10450
1892
+ },
1893
+ {
1894
+ "epoch": 0.3507189738964878,
1895
+ "grad_norm": 3.1875,
1896
+ "learning_rate": 1.4550952867121933e-05,
1897
+ "loss": 0.5654,
1898
+ "mean_token_accuracy": 0.824184655547142,
1899
+ "num_tokens": 31668597.0,
1900
+ "step": 10500
1901
+ },
1902
+ {
1903
+ "epoch": 0.35238906424837585,
1904
+ "grad_norm": 3.6953125,
1905
+ "learning_rate": 1.4504092044539597e-05,
1906
+ "loss": 0.5592,
1907
+ "mean_token_accuracy": 0.8253324276208878,
1908
+ "num_tokens": 31817864.0,
1909
+ "step": 10550
1910
+ },
1911
+ {
1912
+ "epoch": 0.3540591546002639,
1913
+ "grad_norm": 3.45703125,
1914
+ "learning_rate": 1.4457106812813127e-05,
1915
+ "loss": 0.5505,
1916
+ "mean_token_accuracy": 0.827171710729599,
1917
+ "num_tokens": 31973663.0,
1918
+ "step": 10600
1919
+ },
1920
+ {
1921
+ "epoch": 0.3557292449521519,
1922
+ "grad_norm": 3.0234375,
1923
+ "learning_rate": 1.4409998469738486e-05,
1924
+ "loss": 0.5603,
1925
+ "mean_token_accuracy": 0.8264797705411911,
1926
+ "num_tokens": 32122867.0,
1927
+ "step": 10650
1928
+ },
1929
+ {
1930
+ "epoch": 0.35739933530403994,
1931
+ "grad_norm": 3.646484375,
1932
+ "learning_rate": 1.4362768316512143e-05,
1933
+ "loss": 0.5526,
1934
+ "mean_token_accuracy": 0.8274721032381058,
1935
+ "num_tokens": 32271390.0,
1936
+ "step": 10700
1937
+ },
1938
+ {
1939
+ "epoch": 0.359069425655928,
1940
+ "grad_norm": 3.0390625,
1941
+ "learning_rate": 1.431541765769512e-05,
1942
+ "loss": 0.5704,
1943
+ "mean_token_accuracy": 0.8233271539211273,
1944
+ "num_tokens": 32422937.0,
1945
+ "step": 10750
1946
+ },
1947
+ {
1948
+ "epoch": 0.360739516007816,
1949
+ "grad_norm": 3.361328125,
1950
+ "learning_rate": 1.4267947801176972e-05,
1951
+ "loss": 0.5592,
1952
+ "mean_token_accuracy": 0.8274497729539871,
1953
+ "num_tokens": 32575242.0,
1954
+ "step": 10800
1955
+ },
1956
+ {
1957
+ "epoch": 0.36240960635970404,
1958
+ "grad_norm": 3.708984375,
1959
+ "learning_rate": 1.4220360058139658e-05,
1960
+ "loss": 0.563,
1961
+ "mean_token_accuracy": 0.8254307335615159,
1962
+ "num_tokens": 32729992.0,
1963
+ "step": 10850
1964
+ },
1965
+ {
1966
+ "epoch": 0.36407969671159207,
1967
+ "grad_norm": 4.1640625,
1968
+ "learning_rate": 1.4172655743021317e-05,
1969
+ "loss": 0.5382,
1970
+ "mean_token_accuracy": 0.8311225980520248,
1971
+ "num_tokens": 32875452.0,
1972
+ "step": 10900
1973
+ },
1974
+ {
1975
+ "epoch": 0.36574978706348016,
1976
+ "grad_norm": 3.388671875,
1977
+ "learning_rate": 1.4124836173479972e-05,
1978
+ "loss": 0.5514,
1979
+ "mean_token_accuracy": 0.8262929075956345,
1980
+ "num_tokens": 33030316.0,
1981
+ "step": 10950
1982
+ },
1983
+ {
1984
+ "epoch": 0.3674198774153682,
1985
+ "grad_norm": 3.482421875,
1986
+ "learning_rate": 1.4076902670357132e-05,
1987
+ "loss": 0.5602,
1988
+ "mean_token_accuracy": 0.8263766020536423,
1989
+ "num_tokens": 33183725.0,
1990
+ "step": 11000
1991
+ },
1992
+ {
1993
+ "epoch": 0.3690899677672562,
1994
+ "grad_norm": 4.14453125,
1995
+ "learning_rate": 1.4028856557641307e-05,
1996
+ "loss": 0.56,
1997
+ "mean_token_accuracy": 0.8278713476657867,
1998
+ "num_tokens": 33338154.0,
1999
+ "step": 11050
2000
+ },
2001
+ {
2002
+ "epoch": 0.37076005811914425,
2003
+ "grad_norm": 3.494140625,
2004
+ "learning_rate": 1.3980699162431421e-05,
2005
+ "loss": 0.5614,
2006
+ "mean_token_accuracy": 0.8253583014011383,
2007
+ "num_tokens": 33490131.0,
2008
+ "step": 11100
2009
+ },
2010
+ {
2011
+ "epoch": 0.3724301484710323,
2012
+ "grad_norm": 3.3828125,
2013
+ "learning_rate": 1.3932431814900195e-05,
2014
+ "loss": 0.5431,
2015
+ "mean_token_accuracy": 0.8307267421483994,
2016
+ "num_tokens": 33644521.0,
2017
+ "step": 11150
2018
+ },
2019
+ {
2020
+ "epoch": 0.3741002388229203,
2021
+ "grad_norm": 3.28125,
2022
+ "learning_rate": 1.3884055848257369e-05,
2023
+ "loss": 0.5521,
2024
+ "mean_token_accuracy": 0.8285236042737961,
2025
+ "num_tokens": 33800815.0,
2026
+ "step": 11200
2027
+ },
2028
+ {
2029
+ "epoch": 0.37577032917480835,
2030
+ "grad_norm": 3.08203125,
2031
+ "learning_rate": 1.383557259871289e-05,
2032
+ "loss": 0.5586,
2033
+ "mean_token_accuracy": 0.8272363543510437,
2034
+ "num_tokens": 33946173.0,
2035
+ "step": 11250
2036
+ },
2037
+ {
2038
+ "epoch": 0.3774404195266964,
2039
+ "grad_norm": 4.18359375,
2040
+ "learning_rate": 1.3786983405440013e-05,
2041
+ "loss": 0.552,
2042
+ "mean_token_accuracy": 0.8265141195058823,
2043
+ "num_tokens": 34095920.0,
2044
+ "step": 11300
2045
+ },
2046
+ {
2047
+ "epoch": 0.3791105098785844,
2048
+ "grad_norm": 3.103515625,
2049
+ "learning_rate": 1.373828961053829e-05,
2050
+ "loss": 0.5837,
2051
+ "mean_token_accuracy": 0.8216396170854569,
2052
+ "num_tokens": 34248446.0,
2053
+ "step": 11350
2054
+ },
2055
+ {
2056
+ "epoch": 0.38078060023047244,
2057
+ "grad_norm": 3.419921875,
2058
+ "learning_rate": 1.3689492558996529e-05,
2059
+ "loss": 0.5463,
2060
+ "mean_token_accuracy": 0.8293001532554627,
2061
+ "num_tokens": 34400906.0,
2062
+ "step": 11400
2063
+ },
2064
+ {
2065
+ "epoch": 0.38245069058236053,
2066
+ "grad_norm": 3.173828125,
2067
+ "learning_rate": 1.3640593598655611e-05,
2068
+ "loss": 0.5463,
2069
+ "mean_token_accuracy": 0.8295714586973191,
2070
+ "num_tokens": 34553639.0,
2071
+ "step": 11450
2072
+ },
2073
+ {
2074
+ "epoch": 0.38412078093424856,
2075
+ "grad_norm": 3.44140625,
2076
+ "learning_rate": 1.3591594080171286e-05,
2077
+ "loss": 0.5626,
2078
+ "mean_token_accuracy": 0.8255741602182388,
2079
+ "num_tokens": 34705512.0,
2080
+ "step": 11500
2081
+ },
2082
+ {
2083
+ "epoch": 0.3857908712861366,
2084
+ "grad_norm": 2.962890625,
2085
+ "learning_rate": 1.3542495356976851e-05,
2086
+ "loss": 0.5529,
2087
+ "mean_token_accuracy": 0.8291010856628418,
2088
+ "num_tokens": 34851485.0,
2089
+ "step": 11550
2090
+ },
2091
+ {
2092
+ "epoch": 0.3874609616380246,
2093
+ "grad_norm": 3.091796875,
2094
+ "learning_rate": 1.3493298785245778e-05,
2095
+ "loss": 0.5588,
2096
+ "mean_token_accuracy": 0.826916481256485,
2097
+ "num_tokens": 35008468.0,
2098
+ "step": 11600
2099
+ },
2100
+ {
2101
+ "epoch": 0.38913105198991266,
2102
+ "grad_norm": 3.2109375,
2103
+ "learning_rate": 1.344400572385424e-05,
2104
+ "loss": 0.5458,
2105
+ "mean_token_accuracy": 0.8301201510429382,
2106
+ "num_tokens": 35157833.0,
2107
+ "step": 11650
2108
+ },
2109
+ {
2110
+ "epoch": 0.3908011423418007,
2111
+ "grad_norm": 4.0859375,
2112
+ "learning_rate": 1.3394617534343596e-05,
2113
+ "loss": 0.5461,
2114
+ "mean_token_accuracy": 0.8286441028118133,
2115
+ "num_tokens": 35310802.0,
2116
+ "step": 11700
2117
+ },
2118
+ {
2119
+ "epoch": 0.3924712326936887,
2120
+ "grad_norm": 3.578125,
2121
+ "learning_rate": 1.3345135580882764e-05,
2122
+ "loss": 0.553,
2123
+ "mean_token_accuracy": 0.8291445529460907,
2124
+ "num_tokens": 35456677.0,
2125
+ "step": 11750
2126
+ },
2127
+ {
2128
+ "epoch": 0.39414132304557675,
2129
+ "grad_norm": 4.30078125,
2130
+ "learning_rate": 1.3295561230230555e-05,
2131
+ "loss": 0.557,
2132
+ "mean_token_accuracy": 0.827009784579277,
2133
+ "num_tokens": 35604720.0,
2134
+ "step": 11800
2135
+ },
2136
+ {
2137
+ "epoch": 0.3958114133974648,
2138
+ "grad_norm": 3.46875,
2139
+ "learning_rate": 1.3245895851697914e-05,
2140
+ "loss": 0.5387,
2141
+ "mean_token_accuracy": 0.832523928284645,
2142
+ "num_tokens": 35758780.0,
2143
+ "step": 11850
2144
+ },
2145
+ {
2146
+ "epoch": 0.3974815037493528,
2147
+ "grad_norm": 3.318359375,
2148
+ "learning_rate": 1.3196140817110104e-05,
2149
+ "loss": 0.5501,
2150
+ "mean_token_accuracy": 0.8280089855194092,
2151
+ "num_tokens": 35908459.0,
2152
+ "step": 11900
2153
+ },
2154
+ {
2155
+ "epoch": 0.3991515941012409,
2156
+ "grad_norm": 3.591796875,
2157
+ "learning_rate": 1.3146297500768799e-05,
2158
+ "loss": 0.5518,
2159
+ "mean_token_accuracy": 0.8280233424901963,
2160
+ "num_tokens": 36054122.0,
2161
+ "step": 11950
2162
+ },
2163
+ {
2164
+ "epoch": 0.40082168445312893,
2165
+ "grad_norm": 3.591796875,
2166
+ "learning_rate": 1.3096367279414142e-05,
2167
+ "loss": 0.557,
2168
+ "mean_token_accuracy": 0.8281977719068527,
2169
+ "num_tokens": 36206481.0,
2170
+ "step": 12000
2171
+ },
2172
+ {
2173
+ "epoch": 0.40249177480501697,
2174
+ "grad_norm": 3.64453125,
2175
+ "learning_rate": 1.3046351532186718e-05,
2176
+ "loss": 0.5341,
2177
+ "mean_token_accuracy": 0.8324338626861573,
2178
+ "num_tokens": 36357066.0,
2179
+ "step": 12050
2180
+ },
2181
+ {
2182
+ "epoch": 0.404161865156905,
2183
+ "grad_norm": 4.046875,
2184
+ "learning_rate": 1.2996251640589443e-05,
2185
+ "loss": 0.5597,
2186
+ "mean_token_accuracy": 0.8256447410583496,
2187
+ "num_tokens": 36509298.0,
2188
+ "step": 12100
2189
+ },
2190
+ {
2191
+ "epoch": 0.40583195550879303,
2192
+ "grad_norm": 3.501953125,
2193
+ "learning_rate": 1.2946068988449416e-05,
2194
+ "loss": 0.5461,
2195
+ "mean_token_accuracy": 0.8304660183191299,
2196
+ "num_tokens": 36664284.0,
2197
+ "step": 12150
2198
+ },
2199
+ {
2200
+ "epoch": 0.40750204586068106,
2201
+ "grad_norm": 3.775390625,
2202
+ "learning_rate": 1.2895804961879697e-05,
2203
+ "loss": 0.5452,
2204
+ "mean_token_accuracy": 0.8312527567148209,
2205
+ "num_tokens": 36815575.0,
2206
+ "step": 12200
2207
+ },
2208
+ {
2209
+ "epoch": 0.4091721362125691,
2210
+ "grad_norm": 2.853515625,
2211
+ "learning_rate": 1.2845460949241018e-05,
2212
+ "loss": 0.5481,
2213
+ "mean_token_accuracy": 0.8277529460191727,
2214
+ "num_tokens": 36974534.0,
2215
+ "step": 12250
2216
+ },
2217
+ {
2218
+ "epoch": 0.4108422265644571,
2219
+ "grad_norm": 3.171875,
2220
+ "learning_rate": 1.2795038341103432e-05,
2221
+ "loss": 0.5543,
2222
+ "mean_token_accuracy": 0.8282213443517685,
2223
+ "num_tokens": 37125168.0,
2224
+ "step": 12300
2225
+ },
2226
+ {
2227
+ "epoch": 0.41251231691634516,
2228
+ "grad_norm": 3.2734375,
2229
+ "learning_rate": 1.2744538530207916e-05,
2230
+ "loss": 0.547,
2231
+ "mean_token_accuracy": 0.8285178303718567,
2232
+ "num_tokens": 37278364.0,
2233
+ "step": 12350
2234
+ },
2235
+ {
2236
+ "epoch": 0.4141824072682332,
2237
+ "grad_norm": 2.806640625,
2238
+ "learning_rate": 1.2693962911427877e-05,
2239
+ "loss": 0.568,
2240
+ "mean_token_accuracy": 0.8246202218532562,
2241
+ "num_tokens": 37431928.0,
2242
+ "step": 12400
2243
+ },
2244
+ {
2245
+ "epoch": 0.4158524976201213,
2246
+ "grad_norm": 2.970703125,
2247
+ "learning_rate": 1.264331288173065e-05,
2248
+ "loss": 0.5553,
2249
+ "mean_token_accuracy": 0.8278548353910447,
2250
+ "num_tokens": 37584284.0,
2251
+ "step": 12450
2252
+ },
2253
+ {
2254
+ "epoch": 0.4175225879720093,
2255
+ "grad_norm": 3.25390625,
2256
+ "learning_rate": 1.2592589840138897e-05,
2257
+ "loss": 0.5417,
2258
+ "mean_token_accuracy": 0.8309043395519257,
2259
+ "num_tokens": 37729858.0,
2260
+ "step": 12500
2261
+ },
2262
+ {
2263
+ "epoch": 0.41919267832389734,
2264
+ "grad_norm": 3.705078125,
2265
+ "learning_rate": 1.2541795187691971e-05,
2266
+ "loss": 0.5514,
2267
+ "mean_token_accuracy": 0.8308093577623368,
2268
+ "num_tokens": 37876100.0,
2269
+ "step": 12550
2270
+ },
2271
+ {
2272
+ "epoch": 0.42086276867578537,
2273
+ "grad_norm": 3.5,
2274
+ "learning_rate": 1.2490930327407209e-05,
2275
+ "loss": 0.5575,
2276
+ "mean_token_accuracy": 0.8272964191436768,
2277
+ "num_tokens": 38025185.0,
2278
+ "step": 12600
2279
+ },
2280
+ {
2281
+ "epoch": 0.4225328590276734,
2282
+ "grad_norm": 3.21484375,
2283
+ "learning_rate": 1.243999666424119e-05,
2284
+ "loss": 0.5485,
2285
+ "mean_token_accuracy": 0.8294407540559768,
2286
+ "num_tokens": 38181591.0,
2287
+ "step": 12650
2288
+ },
2289
+ {
2290
+ "epoch": 0.42420294937956143,
2291
+ "grad_norm": 2.900390625,
2292
+ "learning_rate": 1.2388995605050914e-05,
2293
+ "loss": 0.5449,
2294
+ "mean_token_accuracy": 0.8301975548267364,
2295
+ "num_tokens": 38327561.0,
2296
+ "step": 12700
2297
+ },
2298
+ {
2299
+ "epoch": 0.42587303973144947,
2300
+ "grad_norm": 3.689453125,
2301
+ "learning_rate": 1.2337928558554956e-05,
2302
+ "loss": 0.5332,
2303
+ "mean_token_accuracy": 0.8314021265506745,
2304
+ "num_tokens": 38473244.0,
2305
+ "step": 12750
2306
+ },
2307
+ {
2308
+ "epoch": 0.4275431300833375,
2309
+ "grad_norm": 3.4921875,
2310
+ "learning_rate": 1.2286796935294551e-05,
2311
+ "loss": 0.5471,
2312
+ "mean_token_accuracy": 0.8304495525360107,
2313
+ "num_tokens": 38624223.0,
2314
+ "step": 12800
2315
+ },
2316
+ {
2317
+ "epoch": 0.42921322043522553,
2318
+ "grad_norm": 3.779296875,
2319
+ "learning_rate": 1.2235602147594631e-05,
2320
+ "loss": 0.5479,
2321
+ "mean_token_accuracy": 0.8294032102823258,
2322
+ "num_tokens": 38773521.0,
2323
+ "step": 12850
2324
+ },
2325
+ {
2326
+ "epoch": 0.43088331078711356,
2327
+ "grad_norm": 3.548828125,
2328
+ "learning_rate": 1.2184345609524813e-05,
2329
+ "loss": 0.5482,
2330
+ "mean_token_accuracy": 0.8288749468326568,
2331
+ "num_tokens": 38926115.0,
2332
+ "step": 12900
2333
+ },
2334
+ {
2335
+ "epoch": 0.43255340113900165,
2336
+ "grad_norm": 2.939453125,
2337
+ "learning_rate": 1.2133028736860348e-05,
2338
+ "loss": 0.5353,
2339
+ "mean_token_accuracy": 0.8340490102767945,
2340
+ "num_tokens": 39077937.0,
2341
+ "step": 12950
2342
+ },
2343
+ {
2344
+ "epoch": 0.4342234914908897,
2345
+ "grad_norm": 3.244140625,
2346
+ "learning_rate": 1.2081652947043003e-05,
2347
+ "loss": 0.5501,
2348
+ "mean_token_accuracy": 0.8298387557268143,
2349
+ "num_tokens": 39229394.0,
2350
+ "step": 13000
2351
+ },
2352
+ {
2353
+ "epoch": 0.4358935818427777,
2354
+ "grad_norm": 2.775390625,
2355
+ "learning_rate": 1.2030219659141928e-05,
2356
+ "loss": 0.5353,
2357
+ "mean_token_accuracy": 0.8325220793485641,
2358
+ "num_tokens": 39378856.0,
2359
+ "step": 13050
2360
+ },
2361
+ {
2362
+ "epoch": 0.43756367219466574,
2363
+ "grad_norm": 3.701171875,
2364
+ "learning_rate": 1.1978730293814438e-05,
2365
+ "loss": 0.5334,
2366
+ "mean_token_accuracy": 0.8330723947286606,
2367
+ "num_tokens": 39531138.0,
2368
+ "step": 13100
2369
+ },
2370
+ {
2371
+ "epoch": 0.4392337625465538,
2372
+ "grad_norm": 3.646484375,
2373
+ "learning_rate": 1.1927186273266784e-05,
2374
+ "loss": 0.5345,
2375
+ "mean_token_accuracy": 0.8333028835058213,
2376
+ "num_tokens": 39677471.0,
2377
+ "step": 13150
2378
+ },
2379
+ {
2380
+ "epoch": 0.4409038528984418,
2381
+ "grad_norm": 3.234375,
2382
+ "learning_rate": 1.1875589021214877e-05,
2383
+ "loss": 0.5436,
2384
+ "mean_token_accuracy": 0.8311423152685166,
2385
+ "num_tokens": 39825562.0,
2386
+ "step": 13200
2387
+ },
2388
+ {
2389
+ "epoch": 0.44257394325032984,
2390
+ "grad_norm": 3.224609375,
2391
+ "learning_rate": 1.1823939962844944e-05,
2392
+ "loss": 0.5326,
2393
+ "mean_token_accuracy": 0.8338733261823654,
2394
+ "num_tokens": 39979460.0,
2395
+ "step": 13250
2396
+ },
2397
+ {
2398
+ "epoch": 0.44424403360221787,
2399
+ "grad_norm": 3.4765625,
2400
+ "learning_rate": 1.1772240524774179e-05,
2401
+ "loss": 0.5395,
2402
+ "mean_token_accuracy": 0.8336213880777359,
2403
+ "num_tokens": 40121774.0,
2404
+ "step": 13300
2405
+ },
2406
+ {
2407
+ "epoch": 0.4459141239541059,
2408
+ "grad_norm": 3.6875,
2409
+ "learning_rate": 1.172049213501133e-05,
2410
+ "loss": 0.5434,
2411
+ "mean_token_accuracy": 0.8318886595964432,
2412
+ "num_tokens": 40272793.0,
2413
+ "step": 13350
2414
+ },
2415
+ {
2416
+ "epoch": 0.44758421430599393,
2417
+ "grad_norm": 3.65234375,
2418
+ "learning_rate": 1.166869622291726e-05,
2419
+ "loss": 0.5425,
2420
+ "mean_token_accuracy": 0.8311206418275833,
2421
+ "num_tokens": 40425954.0,
2422
+ "step": 13400
2423
+ },
2424
+ {
2425
+ "epoch": 0.44925430465788196,
2426
+ "grad_norm": 3.109375,
2427
+ "learning_rate": 1.1616854219165452e-05,
2428
+ "loss": 0.5352,
2429
+ "mean_token_accuracy": 0.8340102022886277,
2430
+ "num_tokens": 40578555.0,
2431
+ "step": 13450
2432
+ },
2433
+ {
2434
+ "epoch": 0.45092439500977005,
2435
+ "grad_norm": 3.138671875,
2436
+ "learning_rate": 1.1564967555702526e-05,
2437
+ "loss": 0.5593,
2438
+ "mean_token_accuracy": 0.8278708016872406,
2439
+ "num_tokens": 40725781.0,
2440
+ "step": 13500
2441
+ },
2442
+ {
2443
+ "epoch": 0.4525944853616581,
2444
+ "grad_norm": 3.943359375,
2445
+ "learning_rate": 1.1513037665708639e-05,
2446
+ "loss": 0.5105,
2447
+ "mean_token_accuracy": 0.8397551566362381,
2448
+ "num_tokens": 40874427.0,
2449
+ "step": 13550
2450
+ },
2451
+ {
2452
+ "epoch": 0.4542645757135461,
2453
+ "grad_norm": 3.845703125,
2454
+ "learning_rate": 1.146106598355794e-05,
2455
+ "loss": 0.5195,
2456
+ "mean_token_accuracy": 0.8360963726043701,
2457
+ "num_tokens": 41027698.0,
2458
+ "step": 13600
2459
+ },
2460
+ {
2461
+ "epoch": 0.45593466606543415,
2462
+ "grad_norm": 4.0546875,
2463
+ "learning_rate": 1.1409053944778925e-05,
2464
+ "loss": 0.5367,
2465
+ "mean_token_accuracy": 0.8321689450740815,
2466
+ "num_tokens": 41177523.0,
2467
+ "step": 13650
2468
+ },
2469
+ {
2470
+ "epoch": 0.4576047564173222,
2471
+ "grad_norm": 3.27734375,
2472
+ "learning_rate": 1.1357002986014803e-05,
2473
+ "loss": 0.5253,
2474
+ "mean_token_accuracy": 0.8352927279472351,
2475
+ "num_tokens": 41329674.0,
2476
+ "step": 13700
2477
+ },
2478
+ {
2479
+ "epoch": 0.4592748467692102,
2480
+ "grad_norm": 3.017578125,
2481
+ "learning_rate": 1.1304914544983794e-05,
2482
+ "loss": 0.5232,
2483
+ "mean_token_accuracy": 0.835522665977478,
2484
+ "num_tokens": 41476485.0,
2485
+ "step": 13750
2486
+ },
2487
+ {
2488
+ "epoch": 0.46094493712109824,
2489
+ "grad_norm": 3.44921875,
2490
+ "learning_rate": 1.125279006043945e-05,
2491
+ "loss": 0.5674,
2492
+ "mean_token_accuracy": 0.824871341586113,
2493
+ "num_tokens": 41628555.0,
2494
+ "step": 13800
2495
+ },
2496
+ {
2497
+ "epoch": 0.4626150274729863,
2498
+ "grad_norm": 3.52734375,
2499
+ "learning_rate": 1.1200630972130867e-05,
2500
+ "loss": 0.5282,
2501
+ "mean_token_accuracy": 0.8343129450082779,
2502
+ "num_tokens": 41775801.0,
2503
+ "step": 13850
2504
+ },
2505
+ {
2506
+ "epoch": 0.4642851178248743,
2507
+ "grad_norm": 2.576171875,
2508
+ "learning_rate": 1.1148438720762961e-05,
2509
+ "loss": 0.5106,
2510
+ "mean_token_accuracy": 0.8398624455928803,
2511
+ "num_tokens": 41925501.0,
2512
+ "step": 13900
2513
+ },
2514
+ {
2515
+ "epoch": 0.46595520817676234,
2516
+ "grad_norm": 4.1015625,
2517
+ "learning_rate": 1.109621474795666e-05,
2518
+ "loss": 0.5338,
2519
+ "mean_token_accuracy": 0.8335544627904892,
2520
+ "num_tokens": 42079129.0,
2521
+ "step": 13950
2522
+ },
2523
+ {
2524
+ "epoch": 0.4676252985286504,
2525
+ "grad_norm": 3.314453125,
2526
+ "learning_rate": 1.104396049620907e-05,
2527
+ "loss": 0.5235,
2528
+ "mean_token_accuracy": 0.8374920153617859,
2529
+ "num_tokens": 42225486.0,
2530
+ "step": 14000
2531
+ },
2532
+ {
2533
+ "epoch": 0.46929538888053846,
2534
+ "grad_norm": 3.83203125,
2535
+ "learning_rate": 1.0991677408853643e-05,
2536
+ "loss": 0.5387,
2537
+ "mean_token_accuracy": 0.8317029625177383,
2538
+ "num_tokens": 42379482.0,
2539
+ "step": 14050
2540
+ },
2541
+ {
2542
+ "epoch": 0.4709654792324265,
2543
+ "grad_norm": 4.4296875,
2544
+ "learning_rate": 1.093936693002032e-05,
2545
+ "loss": 0.5135,
2546
+ "mean_token_accuracy": 0.8377016627788544,
2547
+ "num_tokens": 42531797.0,
2548
+ "step": 14100
2549
+ },
2550
+ {
2551
+ "epoch": 0.4726355695843145,
2552
+ "grad_norm": 2.8984375,
2553
+ "learning_rate": 1.0887030504595624e-05,
2554
+ "loss": 0.5392,
2555
+ "mean_token_accuracy": 0.8313734871149063,
2556
+ "num_tokens": 42686643.0,
2557
+ "step": 14150
2558
+ },
2559
+ {
2560
+ "epoch": 0.47430565993620255,
2561
+ "grad_norm": 3.212890625,
2562
+ "learning_rate": 1.0834669578182757e-05,
2563
+ "loss": 0.5475,
2564
+ "mean_token_accuracy": 0.8304911911487579,
2565
+ "num_tokens": 42836786.0,
2566
+ "step": 14200
2567
+ },
2568
+ {
2569
+ "epoch": 0.4759757502880906,
2570
+ "grad_norm": 3.361328125,
2571
+ "learning_rate": 1.0782285597061679e-05,
2572
+ "loss": 0.5259,
2573
+ "mean_token_accuracy": 0.8347867679595947,
2574
+ "num_tokens": 42982714.0,
2575
+ "step": 14250
2576
+ },
2577
+ {
2578
+ "epoch": 0.4776458406399786,
2579
+ "grad_norm": 4.04296875,
2580
+ "learning_rate": 1.0729880008149142e-05,
2581
+ "loss": 0.5485,
2582
+ "mean_token_accuracy": 0.829542515873909,
2583
+ "num_tokens": 43142510.0,
2584
+ "step": 14300
2585
+ },
2586
+ {
2587
+ "epoch": 0.47931593099186665,
2588
+ "grad_norm": 3.369140625,
2589
+ "learning_rate": 1.0677454258958746e-05,
2590
+ "loss": 0.5244,
2591
+ "mean_token_accuracy": 0.8368889981508255,
2592
+ "num_tokens": 43288399.0,
2593
+ "step": 14350
2594
+ },
2595
+ {
2596
+ "epoch": 0.4809860213437547,
2597
+ "grad_norm": 3.76171875,
2598
+ "learning_rate": 1.0625009797560936e-05,
2599
+ "loss": 0.5142,
2600
+ "mean_token_accuracy": 0.8388600462675094,
2601
+ "num_tokens": 43444562.0,
2602
+ "step": 14400
2603
+ },
2604
+ {
2605
+ "epoch": 0.4826561116956427,
2606
+ "grad_norm": 2.94140625,
2607
+ "learning_rate": 1.0572548072543021e-05,
2608
+ "loss": 0.5284,
2609
+ "mean_token_accuracy": 0.8348941618204117,
2610
+ "num_tokens": 43591116.0,
2611
+ "step": 14450
2612
+ },
2613
+ {
2614
+ "epoch": 0.4843262020475308,
2615
+ "grad_norm": 3.28125,
2616
+ "learning_rate": 1.0520070532969142e-05,
2617
+ "loss": 0.5208,
2618
+ "mean_token_accuracy": 0.8390517139434814,
2619
+ "num_tokens": 43740326.0,
2620
+ "step": 14500
2621
+ },
2622
+ {
2623
+ "epoch": 0.4859962923994188,
2624
+ "grad_norm": 3.853515625,
2625
+ "learning_rate": 1.0467578628340278e-05,
2626
+ "loss": 0.5075,
2627
+ "mean_token_accuracy": 0.8397658413648605,
2628
+ "num_tokens": 43891895.0,
2629
+ "step": 14550
2630
+ },
2631
+ {
2632
+ "epoch": 0.48766638275130686,
2633
+ "grad_norm": 3.06640625,
2634
+ "learning_rate": 1.0415073808554178e-05,
2635
+ "loss": 0.5151,
2636
+ "mean_token_accuracy": 0.8379220014810562,
2637
+ "num_tokens": 44046860.0,
2638
+ "step": 14600
2639
+ },
2640
+ {
2641
+ "epoch": 0.4893364731031949,
2642
+ "grad_norm": 3.13671875,
2643
+ "learning_rate": 1.0362557523865323e-05,
2644
+ "loss": 0.5281,
2645
+ "mean_token_accuracy": 0.8356298345327378,
2646
+ "num_tokens": 44203029.0,
2647
+ "step": 14650
2648
+ },
2649
+ {
2650
+ "epoch": 0.4910065634550829,
2651
+ "grad_norm": 3.08984375,
2652
+ "learning_rate": 1.0310031224844878e-05,
2653
+ "loss": 0.5365,
2654
+ "mean_token_accuracy": 0.8323665148019791,
2655
+ "num_tokens": 44355830.0,
2656
+ "step": 14700
2657
+ },
2658
+ {
2659
+ "epoch": 0.49267665380697095,
2660
+ "grad_norm": 3.234375,
2661
+ "learning_rate": 1.0257496362340622e-05,
2662
+ "loss": 0.5168,
2663
+ "mean_token_accuracy": 0.8371515053510666,
2664
+ "num_tokens": 44506850.0,
2665
+ "step": 14750
2666
+ },
2667
+ {
2668
+ "epoch": 0.494346744158859,
2669
+ "grad_norm": 3.34765625,
2670
+ "learning_rate": 1.0204954387436847e-05,
2671
+ "loss": 0.5228,
2672
+ "mean_token_accuracy": 0.8363370823860169,
2673
+ "num_tokens": 44658966.0,
2674
+ "step": 14800
2675
+ },
2676
+ {
2677
+ "epoch": 0.496016834510747,
2678
+ "grad_norm": 3.560546875,
2679
+ "learning_rate": 1.0152406751414328e-05,
2680
+ "loss": 0.5151,
2681
+ "mean_token_accuracy": 0.8392669379711151,
2682
+ "num_tokens": 44805368.0,
2683
+ "step": 14850
2684
+ },
2685
+ {
2686
+ "epoch": 0.49768692486263505,
2687
+ "grad_norm": 3.609375,
2688
+ "learning_rate": 1.0099854905710183e-05,
2689
+ "loss": 0.5285,
2690
+ "mean_token_accuracy": 0.8332407021522522,
2691
+ "num_tokens": 44951969.0,
2692
+ "step": 14900
2693
+ },
2694
+ {
2695
+ "epoch": 0.4993570152145231,
2696
+ "grad_norm": 3.0078125,
2697
+ "learning_rate": 1.0047300301877822e-05,
2698
+ "loss": 0.5289,
2699
+ "mean_token_accuracy": 0.8356048625707626,
2700
+ "num_tokens": 45108518.0,
2701
+ "step": 14950
2702
+ },
2703
+ {
2704
+ "epoch": 0.5010271055664112,
2705
+ "grad_norm": 3.59765625,
2706
+ "learning_rate": 9.994744391546838e-06,
2707
+ "loss": 0.5206,
2708
+ "mean_token_accuracy": 0.8376101624965667,
2709
+ "num_tokens": 45263447.0,
2710
+ "step": 15000
2711
+ }
2712
+ ],
2713
+ "logging_steps": 50,
2714
+ "max_steps": 29938,
2715
+ "num_input_tokens_seen": 0,
2716
+ "num_train_epochs": 1,
2717
+ "save_steps": 1000,
2718
+ "stateful_callbacks": {
2719
+ "TrainerControl": {
2720
+ "args": {
2721
+ "should_epoch_stop": false,
2722
+ "should_evaluate": false,
2723
+ "should_log": false,
2724
+ "should_save": true,
2725
+ "should_training_stop": false
2726
+ },
2727
+ "attributes": {}
2728
+ }
2729
+ },
2730
+ "total_flos": 1.6088218656099533e+17,
2731
+ "train_batch_size": 4,
2732
+ "trial_name": null,
2733
+ "trial_params": null
2734
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:657cbef08b1c3ef63f4593deeb2ba73805f587a3794c3091761b43f1d736d144
3
+ size 5624
vocab.json ADDED
The diff for this file is too large to render. See raw diff