HAissa commited on
Commit
7301ccd
·
verified ·
1 Parent(s): 5176f02

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
added_tokens.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</think>": 151668,
3
+ "</tool_call>": 151658,
4
+ "</tool_response>": 151666,
5
+ "<think>": 151667,
6
+ "<tool_call>": 151657,
7
+ "<tool_response>": 151665,
8
+ "<|box_end|>": 151649,
9
+ "<|box_start|>": 151648,
10
+ "<|endoftext|>": 151643,
11
+ "<|file_sep|>": 151664,
12
+ "<|fim_middle|>": 151660,
13
+ "<|fim_pad|>": 151662,
14
+ "<|fim_prefix|>": 151659,
15
+ "<|fim_suffix|>": 151661,
16
+ "<|im_end|>": 151645,
17
+ "<|im_start|>": 151644,
18
+ "<|image_pad|>": 151655,
19
+ "<|object_ref_end|>": 151647,
20
+ "<|object_ref_start|>": 151646,
21
+ "<|quad_end|>": 151651,
22
+ "<|quad_start|>": 151650,
23
+ "<|repo_name|>": 151663,
24
+ "<|video_pad|>": 151656,
25
+ "<|vision_end|>": 151653,
26
+ "<|vision_pad|>": 151654,
27
+ "<|vision_start|>": 151652
28
+ }
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen3ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "head_dim": 128,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 1024,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 28,
16
+ "model_type": "qwen3",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 28,
19
+ "num_key_value_heads": 8,
20
+ "rms_norm_eps": 1e-06,
21
+ "rope_scaling": null,
22
+ "rope_theta": 1000000,
23
+ "sliding_window": null,
24
+ "tie_word_embeddings": true,
25
+ "torch_dtype": "float16",
26
+ "transformers_version": "4.51.3",
27
+ "use_cache": true,
28
+ "use_sliding_window": false,
29
+ "vocab_size": 151936
30
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_new_tokens": 2048,
5
+ "transformers_version": "4.51.3"
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f95ffa229dadd231079a8a87a9f00591c78a28250b6b9e55384fac04c769a5b
3
+ size 1192134784
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55081900245edda70945f546e6b163bcbcde54b0c0cf43dc35297527fd3b3520
3
+ size 1211416522
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ea1de72b328a933462665380d1b401f4c2949d717ec4fe1ca4671cf96135190
3
+ size 14244
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bee2c236427f40a4b100f2321c3b97769d80262935b385180e2a7b082d227f2
3
+ size 1064
special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
+ "eos_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "pad_token": {
25
+ "content": "<|endoftext|>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeb13307a71acd8fe81861d94ad54ab689df773318809eed3cbe794b4492dae4
3
+ size 11422654
tokenizer_config.json ADDED
@@ -0,0 +1,240 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ },
181
+ "151665": {
182
+ "content": "<tool_response>",
183
+ "lstrip": false,
184
+ "normalized": false,
185
+ "rstrip": false,
186
+ "single_word": false,
187
+ "special": false
188
+ },
189
+ "151666": {
190
+ "content": "</tool_response>",
191
+ "lstrip": false,
192
+ "normalized": false,
193
+ "rstrip": false,
194
+ "single_word": false,
195
+ "special": false
196
+ },
197
+ "151667": {
198
+ "content": "<think>",
199
+ "lstrip": false,
200
+ "normalized": false,
201
+ "rstrip": false,
202
+ "single_word": false,
203
+ "special": false
204
+ },
205
+ "151668": {
206
+ "content": "</think>",
207
+ "lstrip": false,
208
+ "normalized": false,
209
+ "rstrip": false,
210
+ "single_word": false,
211
+ "special": false
212
+ }
213
+ },
214
+ "additional_special_tokens": [
215
+ "<|im_start|>",
216
+ "<|im_end|>",
217
+ "<|object_ref_start|>",
218
+ "<|object_ref_end|>",
219
+ "<|box_start|>",
220
+ "<|box_end|>",
221
+ "<|quad_start|>",
222
+ "<|quad_end|>",
223
+ "<|vision_start|>",
224
+ "<|vision_end|>",
225
+ "<|vision_pad|>",
226
+ "<|image_pad|>",
227
+ "<|video_pad|>"
228
+ ],
229
+ "bos_token": null,
230
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0].role == 'system' %}\n {{- messages[0].content + '\\n\\n' }}\n {%- endif %}\n {{- \"# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0].role == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0].content + '<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- set ns = namespace(multi_step_tool=true, last_query_index=messages|length - 1) %}\n{%- for message in messages[::-1] %}\n {%- set index = (messages|length - 1) - loop.index0 %}\n {%- if ns.multi_step_tool and message.role == \"user\" and not(message.content.startswith('<tool_response>') and message.content.endswith('</tool_response>')) %}\n {%- set ns.multi_step_tool = false %}\n {%- set ns.last_query_index = index %}\n {%- endif %}\n{%- endfor %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {%- set content = message.content %}\n {%- set reasoning_content = '' %}\n {%- if message.reasoning_content is defined and message.reasoning_content is not none %}\n {%- set reasoning_content = message.reasoning_content %}\n {%- else %}\n {%- if '</think>' in message.content %}\n {%- set content = message.content.split('</think>')[-1].lstrip('\\n') %}\n {%- set reasoning_content = message.content.split('</think>')[0].rstrip('\\n').split('<think>')[-1].lstrip('\\n') %}\n {%- endif %}\n {%- endif %}\n {%- if loop.index0 > ns.last_query_index %}\n {%- if loop.last or (not loop.last and reasoning_content) %}\n {{- '<|im_start|>' + message.role + '\\n<think>\\n' + reasoning_content.strip('\\n') + '\\n</think>\\n\\n' + content.lstrip('\\n') }}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- else %}\n {{- '<|im_start|>' + message.role + '\\n' + content }}\n {%- endif %}\n {%- if message.tool_calls %}\n {%- for tool_call in message.tool_calls %}\n {%- if (loop.first and content) or (not loop.first) %}\n {{- '\\n' }}\n {%- endif %}\n {%- if tool_call.function %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {%- if tool_call.arguments is string %}\n {{- tool_call.arguments }}\n {%- else %}\n {{- tool_call.arguments | tojson }}\n {%- endif %}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {%- endif %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if loop.first or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n {%- if enable_thinking is defined and enable_thinking is false %}\n {{- '<think>\\n\\n</think>\\n\\n' }}\n {%- endif %}\n{%- endif %}",
231
+ "clean_up_tokenization_spaces": false,
232
+ "eos_token": "<|endoftext|>",
233
+ "errors": "replace",
234
+ "extra_special_tokens": {},
235
+ "model_max_length": 131072,
236
+ "pad_token": "<|endoftext|>",
237
+ "split_special_tokens": false,
238
+ "tokenizer_class": "Qwen2Tokenizer",
239
+ "unk_token": null
240
+ }
trainer_state.json ADDED
@@ -0,0 +1,1014 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 4.93339911198816,
6
+ "eval_steps": 1000,
7
+ "global_step": 10000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.0493339911198816,
14
+ "grad_norm": 3.404296875,
15
+ "learning_rate": 1.999987141198133e-05,
16
+ "loss": 1.1607,
17
+ "mean_token_accuracy": 0.7450588776171208,
18
+ "num_tokens": 817724.0,
19
+ "step": 100
20
+ },
21
+ {
22
+ "epoch": 0.0986679822397632,
23
+ "grad_norm": 3.275390625,
24
+ "learning_rate": 1.9998811023678148e-05,
25
+ "loss": 1.0497,
26
+ "mean_token_accuracy": 0.764276393353939,
27
+ "num_tokens": 1635495.0,
28
+ "step": 200
29
+ },
30
+ {
31
+ "epoch": 0.1480019733596448,
32
+ "grad_norm": 3.578125,
33
+ "learning_rate": 1.999667964947103e-05,
34
+ "loss": 1.0262,
35
+ "mean_token_accuracy": 0.7685012140870094,
36
+ "num_tokens": 2453891.0,
37
+ "step": 300
38
+ },
39
+ {
40
+ "epoch": 0.1973359644795264,
41
+ "grad_norm": 2.833984375,
42
+ "learning_rate": 1.999347751765429e-05,
43
+ "loss": 1.0036,
44
+ "mean_token_accuracy": 0.7728290051221848,
45
+ "num_tokens": 3271228.0,
46
+ "step": 400
47
+ },
48
+ {
49
+ "epoch": 0.246669955599408,
50
+ "grad_norm": 2.86328125,
51
+ "learning_rate": 1.998920497121251e-05,
52
+ "loss": 0.9978,
53
+ "mean_token_accuracy": 0.77385304749012,
54
+ "num_tokens": 4090128.0,
55
+ "step": 500
56
+ },
57
+ {
58
+ "epoch": 0.2960039467192896,
59
+ "grad_norm": 2.96484375,
60
+ "learning_rate": 1.998386246778381e-05,
61
+ "loss": 0.9859,
62
+ "mean_token_accuracy": 0.7768463261425496,
63
+ "num_tokens": 4908319.0,
64
+ "step": 600
65
+ },
66
+ {
67
+ "epoch": 0.3453379378391712,
68
+ "grad_norm": 3.03515625,
69
+ "learning_rate": 1.997745057961081e-05,
70
+ "loss": 0.9702,
71
+ "mean_token_accuracy": 0.7799235332012177,
72
+ "num_tokens": 5726519.0,
73
+ "step": 700
74
+ },
75
+ {
76
+ "epoch": 0.3946719289590528,
77
+ "grad_norm": 2.759765625,
78
+ "learning_rate": 1.9969969993479352e-05,
79
+ "loss": 0.9674,
80
+ "mean_token_accuracy": 0.7801838928461075,
81
+ "num_tokens": 6545163.0,
82
+ "step": 800
83
+ },
84
+ {
85
+ "epoch": 0.4440059200789344,
86
+ "grad_norm": 3.3046875,
87
+ "learning_rate": 1.9961421510644935e-05,
88
+ "loss": 0.9597,
89
+ "mean_token_accuracy": 0.7807393775880337,
90
+ "num_tokens": 7363259.0,
91
+ "step": 900
92
+ },
93
+ {
94
+ "epoch": 0.493339911198816,
95
+ "grad_norm": 3.087890625,
96
+ "learning_rate": 1.9951806046746892e-05,
97
+ "loss": 0.9476,
98
+ "step": 1000
99
+ },
100
+ {
101
+ "epoch": 0.493339911198816,
102
+ "eval_loss": 0.9531487822532654,
103
+ "eval_mean_token_accuracy": 0.784176768076837,
104
+ "eval_num_tokens": 8182192.0,
105
+ "eval_runtime": 232.2542,
106
+ "eval_samples_per_second": 69.863,
107
+ "eval_steps_per_second": 8.736,
108
+ "step": 1000
109
+ },
110
+ {
111
+ "epoch": 0.5426739023186976,
112
+ "grad_norm": 2.607421875,
113
+ "learning_rate": 1.9941124631710298e-05,
114
+ "loss": 0.9412,
115
+ "mean_token_accuracy": 0.7848582742363215,
116
+ "num_tokens": 9000632.0,
117
+ "step": 1100
118
+ },
119
+ {
120
+ "epoch": 0.5920078934385792,
121
+ "grad_norm": 2.814453125,
122
+ "learning_rate": 1.9929378409635686e-05,
123
+ "loss": 0.9383,
124
+ "mean_token_accuracy": 0.7856255556643009,
125
+ "num_tokens": 9817706.0,
126
+ "step": 1200
127
+ },
128
+ {
129
+ "epoch": 0.6413418845584608,
130
+ "grad_norm": 2.822265625,
131
+ "learning_rate": 1.9916568638676474e-05,
132
+ "loss": 0.9302,
133
+ "mean_token_accuracy": 0.7883923088014125,
134
+ "num_tokens": 10635361.0,
135
+ "step": 1300
136
+ },
137
+ {
138
+ "epoch": 0.6906758756783424,
139
+ "grad_norm": 2.779296875,
140
+ "learning_rate": 1.9902696690904203e-05,
141
+ "loss": 0.928,
142
+ "mean_token_accuracy": 0.788385953605175,
143
+ "num_tokens": 11452193.0,
144
+ "step": 1400
145
+ },
146
+ {
147
+ "epoch": 0.740009866798224,
148
+ "grad_norm": 2.681640625,
149
+ "learning_rate": 1.9887764052161598e-05,
150
+ "loss": 0.9187,
151
+ "mean_token_accuracy": 0.7905992804467679,
152
+ "num_tokens": 12270285.0,
153
+ "step": 1500
154
+ },
155
+ {
156
+ "epoch": 0.7893438579181056,
157
+ "grad_norm": 2.6796875,
158
+ "learning_rate": 1.9871772321903388e-05,
159
+ "loss": 0.9069,
160
+ "mean_token_accuracy": 0.7929588802158832,
161
+ "num_tokens": 13088068.0,
162
+ "step": 1600
163
+ },
164
+ {
165
+ "epoch": 0.8386778490379871,
166
+ "grad_norm": 2.859375,
167
+ "learning_rate": 1.9854723213024996e-05,
168
+ "loss": 0.9023,
169
+ "mean_token_accuracy": 0.7939252285659313,
170
+ "num_tokens": 13906129.0,
171
+ "step": 1700
172
+ },
173
+ {
174
+ "epoch": 0.8880118401578688,
175
+ "grad_norm": 2.439453125,
176
+ "learning_rate": 1.983661855167908e-05,
177
+ "loss": 0.895,
178
+ "mean_token_accuracy": 0.7945813001692295,
179
+ "num_tokens": 14724436.0,
180
+ "step": 1800
181
+ },
182
+ {
183
+ "epoch": 0.9373458312777504,
184
+ "grad_norm": 2.708984375,
185
+ "learning_rate": 1.981746027707992e-05,
186
+ "loss": 0.8826,
187
+ "mean_token_accuracy": 0.7975821754336357,
188
+ "num_tokens": 15542251.0,
189
+ "step": 1900
190
+ },
191
+ {
192
+ "epoch": 0.986679822397632,
193
+ "grad_norm": 2.47265625,
194
+ "learning_rate": 1.9797250441295696e-05,
195
+ "loss": 0.8814,
196
+ "step": 2000
197
+ },
198
+ {
199
+ "epoch": 0.986679822397632,
200
+ "eval_loss": 0.8888025879859924,
201
+ "eval_mean_token_accuracy": 0.7973491149679554,
202
+ "eval_num_tokens": 16360700.0,
203
+ "eval_runtime": 230.6477,
204
+ "eval_samples_per_second": 70.35,
205
+ "eval_steps_per_second": 8.797,
206
+ "step": 2000
207
+ },
208
+ {
209
+ "epoch": 1.0360138135175136,
210
+ "grad_norm": 2.748046875,
211
+ "learning_rate": 1.9775991209028724e-05,
212
+ "loss": 0.8088,
213
+ "mean_token_accuracy": 0.8046260391920805,
214
+ "num_tokens": 17176613.0,
215
+ "step": 2100
216
+ },
217
+ {
218
+ "epoch": 1.0853478046373952,
219
+ "grad_norm": 2.654296875,
220
+ "learning_rate": 1.9753684857383547e-05,
221
+ "loss": 0.7927,
222
+ "mean_token_accuracy": 0.813809906244278,
223
+ "num_tokens": 17994211.0,
224
+ "step": 2200
225
+ },
226
+ {
227
+ "epoch": 1.1346817957572768,
228
+ "grad_norm": 2.775390625,
229
+ "learning_rate": 1.9730333775623062e-05,
230
+ "loss": 0.7785,
231
+ "mean_token_accuracy": 0.8165469121932983,
232
+ "num_tokens": 18812403.0,
233
+ "step": 2300
234
+ },
235
+ {
236
+ "epoch": 1.1840157868771584,
237
+ "grad_norm": 2.703125,
238
+ "learning_rate": 1.9705940464912593e-05,
239
+ "loss": 0.7799,
240
+ "mean_token_accuracy": 0.8157392312586308,
241
+ "num_tokens": 19630645.0,
242
+ "step": 2400
243
+ },
244
+ {
245
+ "epoch": 1.23334977799704,
246
+ "grad_norm": 2.24609375,
247
+ "learning_rate": 1.9680507538051982e-05,
248
+ "loss": 0.7792,
249
+ "mean_token_accuracy": 0.8161792799830436,
250
+ "num_tokens": 20448907.0,
251
+ "step": 2500
252
+ },
253
+ {
254
+ "epoch": 1.2826837691169215,
255
+ "grad_norm": 2.34765625,
256
+ "learning_rate": 1.9654037719195753e-05,
257
+ "loss": 0.7791,
258
+ "mean_token_accuracy": 0.8167361421883106,
259
+ "num_tokens": 21267549.0,
260
+ "step": 2600
261
+ },
262
+ {
263
+ "epoch": 1.3320177602368033,
264
+ "grad_norm": 2.255859375,
265
+ "learning_rate": 1.9626533843561284e-05,
266
+ "loss": 0.7712,
267
+ "mean_token_accuracy": 0.8178707587718964,
268
+ "num_tokens": 22085783.0,
269
+ "step": 2700
270
+ },
271
+ {
272
+ "epoch": 1.3813517513566849,
273
+ "grad_norm": 2.62109375,
274
+ "learning_rate": 1.959799885712516e-05,
275
+ "loss": 0.7802,
276
+ "mean_token_accuracy": 0.8168295152485371,
277
+ "num_tokens": 22903659.0,
278
+ "step": 2800
279
+ },
280
+ {
281
+ "epoch": 1.4306857424765664,
282
+ "grad_norm": 2.36328125,
283
+ "learning_rate": 1.9568435816307606e-05,
284
+ "loss": 0.7786,
285
+ "mean_token_accuracy": 0.817071139216423,
286
+ "num_tokens": 23721992.0,
287
+ "step": 2900
288
+ },
289
+ {
290
+ "epoch": 1.480019733596448,
291
+ "grad_norm": 2.35546875,
292
+ "learning_rate": 1.9537847887645114e-05,
293
+ "loss": 0.7726,
294
+ "step": 3000
295
+ },
296
+ {
297
+ "epoch": 1.480019733596448,
298
+ "eval_loss": 0.8628731966018677,
299
+ "eval_mean_token_accuracy": 0.8050215809494478,
300
+ "eval_num_tokens": 24539842.0,
301
+ "eval_runtime": 258.3525,
302
+ "eval_samples_per_second": 62.806,
303
+ "eval_steps_per_second": 7.854,
304
+ "step": 3000
305
+ },
306
+ {
307
+ "epoch": 1.5293537247163296,
308
+ "grad_norm": 2.36328125,
309
+ "learning_rate": 1.9506238347451272e-05,
310
+ "loss": 0.7797,
311
+ "mean_token_accuracy": 0.8179577070474625,
312
+ "num_tokens": 25357355.0,
313
+ "step": 3100
314
+ },
315
+ {
316
+ "epoch": 1.5786877158362111,
317
+ "grad_norm": 2.5078125,
318
+ "learning_rate": 1.9473610581465835e-05,
319
+ "loss": 0.7734,
320
+ "mean_token_accuracy": 0.818172342479229,
321
+ "num_tokens": 26175340.0,
322
+ "step": 3200
323
+ },
324
+ {
325
+ "epoch": 1.6280217069560927,
326
+ "grad_norm": 2.435546875,
327
+ "learning_rate": 1.943996808449207e-05,
328
+ "loss": 0.7736,
329
+ "mean_token_accuracy": 0.8182803516089916,
330
+ "num_tokens": 26993802.0,
331
+ "step": 3300
332
+ },
333
+ {
334
+ "epoch": 1.6773556980759743,
335
+ "grad_norm": 2.40234375,
336
+ "learning_rate": 1.940531446002243e-05,
337
+ "loss": 0.7706,
338
+ "mean_token_accuracy": 0.8192763808369636,
339
+ "num_tokens": 27811947.0,
340
+ "step": 3400
341
+ },
342
+ {
343
+ "epoch": 1.7266896891958559,
344
+ "grad_norm": 2.5390625,
345
+ "learning_rate": 1.9369653419852568e-05,
346
+ "loss": 0.7673,
347
+ "mean_token_accuracy": 0.8196577854454518,
348
+ "num_tokens": 28629348.0,
349
+ "step": 3500
350
+ },
351
+ {
352
+ "epoch": 1.7760236803157374,
353
+ "grad_norm": 2.453125,
354
+ "learning_rate": 1.933298878368378e-05,
355
+ "loss": 0.7637,
356
+ "mean_token_accuracy": 0.8204946468770504,
357
+ "num_tokens": 29447420.0,
358
+ "step": 3600
359
+ },
360
+ {
361
+ "epoch": 1.825357671435619,
362
+ "grad_norm": 2.337890625,
363
+ "learning_rate": 1.929532447871384e-05,
364
+ "loss": 0.7558,
365
+ "mean_token_accuracy": 0.8219007922708989,
366
+ "num_tokens": 30265221.0,
367
+ "step": 3700
368
+ },
369
+ {
370
+ "epoch": 1.8746916625555008,
371
+ "grad_norm": 2.48828125,
372
+ "learning_rate": 1.925666453921639e-05,
373
+ "loss": 0.7587,
374
+ "mean_token_accuracy": 0.82237908706069,
375
+ "num_tokens": 31082572.0,
376
+ "step": 3800
377
+ },
378
+ {
379
+ "epoch": 1.9240256536753824,
380
+ "grad_norm": 2.51953125,
381
+ "learning_rate": 1.9217013106108798e-05,
382
+ "loss": 0.7501,
383
+ "mean_token_accuracy": 0.8232213893532753,
384
+ "num_tokens": 31900505.0,
385
+ "step": 3900
386
+ },
387
+ {
388
+ "epoch": 1.973359644795264,
389
+ "grad_norm": 2.33203125,
390
+ "learning_rate": 1.917637442650863e-05,
391
+ "loss": 0.7557,
392
+ "step": 4000
393
+ },
394
+ {
395
+ "epoch": 1.973359644795264,
396
+ "eval_loss": 0.8336860537528992,
397
+ "eval_mean_token_accuracy": 0.8114011559420586,
398
+ "eval_num_tokens": 32718869.0,
399
+ "eval_runtime": 301.9223,
400
+ "eval_samples_per_second": 53.742,
401
+ "eval_steps_per_second": 6.72,
402
+ "step": 4000
403
+ },
404
+ {
405
+ "epoch": 2.0226936359151457,
406
+ "grad_norm": 2.255859375,
407
+ "learning_rate": 1.913475285327874e-05,
408
+ "loss": 0.7076,
409
+ "mean_token_accuracy": 0.826932647228241,
410
+ "num_tokens": 33535218.0,
411
+ "step": 4100
412
+ },
413
+ {
414
+ "epoch": 2.0720276270350273,
415
+ "grad_norm": 2.8046875,
416
+ "learning_rate": 1.9092152844561e-05,
417
+ "loss": 0.6484,
418
+ "mean_token_accuracy": 0.8430611005425453,
419
+ "num_tokens": 34353569.0,
420
+ "step": 4200
421
+ },
422
+ {
423
+ "epoch": 2.121361618154909,
424
+ "grad_norm": 2.31640625,
425
+ "learning_rate": 1.904857896329882e-05,
426
+ "loss": 0.6535,
427
+ "mean_token_accuracy": 0.8420212762057782,
428
+ "num_tokens": 35171634.0,
429
+ "step": 4300
430
+ },
431
+ {
432
+ "epoch": 2.1706956092747904,
433
+ "grad_norm": 2.474609375,
434
+ "learning_rate": 1.9004035876748393e-05,
435
+ "loss": 0.65,
436
+ "mean_token_accuracy": 0.8425224512815476,
437
+ "num_tokens": 35989510.0,
438
+ "step": 4400
439
+ },
440
+ {
441
+ "epoch": 2.220029600394672,
442
+ "grad_norm": 2.37109375,
443
+ "learning_rate": 1.8958528355978767e-05,
444
+ "loss": 0.6482,
445
+ "mean_token_accuracy": 0.8430168768763542,
446
+ "num_tokens": 36807593.0,
447
+ "step": 4500
448
+ },
449
+ {
450
+ "epoch": 2.2693635915145536,
451
+ "grad_norm": 2.544921875,
452
+ "learning_rate": 1.8912061275360817e-05,
453
+ "loss": 0.6563,
454
+ "mean_token_accuracy": 0.8413836374878884,
455
+ "num_tokens": 37625695.0,
456
+ "step": 4600
457
+ },
458
+ {
459
+ "epoch": 2.318697582634435,
460
+ "grad_norm": 2.4140625,
461
+ "learning_rate": 1.8864639612045153e-05,
462
+ "loss": 0.6556,
463
+ "mean_token_accuracy": 0.841529670804739,
464
+ "num_tokens": 38443841.0,
465
+ "step": 4700
466
+ },
467
+ {
468
+ "epoch": 2.3680315737543167,
469
+ "grad_norm": 2.556640625,
470
+ "learning_rate": 1.8816268445428996e-05,
471
+ "loss": 0.6495,
472
+ "mean_token_accuracy": 0.8425976119935512,
473
+ "num_tokens": 39262075.0,
474
+ "step": 4800
475
+ },
476
+ {
477
+ "epoch": 2.4173655648741983,
478
+ "grad_norm": 2.5078125,
479
+ "learning_rate": 1.8766952956612123e-05,
480
+ "loss": 0.655,
481
+ "mean_token_accuracy": 0.8413432243466378,
482
+ "num_tokens": 40079953.0,
483
+ "step": 4900
484
+ },
485
+ {
486
+ "epoch": 2.46669955599408,
487
+ "grad_norm": 2.724609375,
488
+ "learning_rate": 1.8716698427841926e-05,
489
+ "loss": 0.6576,
490
+ "step": 5000
491
+ },
492
+ {
493
+ "epoch": 2.46669955599408,
494
+ "eval_loss": 0.836199939250946,
495
+ "eval_mean_token_accuracy": 0.8145093770377327,
496
+ "eval_num_tokens": 40898183.0,
497
+ "eval_runtime": 227.0413,
498
+ "eval_samples_per_second": 71.467,
499
+ "eval_steps_per_second": 8.937,
500
+ "step": 5000
501
+ },
502
+ {
503
+ "epoch": 2.5160335471139614,
504
+ "grad_norm": 2.349609375,
505
+ "learning_rate": 1.8665510241947596e-05,
506
+ "loss": 0.6579,
507
+ "mean_token_accuracy": 0.8412897626310587,
508
+ "num_tokens": 41716178.0,
509
+ "step": 5100
510
+ },
511
+ {
512
+ "epoch": 2.565367538233843,
513
+ "grad_norm": 2.609375,
514
+ "learning_rate": 1.8613393881763583e-05,
515
+ "loss": 0.6552,
516
+ "mean_token_accuracy": 0.8418721158802509,
517
+ "num_tokens": 42533359.0,
518
+ "step": 5200
519
+ },
520
+ {
521
+ "epoch": 2.6147015293537246,
522
+ "grad_norm": 2.529296875,
523
+ "learning_rate": 1.8560354929542322e-05,
524
+ "loss": 0.6504,
525
+ "mean_token_accuracy": 0.8428675523400306,
526
+ "num_tokens": 43352238.0,
527
+ "step": 5300
528
+ },
529
+ {
530
+ "epoch": 2.6640355204736066,
531
+ "grad_norm": 2.5703125,
532
+ "learning_rate": 1.8506399066356294e-05,
533
+ "loss": 0.6562,
534
+ "mean_token_accuracy": 0.8421946428716183,
535
+ "num_tokens": 44170387.0,
536
+ "step": 5400
537
+ },
538
+ {
539
+ "epoch": 2.7133695115934877,
540
+ "grad_norm": 2.55078125,
541
+ "learning_rate": 1.8451532071489532e-05,
542
+ "loss": 0.6579,
543
+ "mean_token_accuracy": 0.8413851109147071,
544
+ "num_tokens": 44988361.0,
545
+ "step": 5500
546
+ },
547
+ {
548
+ "epoch": 2.7627035027133697,
549
+ "grad_norm": 2.5703125,
550
+ "learning_rate": 1.839575982181859e-05,
551
+ "loss": 0.6559,
552
+ "mean_token_accuracy": 0.8422524558007717,
553
+ "num_tokens": 45805576.0,
554
+ "step": 5600
555
+ },
556
+ {
557
+ "epoch": 2.812037493833251,
558
+ "grad_norm": 2.552734375,
559
+ "learning_rate": 1.8339088291183072e-05,
560
+ "loss": 0.6524,
561
+ "mean_token_accuracy": 0.8429344496130944,
562
+ "num_tokens": 46623168.0,
563
+ "step": 5700
564
+ },
565
+ {
566
+ "epoch": 2.861371484953133,
567
+ "grad_norm": 2.591796875,
568
+ "learning_rate": 1.828152354974575e-05,
569
+ "loss": 0.6572,
570
+ "mean_token_accuracy": 0.841839095801115,
571
+ "num_tokens": 47441605.0,
572
+ "step": 5800
573
+ },
574
+ {
575
+ "epoch": 2.9107054760730144,
576
+ "grad_norm": 2.5625,
577
+ "learning_rate": 1.8223071763342388e-05,
578
+ "loss": 0.6569,
579
+ "mean_token_accuracy": 0.8422958692908287,
580
+ "num_tokens": 48259542.0,
581
+ "step": 5900
582
+ },
583
+ {
584
+ "epoch": 2.960039467192896,
585
+ "grad_norm": 2.3671875,
586
+ "learning_rate": 1.8163739192821325e-05,
587
+ "loss": 0.6521,
588
+ "step": 6000
589
+ },
590
+ {
591
+ "epoch": 2.960039467192896,
592
+ "eval_loss": 0.8176103830337524,
593
+ "eval_mean_token_accuracy": 0.8183424575626116,
594
+ "eval_num_tokens": 49077499.0,
595
+ "eval_runtime": 227.0382,
596
+ "eval_samples_per_second": 71.468,
597
+ "eval_steps_per_second": 8.937,
598
+ "step": 6000
599
+ },
600
+ {
601
+ "epoch": 3.0093734583127776,
602
+ "grad_norm": 2.68359375,
603
+ "learning_rate": 1.8103532193372832e-05,
604
+ "loss": 0.6343,
605
+ "mean_token_accuracy": 0.8449597600102424,
606
+ "num_tokens": 49893961.0,
607
+ "step": 6100
608
+ },
609
+ {
610
+ "epoch": 3.058707449432659,
611
+ "grad_norm": 2.630859375,
612
+ "learning_rate": 1.8042457213848448e-05,
613
+ "loss": 0.5458,
614
+ "mean_token_accuracy": 0.8658474875986576,
615
+ "num_tokens": 50711532.0,
616
+ "step": 6200
617
+ },
618
+ {
619
+ "epoch": 3.1080414405525407,
620
+ "grad_norm": 2.708984375,
621
+ "learning_rate": 1.798052079607019e-05,
622
+ "loss": 0.538,
623
+ "mean_token_accuracy": 0.8663399314880371,
624
+ "num_tokens": 51528714.0,
625
+ "step": 6300
626
+ },
627
+ {
628
+ "epoch": 3.1573754316724223,
629
+ "grad_norm": 2.62890625,
630
+ "learning_rate": 1.791772957412987e-05,
631
+ "loss": 0.5517,
632
+ "mean_token_accuracy": 0.8640760770440101,
633
+ "num_tokens": 52345966.0,
634
+ "step": 6400
635
+ },
636
+ {
637
+ "epoch": 3.206709422792304,
638
+ "grad_norm": 2.859375,
639
+ "learning_rate": 1.785409027367852e-05,
640
+ "loss": 0.5442,
641
+ "mean_token_accuracy": 0.8657858520746231,
642
+ "num_tokens": 53163793.0,
643
+ "step": 6500
644
+ },
645
+ {
646
+ "epoch": 3.2560434139121854,
647
+ "grad_norm": 2.669921875,
648
+ "learning_rate": 1.7789609711205967e-05,
649
+ "loss": 0.5516,
650
+ "mean_token_accuracy": 0.863896958976984,
651
+ "num_tokens": 53981672.0,
652
+ "step": 6600
653
+ },
654
+ {
655
+ "epoch": 3.305377405032067,
656
+ "grad_norm": 2.638671875,
657
+ "learning_rate": 1.7724294793310742e-05,
658
+ "loss": 0.552,
659
+ "mean_token_accuracy": 0.8638136276602745,
660
+ "num_tokens": 54800133.0,
661
+ "step": 6700
662
+ },
663
+ {
664
+ "epoch": 3.3547113961519486,
665
+ "grad_norm": 2.669921875,
666
+ "learning_rate": 1.765815251596029e-05,
667
+ "loss": 0.5537,
668
+ "mean_token_accuracy": 0.8634848801791668,
669
+ "num_tokens": 55617519.0,
670
+ "step": 6800
671
+ },
672
+ {
673
+ "epoch": 3.40404538727183,
674
+ "grad_norm": 2.712890625,
675
+ "learning_rate": 1.7591189963741614e-05,
676
+ "loss": 0.5526,
677
+ "mean_token_accuracy": 0.8636865784227848,
678
+ "num_tokens": 56435580.0,
679
+ "step": 6900
680
+ },
681
+ {
682
+ "epoch": 3.4533793783917117,
683
+ "grad_norm": 2.89453125,
684
+ "learning_rate": 1.7523414309102462e-05,
685
+ "loss": 0.5566,
686
+ "step": 7000
687
+ },
688
+ {
689
+ "epoch": 3.4533793783917117,
690
+ "eval_loss": 0.8430932760238647,
691
+ "eval_mean_token_accuracy": 0.8191284864274312,
692
+ "eval_num_tokens": 57254348.0,
693
+ "eval_runtime": 227.0935,
694
+ "eval_samples_per_second": 71.451,
695
+ "eval_steps_per_second": 8.935,
696
+ "step": 7000
697
+ },
698
+ {
699
+ "epoch": 3.5027133695115937,
700
+ "grad_norm": 2.541015625,
701
+ "learning_rate": 1.7454832811583045e-05,
702
+ "loss": 0.553,
703
+ "mean_token_accuracy": 0.8629767662286758,
704
+ "num_tokens": 58072994.0,
705
+ "step": 7100
706
+ },
707
+ {
708
+ "epoch": 3.552047360631475,
709
+ "grad_norm": 2.796875,
710
+ "learning_rate": 1.738545281703848e-05,
711
+ "loss": 0.5539,
712
+ "mean_token_accuracy": 0.8632567670941352,
713
+ "num_tokens": 58890672.0,
714
+ "step": 7200
715
+ },
716
+ {
717
+ "epoch": 3.601381351751357,
718
+ "grad_norm": 2.82421875,
719
+ "learning_rate": 1.731528175685196e-05,
720
+ "loss": 0.5572,
721
+ "mean_token_accuracy": 0.8626312711834907,
722
+ "num_tokens": 59708564.0,
723
+ "step": 7300
724
+ },
725
+ {
726
+ "epoch": 3.6507153428712384,
727
+ "grad_norm": 2.693359375,
728
+ "learning_rate": 1.7244327147138765e-05,
729
+ "loss": 0.5553,
730
+ "mean_token_accuracy": 0.8632141479849815,
731
+ "num_tokens": 60526575.0,
732
+ "step": 7400
733
+ },
734
+ {
735
+ "epoch": 3.70004933399112,
736
+ "grad_norm": 2.8125,
737
+ "learning_rate": 1.7172596587941203e-05,
738
+ "loss": 0.5573,
739
+ "mean_token_accuracy": 0.8632130342721939,
740
+ "num_tokens": 61345050.0,
741
+ "step": 7500
742
+ },
743
+ {
744
+ "epoch": 3.7493833251110016,
745
+ "grad_norm": 2.921875,
746
+ "learning_rate": 1.710009776241456e-05,
747
+ "loss": 0.5579,
748
+ "mean_token_accuracy": 0.8623562103509903,
749
+ "num_tokens": 62161790.0,
750
+ "step": 7600
751
+ },
752
+ {
753
+ "epoch": 3.798717316230883,
754
+ "grad_norm": 2.767578125,
755
+ "learning_rate": 1.702683843600415e-05,
756
+ "loss": 0.5567,
757
+ "mean_token_accuracy": 0.8631315796077251,
758
+ "num_tokens": 62980263.0,
759
+ "step": 7700
760
+ },
761
+ {
762
+ "epoch": 3.8480513073507647,
763
+ "grad_norm": 2.58984375,
764
+ "learning_rate": 1.6952826455613546e-05,
765
+ "loss": 0.5595,
766
+ "mean_token_accuracy": 0.862657565176487,
767
+ "num_tokens": 63798977.0,
768
+ "step": 7800
769
+ },
770
+ {
771
+ "epoch": 3.8973852984706463,
772
+ "grad_norm": 2.658203125,
773
+ "learning_rate": 1.687806974876408e-05,
774
+ "loss": 0.5531,
775
+ "mean_token_accuracy": 0.8633909998834133,
776
+ "num_tokens": 64617733.0,
777
+ "step": 7900
778
+ },
779
+ {
780
+ "epoch": 3.946719289590528,
781
+ "grad_norm": 3.1953125,
782
+ "learning_rate": 1.680257632274572e-05,
783
+ "loss": 0.5546,
784
+ "step": 8000
785
+ },
786
+ {
787
+ "epoch": 3.946719289590528,
788
+ "eval_loss": 0.8310380578041077,
789
+ "eval_mean_token_accuracy": 0.8215583155107005,
790
+ "eval_num_tokens": 65435962.0,
791
+ "eval_runtime": 227.0562,
792
+ "eval_samples_per_second": 71.462,
793
+ "eval_steps_per_second": 8.936,
794
+ "step": 8000
795
+ },
796
+ {
797
+ "epoch": 3.9960532807104094,
798
+ "grad_norm": 2.80078125,
799
+ "learning_rate": 1.6726354263759423e-05,
800
+ "loss": 0.5585,
801
+ "mean_token_accuracy": 0.8631708553433418,
802
+ "num_tokens": 66254184.0,
803
+ "step": 8100
804
+ },
805
+ {
806
+ "epoch": 4.0453872718302915,
807
+ "grad_norm": 2.806640625,
808
+ "learning_rate": 1.6649411736050957e-05,
809
+ "loss": 0.4575,
810
+ "mean_token_accuracy": 0.8858809275925159,
811
+ "num_tokens": 67070561.0,
812
+ "step": 8200
813
+ },
814
+ {
815
+ "epoch": 4.094721262950173,
816
+ "grad_norm": 3.056640625,
817
+ "learning_rate": 1.6571756981036476e-05,
818
+ "loss": 0.4536,
819
+ "mean_token_accuracy": 0.8867035652697086,
820
+ "num_tokens": 67889388.0,
821
+ "step": 8300
822
+ },
823
+ {
824
+ "epoch": 4.144055254070055,
825
+ "grad_norm": 2.76171875,
826
+ "learning_rate": 1.6493398316419727e-05,
827
+ "loss": 0.4523,
828
+ "mean_token_accuracy": 0.8867955373227596,
829
+ "num_tokens": 68706133.0,
830
+ "step": 8400
831
+ },
832
+ {
833
+ "epoch": 4.193389245189936,
834
+ "grad_norm": 3.005859375,
835
+ "learning_rate": 1.641434413530116e-05,
836
+ "loss": 0.458,
837
+ "mean_token_accuracy": 0.8855340279638767,
838
+ "num_tokens": 69523848.0,
839
+ "step": 8500
840
+ },
841
+ {
842
+ "epoch": 4.242723236309818,
843
+ "grad_norm": 2.94140625,
844
+ "learning_rate": 1.6334602905278917e-05,
845
+ "loss": 0.4607,
846
+ "mean_token_accuracy": 0.8852245907485485,
847
+ "num_tokens": 70342124.0,
848
+ "step": 8600
849
+ },
850
+ {
851
+ "epoch": 4.292057227429699,
852
+ "grad_norm": 2.60546875,
853
+ "learning_rate": 1.6254183167541848e-05,
854
+ "loss": 0.4618,
855
+ "mean_token_accuracy": 0.884294263869524,
856
+ "num_tokens": 71160314.0,
857
+ "step": 8700
858
+ },
859
+ {
860
+ "epoch": 4.341391218549581,
861
+ "grad_norm": 2.896484375,
862
+ "learning_rate": 1.617309353595468e-05,
863
+ "loss": 0.462,
864
+ "mean_token_accuracy": 0.884438044577837,
865
+ "num_tokens": 71979044.0,
866
+ "step": 8800
867
+ },
868
+ {
869
+ "epoch": 4.390725209669462,
870
+ "grad_norm": 2.900390625,
871
+ "learning_rate": 1.6091342696135354e-05,
872
+ "loss": 0.4601,
873
+ "mean_token_accuracy": 0.8847412486374379,
874
+ "num_tokens": 72797208.0,
875
+ "step": 8900
876
+ },
877
+ {
878
+ "epoch": 4.440059200789344,
879
+ "grad_norm": 2.693359375,
880
+ "learning_rate": 1.6008939404524686e-05,
881
+ "loss": 0.4664,
882
+ "step": 9000
883
+ },
884
+ {
885
+ "epoch": 4.440059200789344,
886
+ "eval_loss": 0.8724967837333679,
887
+ "eval_mean_token_accuracy": 0.8213894294658396,
888
+ "eval_num_tokens": 73615031.0,
889
+ "eval_runtime": 227.1181,
890
+ "eval_samples_per_second": 71.443,
891
+ "eval_steps_per_second": 8.934,
892
+ "step": 9000
893
+ },
894
+ {
895
+ "epoch": 4.489393191909225,
896
+ "grad_norm": 2.93359375,
897
+ "learning_rate": 1.5925892487448492e-05,
898
+ "loss": 0.4624,
899
+ "mean_token_accuracy": 0.8841097000986338,
900
+ "num_tokens": 74433117.0,
901
+ "step": 9100
902
+ },
903
+ {
904
+ "epoch": 4.538727183029107,
905
+ "grad_norm": 3.052734375,
906
+ "learning_rate": 1.584221084017215e-05,
907
+ "loss": 0.4687,
908
+ "mean_token_accuracy": 0.8826957462728023,
909
+ "num_tokens": 75251550.0,
910
+ "step": 9200
911
+ },
912
+ {
913
+ "epoch": 4.588061174148988,
914
+ "grad_norm": 2.978515625,
915
+ "learning_rate": 1.5757903425947833e-05,
916
+ "loss": 0.4656,
917
+ "mean_token_accuracy": 0.8832169409096241,
918
+ "num_tokens": 76069101.0,
919
+ "step": 9300
920
+ },
921
+ {
922
+ "epoch": 4.63739516526887,
923
+ "grad_norm": 2.970703125,
924
+ "learning_rate": 1.5672979275054444e-05,
925
+ "loss": 0.4678,
926
+ "mean_token_accuracy": 0.8832441847026348,
927
+ "num_tokens": 76886994.0,
928
+ "step": 9400
929
+ },
930
+ {
931
+ "epoch": 4.686729156388752,
932
+ "grad_norm": 2.892578125,
933
+ "learning_rate": 1.5587447483830364e-05,
934
+ "loss": 0.4697,
935
+ "mean_token_accuracy": 0.8826014402508736,
936
+ "num_tokens": 77705672.0,
937
+ "step": 9500
938
+ },
939
+ {
940
+ "epoch": 4.736063147508633,
941
+ "grad_norm": 2.853515625,
942
+ "learning_rate": 1.5501317213699145e-05,
943
+ "loss": 0.4668,
944
+ "mean_token_accuracy": 0.8834595142304897,
945
+ "num_tokens": 78523274.0,
946
+ "step": 9600
947
+ },
948
+ {
949
+ "epoch": 4.785397138628515,
950
+ "grad_norm": 2.931640625,
951
+ "learning_rate": 1.5414597690188197e-05,
952
+ "loss": 0.4741,
953
+ "mean_token_accuracy": 0.8814607061445713,
954
+ "num_tokens": 79339785.0,
955
+ "step": 9700
956
+ },
957
+ {
958
+ "epoch": 4.834731129748397,
959
+ "grad_norm": 2.744140625,
960
+ "learning_rate": 1.5327298201940647e-05,
961
+ "loss": 0.4693,
962
+ "mean_token_accuracy": 0.8828557208180428,
963
+ "num_tokens": 80158261.0,
964
+ "step": 9800
965
+ },
966
+ {
967
+ "epoch": 4.884065120868279,
968
+ "grad_norm": 2.93359375,
969
+ "learning_rate": 1.523942809972041e-05,
970
+ "loss": 0.4684,
971
+ "mean_token_accuracy": 0.8827220787107944,
972
+ "num_tokens": 80976546.0,
973
+ "step": 9900
974
+ },
975
+ {
976
+ "epoch": 4.93339911198816,
977
+ "grad_norm": 2.7421875,
978
+ "learning_rate": 1.5150996795410626e-05,
979
+ "loss": 0.4701,
980
+ "step": 10000
981
+ },
982
+ {
983
+ "epoch": 4.93339911198816,
984
+ "eval_loss": 0.8658666014671326,
985
+ "eval_mean_token_accuracy": 0.8231359128002582,
986
+ "eval_num_tokens": 81794889.0,
987
+ "eval_runtime": 227.1166,
988
+ "eval_samples_per_second": 71.443,
989
+ "eval_steps_per_second": 8.934,
990
+ "step": 10000
991
+ }
992
+ ],
993
+ "logging_steps": 100,
994
+ "max_steps": 30405,
995
+ "num_input_tokens_seen": 0,
996
+ "num_train_epochs": 15,
997
+ "save_steps": 5000,
998
+ "stateful_callbacks": {
999
+ "TrainerControl": {
1000
+ "args": {
1001
+ "should_epoch_stop": false,
1002
+ "should_evaluate": false,
1003
+ "should_log": false,
1004
+ "should_save": true,
1005
+ "should_training_stop": false
1006
+ },
1007
+ "attributes": {}
1008
+ }
1009
+ },
1010
+ "total_flos": 2.1647962034877235e+17,
1011
+ "train_batch_size": 8,
1012
+ "trial_name": null,
1013
+ "trial_params": null
1014
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7719ba17d89e08cdd105a499e86ff3f214aa6a822ff7e67bc0320c9b2d8d554a
3
+ size 5624
vocab.json ADDED
The diff for this file is too large to render. See raw diff