mrhuseyn4 commited on
Commit
4a66fbd
·
verified ·
1 Parent(s): 1cdad9c

Upload trained H100 optimized model

Browse files
Files changed (45) hide show
  1. .gitattributes +2 -0
  2. chat_template.jinja +331 -0
  3. checkpoint-1250/chat_template.jinja +331 -0
  4. checkpoint-1250/config.json +69 -0
  5. checkpoint-1250/generation_config.json +11 -0
  6. checkpoint-1250/model-00001-of-00009.safetensors +3 -0
  7. checkpoint-1250/model-00002-of-00009.safetensors +3 -0
  8. checkpoint-1250/model-00003-of-00009.safetensors +3 -0
  9. checkpoint-1250/model-00004-of-00009.safetensors +3 -0
  10. checkpoint-1250/model-00005-of-00009.safetensors +3 -0
  11. checkpoint-1250/model-00006-of-00009.safetensors +3 -0
  12. checkpoint-1250/model-00007-of-00009.safetensors +3 -0
  13. checkpoint-1250/model-00008-of-00009.safetensors +3 -0
  14. checkpoint-1250/model-00009-of-00009.safetensors +3 -0
  15. checkpoint-1250/model.safetensors.index.json +419 -0
  16. checkpoint-1250/optimizer.pt +3 -0
  17. checkpoint-1250/rng_state.pth +3 -0
  18. checkpoint-1250/scheduler.pt +3 -0
  19. checkpoint-1250/special_tokens_map.json +23 -0
  20. checkpoint-1250/tokenizer.json +3 -0
  21. checkpoint-1250/tokenizer_config.json +183 -0
  22. checkpoint-1250/trainer_state.json +916 -0
  23. checkpoint-1250/training_args.bin +3 -0
  24. config.json +69 -0
  25. generation_config.json +11 -0
  26. logs/events.out.tfevents.1763377278.idealistic-pudu-from-uranus-fa7e.10674.0 +3 -0
  27. logs/events.out.tfevents.1763377589.idealistic-pudu-from-uranus-fa7e.11740.0 +3 -0
  28. logs/events.out.tfevents.1763377820.idealistic-pudu-from-uranus-fa7e.12716.0 +3 -0
  29. logs/events.out.tfevents.1763377925.idealistic-pudu-from-uranus-fa7e.13187.0 +3 -0
  30. logs/events.out.tfevents.1763378147.idealistic-pudu-from-uranus-fa7e.14024.0 +3 -0
  31. logs/events.out.tfevents.1763378228.idealistic-pudu-from-uranus-fa7e.14429.0 +3 -0
  32. model-00001-of-00009.safetensors +3 -0
  33. model-00002-of-00009.safetensors +3 -0
  34. model-00003-of-00009.safetensors +3 -0
  35. model-00004-of-00009.safetensors +3 -0
  36. model-00005-of-00009.safetensors +3 -0
  37. model-00006-of-00009.safetensors +3 -0
  38. model-00007-of-00009.safetensors +3 -0
  39. model-00008-of-00009.safetensors +3 -0
  40. model-00009-of-00009.safetensors +3 -0
  41. model.safetensors.index.json +419 -0
  42. special_tokens_map.json +23 -0
  43. tokenizer.json +3 -0
  44. tokenizer_config.json +183 -0
  45. training_args.bin +3 -0
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-1250/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
chat_template.jinja ADDED
@@ -0,0 +1,331 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {#-
2
+ In addition to the normal inputs of `messages` and `tools`, this template also accepts the
3
+ following kwargs:
4
+ - "builtin_tools": A list, can contain "browser" and/or "python".
5
+ - "model_identity": A string that optionally describes the model identity.
6
+ - "reasoning_effort": A string that describes the reasoning effort, defaults to "medium".
7
+ #}
8
+
9
+ {#- Tool Definition Rendering ============================================== #}
10
+ {%- macro render_typescript_type(param_spec, required_params, is_nullable=false) -%}
11
+ {%- if param_spec.type == "array" -%}
12
+ {%- if param_spec['items'] -%}
13
+ {%- if param_spec['items']['type'] == "string" -%}
14
+ {{- "string[]" }}
15
+ {%- elif param_spec['items']['type'] == "number" -%}
16
+ {{- "number[]" }}
17
+ {%- elif param_spec['items']['type'] == "integer" -%}
18
+ {{- "number[]" }}
19
+ {%- elif param_spec['items']['type'] == "boolean" -%}
20
+ {{- "boolean[]" }}
21
+ {%- else -%}
22
+ {%- set inner_type = render_typescript_type(param_spec['items'], required_params) -%}
23
+ {%- if inner_type == "object | object" or inner_type|length > 50 -%}
24
+ {{- "any[]" }}
25
+ {%- else -%}
26
+ {{- inner_type + "[]" }}
27
+ {%- endif -%}
28
+ {%- endif -%}
29
+ {%- if param_spec.nullable -%}
30
+ {{- " | null" }}
31
+ {%- endif -%}
32
+ {%- else -%}
33
+ {{- "any[]" }}
34
+ {%- if param_spec.nullable -%}
35
+ {{- " | null" }}
36
+ {%- endif -%}
37
+ {%- endif -%}
38
+ {%- elif param_spec.type is defined and param_spec.type is iterable and param_spec.type is not string and param_spec.type is not mapping and param_spec.type[0] is defined -%}
39
+ {#- Handle array of types like ["object", "object"] from Union[dict, list] #}
40
+ {%- if param_spec.type | length > 1 -%}
41
+ {{- param_spec.type | join(" | ") }}
42
+ {%- else -%}
43
+ {{- param_spec.type[0] }}
44
+ {%- endif -%}
45
+ {%- elif param_spec.oneOf -%}
46
+ {#- Handle oneOf schemas - check for complex unions and fallback to any #}
47
+ {%- set has_object_variants = false -%}
48
+ {%- for variant in param_spec.oneOf -%}
49
+ {%- if variant.type == "object" -%}
50
+ {%- set has_object_variants = true -%}
51
+ {%- endif -%}
52
+ {%- endfor -%}
53
+ {%- if has_object_variants and param_spec.oneOf|length > 1 -%}
54
+ {{- "any" }}
55
+ {%- else -%}
56
+ {%- for variant in param_spec.oneOf -%}
57
+ {{- render_typescript_type(variant, required_params) -}}
58
+ {%- if variant.description %}
59
+ {{- "// " + variant.description }}
60
+ {%- endif -%}
61
+ {%- if variant.default is defined %}
62
+ {{ "// default: " + variant.default|tojson }}
63
+ {%- endif -%}
64
+ {%- if not loop.last %}
65
+ {{- " | " }}
66
+ {% endif -%}
67
+ {%- endfor -%}
68
+ {%- endif -%}
69
+ {%- elif param_spec.type == "string" -%}
70
+ {%- if param_spec.enum -%}
71
+ {{- '"' + param_spec.enum|join('" | "') + '"' -}}
72
+ {%- else -%}
73
+ {{- "string" }}
74
+ {%- if param_spec.nullable %}
75
+ {{- " | null" }}
76
+ {%- endif -%}
77
+ {%- endif -%}
78
+ {%- elif param_spec.type == "number" -%}
79
+ {{- "number" }}
80
+ {%- elif param_spec.type == "integer" -%}
81
+ {{- "number" }}
82
+ {%- elif param_spec.type == "boolean" -%}
83
+ {{- "boolean" }}
84
+
85
+ {%- elif param_spec.type == "object" -%}
86
+ {%- if param_spec.properties -%}
87
+ {{- "{\n" }}
88
+ {%- for prop_name, prop_spec in param_spec.properties.items() -%}
89
+ {{- prop_name -}}
90
+ {%- if prop_name not in (param_spec.required or []) -%}
91
+ {{- "?" }}
92
+ {%- endif -%}
93
+ {{- ": " }}
94
+ {{ render_typescript_type(prop_spec, param_spec.required or []) }}
95
+ {%- if not loop.last -%}
96
+ {{-", " }}
97
+ {%- endif -%}
98
+ {%- endfor -%}
99
+ {{- "}" }}
100
+ {%- else -%}
101
+ {{- "object" }}
102
+ {%- endif -%}
103
+ {%- else -%}
104
+ {{- "any" }}
105
+ {%- endif -%}
106
+ {%- endmacro -%}
107
+
108
+ {%- macro render_tool_namespace(namespace_name, tools) -%}
109
+ {{- "## " + namespace_name + "\n\n" }}
110
+ {{- "namespace " + namespace_name + " {\n\n" }}
111
+ {%- for tool in tools %}
112
+ {%- set tool = tool.function %}
113
+ {{- "// " + tool.description + "\n" }}
114
+ {{- "type "+ tool.name + " = " }}
115
+ {%- if tool.parameters and tool.parameters.properties %}
116
+ {{- "(_: {\n" }}
117
+ {%- for param_name, param_spec in tool.parameters.properties.items() %}
118
+ {%- if param_spec.description %}
119
+ {{- "// " + param_spec.description + "\n" }}
120
+ {%- endif %}
121
+ {{- param_name }}
122
+ {%- if param_name not in (tool.parameters.required or []) -%}
123
+ {{- "?" }}
124
+ {%- endif -%}
125
+ {{- ": " }}
126
+ {{- render_typescript_type(param_spec, tool.parameters.required or []) }}
127
+ {%- if param_spec.default is defined -%}
128
+ {%- if param_spec.enum %}
129
+ {{- ", // default: " + param_spec.default }}
130
+ {%- elif param_spec.oneOf %}
131
+ {{- "// default: " + param_spec.default }}
132
+ {%- else %}
133
+ {{- ", // default: " + param_spec.default|tojson }}
134
+ {%- endif -%}
135
+ {%- endif -%}
136
+ {%- if not loop.last %}
137
+ {{- ",\n" }}
138
+ {%- else %}
139
+ {{- ",\n" }}
140
+ {%- endif -%}
141
+ {%- endfor %}
142
+ {{- "}) => any;\n\n" }}
143
+ {%- else -%}
144
+ {{- "() => any;\n\n" }}
145
+ {%- endif -%}
146
+ {%- endfor %}
147
+ {{- "} // namespace " + namespace_name }}
148
+ {%- endmacro -%}
149
+
150
+ {%- macro render_builtin_tools(browser_tool, python_tool) -%}
151
+ {%- if browser_tool %}
152
+ {{- "## browser\n\n" }}
153
+ {{- "// Tool for browsing.\n" }}
154
+ {{- "// The `cursor` appears in brackets before each browsing display: `[{cursor}]`.\n" }}
155
+ {{- "// Cite information from the tool using the following format:\n" }}
156
+ {{- "// `【{cursor}†L{line_start}(-L{line_end})?】`, for example: `【6†L9-L11】` or `【8†L3】`.\n" }}
157
+ {{- "// Do not quote more than 10 words directly from the tool output.\n" }}
158
+ {{- "// sources=web (default: web)\n" }}
159
+ {{- "namespace browser {\n\n" }}
160
+ {{- "// Searches for information related to `query` and displays `topn` results.\n" }}
161
+ {{- "type search = (_: {\n" }}
162
+ {{- "query: string,\n" }}
163
+ {{- "topn?: number, // default: 10\n" }}
164
+ {{- "source?: string,\n" }}
165
+ {{- "}) => any;\n\n" }}
166
+ {{- "// Opens the link `id` from the page indicated by `cursor` starting at line number `loc`, showing `num_lines` lines.\n" }}
167
+ {{- "// Valid link ids are displayed with the formatting: `【{id}†.*】`.\n" }}
168
+ {{- "// If `cursor` is not provided, the most recent page is implied.\n" }}
169
+ {{- "// If `id` is a string, it is treated as a fully qualified URL associated with `source`.\n" }}
170
+ {{- "// If `loc` is not provided, the viewport will be positioned at the beginning of the document or centered on the most relevant passage, if available.\n" }}
171
+ {{- "// Use this function without `id` to scroll to a new location of an opened page.\n" }}
172
+ {{- "type open = (_: {\n" }}
173
+ {{- "id?: number | string, // default: -1\n" }}
174
+ {{- "cursor?: number, // default: -1\n" }}
175
+ {{- "loc?: number, // default: -1\n" }}
176
+ {{- "num_lines?: number, // default: -1\n" }}
177
+ {{- "view_source?: boolean, // default: false\n" }}
178
+ {{- "source?: string,\n" }}
179
+ {{- "}) => any;\n\n" }}
180
+ {{- "// Finds exact matches of `pattern` in the current page, or the page given by `cursor`.\n" }}
181
+ {{- "type find = (_: {\n" }}
182
+ {{- "pattern: string,\n" }}
183
+ {{- "cursor?: number, // default: -1\n" }}
184
+ {{- "}) => any;\n\n" }}
185
+ {{- "} // namespace browser\n\n" }}
186
+ {%- endif -%}
187
+
188
+ {%- if python_tool %}
189
+ {{- "## python\n\n" }}
190
+ {{- "Use this tool to execute Python code in your chain of thought. The code will not be shown to the user. This tool should be used for internal reasoning, but not for code that is intended to be visible to the user (e.g. when creating plots, tables, or files).\n\n" }}
191
+ {{- "When you send a message containing Python code to python, it will be executed in a stateful Jupyter notebook environment. python will respond with the output of the execution or time out after 120.0 seconds. The drive at '/mnt/data' can be used to save and persist user files. Internet access for this session is UNKNOWN. Depends on the cluster.\n\n" }}
192
+ {%- endif -%}
193
+ {%- endmacro -%}
194
+
195
+ {#- System Message Construction ============================================ #}
196
+ {%- macro build_system_message() -%}
197
+ {%- if model_identity is not defined %}
198
+ {%- set model_identity = "You are ChatGPT, a large language model trained by OpenAI." %}
199
+ {%- endif %}
200
+ {{- model_identity + "\n" }}
201
+ {{- "Knowledge cutoff: 2024-06\n" }}
202
+ {{- "Current date: " + strftime_now("%Y-%m-%d") + "\n\n" }}
203
+ {%- if reasoning_effort is not defined %}
204
+ {%- set reasoning_effort = "medium" %}
205
+ {%- endif %}
206
+ {{- "Reasoning: " + reasoning_effort + "\n\n" }}
207
+ {%- if builtin_tools %}
208
+ {{- "# Tools\n\n" }}
209
+ {%- set available_builtin_tools = namespace(browser=false, python=false) %}
210
+ {%- for tool in builtin_tools %}
211
+ {%- if tool == "browser" %}
212
+ {%- set available_builtin_tools.browser = true %}
213
+ {%- elif tool == "python" %}
214
+ {%- set available_builtin_tools.python = true %}
215
+ {%- endif %}
216
+ {%- endfor %}
217
+ {{- render_builtin_tools(available_builtin_tools.browser, available_builtin_tools.python) }}
218
+ {%- endif -%}
219
+ {{- "# Valid channels: analysis, commentary, final. Channel must be included for every message." }}
220
+ {%- if tools -%}
221
+ {{- "\nCalls to these tools must go to the commentary channel: 'functions'." }}
222
+ {%- endif -%}
223
+ {%- endmacro -%}
224
+
225
+ {#- Main Template Logic ================================================= #}
226
+ {#- Set defaults #}
227
+
228
+ {#- Render system message #}
229
+ {{- "<|start|>system<|message|>" }}
230
+ {{- build_system_message() }}
231
+ {{- "<|end|>" }}
232
+
233
+ {#- Extract developer message #}
234
+ {%- if messages[0].role == "developer" or messages[0].role == "system" %}
235
+ {%- set developer_message = messages[0].content %}
236
+ {%- set loop_messages = messages[1:] %}
237
+ {%- else %}
238
+ {%- set developer_message = "" %}
239
+ {%- set loop_messages = messages %}
240
+ {%- endif %}
241
+
242
+ {#- Render developer message #}
243
+ {%- if developer_message or tools %}
244
+ {{- "<|start|>developer<|message|>" }}
245
+ {%- if developer_message %}
246
+ {{- "# Instructions\n\n" }}
247
+ {{- developer_message }}
248
+ {{- "\n\n" }}
249
+ {%- endif %}
250
+ {%- if tools -%}
251
+ {{- "# Tools\n\n" }}
252
+ {{- render_tool_namespace("functions", tools) }}
253
+ {%- endif -%}
254
+ {{- "<|end|>" }}
255
+ {%- endif %}
256
+
257
+ {#- Render messages #}
258
+ {%- set last_tool_call = namespace(name=none) %}
259
+ {%- for message in loop_messages -%}
260
+ {#- At this point only assistant/user/tool messages should remain #}
261
+ {%- if message.role == 'assistant' -%}
262
+ {#- Checks to ensure the messages are being passed in the format we expect #}
263
+ {%- if "content" in message %}
264
+ {%- if "<|channel|>analysis<|message|>" in message.content or "<|channel|>final<|message|>" in message.content %}
265
+ {{- raise_exception("You have passed a message containing <|channel|> tags in the content field. Instead of doing this, you should pass analysis messages (the string between '<|message|>' and '<|end|>') in the 'thinking' field, and final messages (the string between '<|message|>' and '<|end|>') in the 'content' field.") }}
266
+ {%- endif %}
267
+ {%- endif %}
268
+ {%- if "thinking" in message %}
269
+ {%- if "<|channel|>analysis<|message|>" in message.thinking or "<|channel|>final<|message|>" in message.thinking %}
270
+ {{- raise_exception("You have passed a message containing <|channel|> tags in the thinking field. Instead of doing this, you should pass analysis messages (the string between '<|message|>' and '<|end|>') in the 'thinking' field, and final messages (the string between '<|message|>' and '<|end|>') in the 'content' field.") }}
271
+ {%- endif %}
272
+ {%- endif %}
273
+ {%- if "tool_calls" in message %}
274
+ {#- We need very careful handling here - we want to drop the tool call analysis message if the model #}
275
+ {#- has output a later <|final|> message, but otherwise we want to retain it. This is the only case #}
276
+ {#- when we render CoT/analysis messages in inference. #}
277
+ {%- set future_final_message = namespace(found=false) %}
278
+ {%- for future_message in loop_messages[loop.index:] %}
279
+ {%- if future_message.role == 'assistant' and "tool_calls" not in future_message %}
280
+ {%- set future_final_message.found = true %}
281
+ {%- endif %}
282
+ {%- endfor %}
283
+ {#- We assume max 1 tool call per message, and so we infer the tool call name #}
284
+ {#- in "tool" messages from the most recent assistant tool call name #}
285
+ {%- set tool_call = message.tool_calls[0] %}
286
+ {%- if tool_call.function %}
287
+ {%- set tool_call = tool_call.function %}
288
+ {%- endif %}
289
+ {%- if message.content and message.thinking %}
290
+ {{- raise_exception("Cannot pass both content and thinking in an assistant message with tool calls! Put the analysis message in one or the other, but not both.") }}
291
+ {%- elif message.content and not future_final_message.found %}
292
+ {{- "<|start|>assistant<|channel|>analysis<|message|>" + message.content + "<|end|>" }}
293
+ {%- elif message.thinking and not future_final_message.found %}
294
+ {{- "<|start|>assistant<|channel|>analysis<|message|>" + message.thinking + "<|end|>" }}
295
+ {%- endif %}
296
+ {{- "<|start|>assistant to=" }}
297
+ {{- "functions." + tool_call.name + "<|channel|>commentary " }}
298
+ {{- (tool_call.content_type if tool_call.content_type is defined else "json") + "<|message|>" }}
299
+ {{- tool_call.arguments|tojson }}
300
+ {{- "<|call|>" }}
301
+ {%- set last_tool_call.name = tool_call.name %}
302
+ {%- elif loop.last and not add_generation_prompt %}
303
+ {#- Only render the CoT if the final turn is an assistant turn and add_generation_prompt is false #}
304
+ {#- This is a situation that should only occur in training, never in inference. #}
305
+ {%- if "thinking" in message %}
306
+ {{- "<|start|>assistant<|channel|>analysis<|message|>" + message.thinking + "<|end|>" }}
307
+ {%- endif %}
308
+ {#- <|return|> indicates the end of generation, but <|end|> does not #}
309
+ {#- <|return|> should never be an input to the model, but we include it as the final token #}
310
+ {#- when training, so the model learns to emit it. #}
311
+ {{- "<|start|>assistant<|channel|>final<|message|>" + message.content + "<|return|>" }}
312
+ {%- else %}
313
+ {#- CoT is dropped during all previous turns, so we never render it for inference #}
314
+ {{- "<|start|>assistant<|channel|>final<|message|>" + message.content + "<|end|>" }}
315
+ {%- set last_tool_call.name = none %}
316
+ {%- endif %}
317
+ {%- elif message.role == 'tool' -%}
318
+ {%- if last_tool_call.name is none %}
319
+ {{- raise_exception("Message has tool role, but there was no previous assistant message with a tool call!") }}
320
+ {%- endif %}
321
+ {{- "<|start|>functions." + last_tool_call.name }}
322
+ {{- " to=assistant<|channel|>commentary<|message|>" + message.content|tojson + "<|end|>" }}
323
+ {%- elif message.role == 'user' -%}
324
+ {{- "<|start|>user<|message|>" + message.content + "<|end|>" }}
325
+ {%- endif -%}
326
+ {%- endfor -%}
327
+
328
+ {#- Generation prompt #}
329
+ {%- if add_generation_prompt -%}
330
+ <|start|>assistant
331
+ {%- endif -%}
checkpoint-1250/chat_template.jinja ADDED
@@ -0,0 +1,331 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {#-
2
+ In addition to the normal inputs of `messages` and `tools`, this template also accepts the
3
+ following kwargs:
4
+ - "builtin_tools": A list, can contain "browser" and/or "python".
5
+ - "model_identity": A string that optionally describes the model identity.
6
+ - "reasoning_effort": A string that describes the reasoning effort, defaults to "medium".
7
+ #}
8
+
9
+ {#- Tool Definition Rendering ============================================== #}
10
+ {%- macro render_typescript_type(param_spec, required_params, is_nullable=false) -%}
11
+ {%- if param_spec.type == "array" -%}
12
+ {%- if param_spec['items'] -%}
13
+ {%- if param_spec['items']['type'] == "string" -%}
14
+ {{- "string[]" }}
15
+ {%- elif param_spec['items']['type'] == "number" -%}
16
+ {{- "number[]" }}
17
+ {%- elif param_spec['items']['type'] == "integer" -%}
18
+ {{- "number[]" }}
19
+ {%- elif param_spec['items']['type'] == "boolean" -%}
20
+ {{- "boolean[]" }}
21
+ {%- else -%}
22
+ {%- set inner_type = render_typescript_type(param_spec['items'], required_params) -%}
23
+ {%- if inner_type == "object | object" or inner_type|length > 50 -%}
24
+ {{- "any[]" }}
25
+ {%- else -%}
26
+ {{- inner_type + "[]" }}
27
+ {%- endif -%}
28
+ {%- endif -%}
29
+ {%- if param_spec.nullable -%}
30
+ {{- " | null" }}
31
+ {%- endif -%}
32
+ {%- else -%}
33
+ {{- "any[]" }}
34
+ {%- if param_spec.nullable -%}
35
+ {{- " | null" }}
36
+ {%- endif -%}
37
+ {%- endif -%}
38
+ {%- elif param_spec.type is defined and param_spec.type is iterable and param_spec.type is not string and param_spec.type is not mapping and param_spec.type[0] is defined -%}
39
+ {#- Handle array of types like ["object", "object"] from Union[dict, list] #}
40
+ {%- if param_spec.type | length > 1 -%}
41
+ {{- param_spec.type | join(" | ") }}
42
+ {%- else -%}
43
+ {{- param_spec.type[0] }}
44
+ {%- endif -%}
45
+ {%- elif param_spec.oneOf -%}
46
+ {#- Handle oneOf schemas - check for complex unions and fallback to any #}
47
+ {%- set has_object_variants = false -%}
48
+ {%- for variant in param_spec.oneOf -%}
49
+ {%- if variant.type == "object" -%}
50
+ {%- set has_object_variants = true -%}
51
+ {%- endif -%}
52
+ {%- endfor -%}
53
+ {%- if has_object_variants and param_spec.oneOf|length > 1 -%}
54
+ {{- "any" }}
55
+ {%- else -%}
56
+ {%- for variant in param_spec.oneOf -%}
57
+ {{- render_typescript_type(variant, required_params) -}}
58
+ {%- if variant.description %}
59
+ {{- "// " + variant.description }}
60
+ {%- endif -%}
61
+ {%- if variant.default is defined %}
62
+ {{ "// default: " + variant.default|tojson }}
63
+ {%- endif -%}
64
+ {%- if not loop.last %}
65
+ {{- " | " }}
66
+ {% endif -%}
67
+ {%- endfor -%}
68
+ {%- endif -%}
69
+ {%- elif param_spec.type == "string" -%}
70
+ {%- if param_spec.enum -%}
71
+ {{- '"' + param_spec.enum|join('" | "') + '"' -}}
72
+ {%- else -%}
73
+ {{- "string" }}
74
+ {%- if param_spec.nullable %}
75
+ {{- " | null" }}
76
+ {%- endif -%}
77
+ {%- endif -%}
78
+ {%- elif param_spec.type == "number" -%}
79
+ {{- "number" }}
80
+ {%- elif param_spec.type == "integer" -%}
81
+ {{- "number" }}
82
+ {%- elif param_spec.type == "boolean" -%}
83
+ {{- "boolean" }}
84
+
85
+ {%- elif param_spec.type == "object" -%}
86
+ {%- if param_spec.properties -%}
87
+ {{- "{\n" }}
88
+ {%- for prop_name, prop_spec in param_spec.properties.items() -%}
89
+ {{- prop_name -}}
90
+ {%- if prop_name not in (param_spec.required or []) -%}
91
+ {{- "?" }}
92
+ {%- endif -%}
93
+ {{- ": " }}
94
+ {{ render_typescript_type(prop_spec, param_spec.required or []) }}
95
+ {%- if not loop.last -%}
96
+ {{-", " }}
97
+ {%- endif -%}
98
+ {%- endfor -%}
99
+ {{- "}" }}
100
+ {%- else -%}
101
+ {{- "object" }}
102
+ {%- endif -%}
103
+ {%- else -%}
104
+ {{- "any" }}
105
+ {%- endif -%}
106
+ {%- endmacro -%}
107
+
108
+ {%- macro render_tool_namespace(namespace_name, tools) -%}
109
+ {{- "## " + namespace_name + "\n\n" }}
110
+ {{- "namespace " + namespace_name + " {\n\n" }}
111
+ {%- for tool in tools %}
112
+ {%- set tool = tool.function %}
113
+ {{- "// " + tool.description + "\n" }}
114
+ {{- "type "+ tool.name + " = " }}
115
+ {%- if tool.parameters and tool.parameters.properties %}
116
+ {{- "(_: {\n" }}
117
+ {%- for param_name, param_spec in tool.parameters.properties.items() %}
118
+ {%- if param_spec.description %}
119
+ {{- "// " + param_spec.description + "\n" }}
120
+ {%- endif %}
121
+ {{- param_name }}
122
+ {%- if param_name not in (tool.parameters.required or []) -%}
123
+ {{- "?" }}
124
+ {%- endif -%}
125
+ {{- ": " }}
126
+ {{- render_typescript_type(param_spec, tool.parameters.required or []) }}
127
+ {%- if param_spec.default is defined -%}
128
+ {%- if param_spec.enum %}
129
+ {{- ", // default: " + param_spec.default }}
130
+ {%- elif param_spec.oneOf %}
131
+ {{- "// default: " + param_spec.default }}
132
+ {%- else %}
133
+ {{- ", // default: " + param_spec.default|tojson }}
134
+ {%- endif -%}
135
+ {%- endif -%}
136
+ {%- if not loop.last %}
137
+ {{- ",\n" }}
138
+ {%- else %}
139
+ {{- ",\n" }}
140
+ {%- endif -%}
141
+ {%- endfor %}
142
+ {{- "}) => any;\n\n" }}
143
+ {%- else -%}
144
+ {{- "() => any;\n\n" }}
145
+ {%- endif -%}
146
+ {%- endfor %}
147
+ {{- "} // namespace " + namespace_name }}
148
+ {%- endmacro -%}
149
+
150
+ {%- macro render_builtin_tools(browser_tool, python_tool) -%}
151
+ {%- if browser_tool %}
152
+ {{- "## browser\n\n" }}
153
+ {{- "// Tool for browsing.\n" }}
154
+ {{- "// The `cursor` appears in brackets before each browsing display: `[{cursor}]`.\n" }}
155
+ {{- "// Cite information from the tool using the following format:\n" }}
156
+ {{- "// `【{cursor}†L{line_start}(-L{line_end})?】`, for example: `【6†L9-L11】` or `【8†L3】`.\n" }}
157
+ {{- "// Do not quote more than 10 words directly from the tool output.\n" }}
158
+ {{- "// sources=web (default: web)\n" }}
159
+ {{- "namespace browser {\n\n" }}
160
+ {{- "// Searches for information related to `query` and displays `topn` results.\n" }}
161
+ {{- "type search = (_: {\n" }}
162
+ {{- "query: string,\n" }}
163
+ {{- "topn?: number, // default: 10\n" }}
164
+ {{- "source?: string,\n" }}
165
+ {{- "}) => any;\n\n" }}
166
+ {{- "// Opens the link `id` from the page indicated by `cursor` starting at line number `loc`, showing `num_lines` lines.\n" }}
167
+ {{- "// Valid link ids are displayed with the formatting: `【{id}†.*】`.\n" }}
168
+ {{- "// If `cursor` is not provided, the most recent page is implied.\n" }}
169
+ {{- "// If `id` is a string, it is treated as a fully qualified URL associated with `source`.\n" }}
170
+ {{- "// If `loc` is not provided, the viewport will be positioned at the beginning of the document or centered on the most relevant passage, if available.\n" }}
171
+ {{- "// Use this function without `id` to scroll to a new location of an opened page.\n" }}
172
+ {{- "type open = (_: {\n" }}
173
+ {{- "id?: number | string, // default: -1\n" }}
174
+ {{- "cursor?: number, // default: -1\n" }}
175
+ {{- "loc?: number, // default: -1\n" }}
176
+ {{- "num_lines?: number, // default: -1\n" }}
177
+ {{- "view_source?: boolean, // default: false\n" }}
178
+ {{- "source?: string,\n" }}
179
+ {{- "}) => any;\n\n" }}
180
+ {{- "// Finds exact matches of `pattern` in the current page, or the page given by `cursor`.\n" }}
181
+ {{- "type find = (_: {\n" }}
182
+ {{- "pattern: string,\n" }}
183
+ {{- "cursor?: number, // default: -1\n" }}
184
+ {{- "}) => any;\n\n" }}
185
+ {{- "} // namespace browser\n\n" }}
186
+ {%- endif -%}
187
+
188
+ {%- if python_tool %}
189
+ {{- "## python\n\n" }}
190
+ {{- "Use this tool to execute Python code in your chain of thought. The code will not be shown to the user. This tool should be used for internal reasoning, but not for code that is intended to be visible to the user (e.g. when creating plots, tables, or files).\n\n" }}
191
+ {{- "When you send a message containing Python code to python, it will be executed in a stateful Jupyter notebook environment. python will respond with the output of the execution or time out after 120.0 seconds. The drive at '/mnt/data' can be used to save and persist user files. Internet access for this session is UNKNOWN. Depends on the cluster.\n\n" }}
192
+ {%- endif -%}
193
+ {%- endmacro -%}
194
+
195
+ {#- System Message Construction ============================================ #}
196
+ {%- macro build_system_message() -%}
197
+ {%- if model_identity is not defined %}
198
+ {%- set model_identity = "You are ChatGPT, a large language model trained by OpenAI." %}
199
+ {%- endif %}
200
+ {{- model_identity + "\n" }}
201
+ {{- "Knowledge cutoff: 2024-06\n" }}
202
+ {{- "Current date: " + strftime_now("%Y-%m-%d") + "\n\n" }}
203
+ {%- if reasoning_effort is not defined %}
204
+ {%- set reasoning_effort = "medium" %}
205
+ {%- endif %}
206
+ {{- "Reasoning: " + reasoning_effort + "\n\n" }}
207
+ {%- if builtin_tools %}
208
+ {{- "# Tools\n\n" }}
209
+ {%- set available_builtin_tools = namespace(browser=false, python=false) %}
210
+ {%- for tool in builtin_tools %}
211
+ {%- if tool == "browser" %}
212
+ {%- set available_builtin_tools.browser = true %}
213
+ {%- elif tool == "python" %}
214
+ {%- set available_builtin_tools.python = true %}
215
+ {%- endif %}
216
+ {%- endfor %}
217
+ {{- render_builtin_tools(available_builtin_tools.browser, available_builtin_tools.python) }}
218
+ {%- endif -%}
219
+ {{- "# Valid channels: analysis, commentary, final. Channel must be included for every message." }}
220
+ {%- if tools -%}
221
+ {{- "\nCalls to these tools must go to the commentary channel: 'functions'." }}
222
+ {%- endif -%}
223
+ {%- endmacro -%}
224
+
225
+ {#- Main Template Logic ================================================= #}
226
+ {#- Set defaults #}
227
+
228
+ {#- Render system message #}
229
+ {{- "<|start|>system<|message|>" }}
230
+ {{- build_system_message() }}
231
+ {{- "<|end|>" }}
232
+
233
+ {#- Extract developer message #}
234
+ {%- if messages[0].role == "developer" or messages[0].role == "system" %}
235
+ {%- set developer_message = messages[0].content %}
236
+ {%- set loop_messages = messages[1:] %}
237
+ {%- else %}
238
+ {%- set developer_message = "" %}
239
+ {%- set loop_messages = messages %}
240
+ {%- endif %}
241
+
242
+ {#- Render developer message #}
243
+ {%- if developer_message or tools %}
244
+ {{- "<|start|>developer<|message|>" }}
245
+ {%- if developer_message %}
246
+ {{- "# Instructions\n\n" }}
247
+ {{- developer_message }}
248
+ {{- "\n\n" }}
249
+ {%- endif %}
250
+ {%- if tools -%}
251
+ {{- "# Tools\n\n" }}
252
+ {{- render_tool_namespace("functions", tools) }}
253
+ {%- endif -%}
254
+ {{- "<|end|>" }}
255
+ {%- endif %}
256
+
257
+ {#- Render messages #}
258
+ {%- set last_tool_call = namespace(name=none) %}
259
+ {%- for message in loop_messages -%}
260
+ {#- At this point only assistant/user/tool messages should remain #}
261
+ {%- if message.role == 'assistant' -%}
262
+ {#- Checks to ensure the messages are being passed in the format we expect #}
263
+ {%- if "content" in message %}
264
+ {%- if "<|channel|>analysis<|message|>" in message.content or "<|channel|>final<|message|>" in message.content %}
265
+ {{- raise_exception("You have passed a message containing <|channel|> tags in the content field. Instead of doing this, you should pass analysis messages (the string between '<|message|>' and '<|end|>') in the 'thinking' field, and final messages (the string between '<|message|>' and '<|end|>') in the 'content' field.") }}
266
+ {%- endif %}
267
+ {%- endif %}
268
+ {%- if "thinking" in message %}
269
+ {%- if "<|channel|>analysis<|message|>" in message.thinking or "<|channel|>final<|message|>" in message.thinking %}
270
+ {{- raise_exception("You have passed a message containing <|channel|> tags in the thinking field. Instead of doing this, you should pass analysis messages (the string between '<|message|>' and '<|end|>') in the 'thinking' field, and final messages (the string between '<|message|>' and '<|end|>') in the 'content' field.") }}
271
+ {%- endif %}
272
+ {%- endif %}
273
+ {%- if "tool_calls" in message %}
274
+ {#- We need very careful handling here - we want to drop the tool call analysis message if the model #}
275
+ {#- has output a later <|final|> message, but otherwise we want to retain it. This is the only case #}
276
+ {#- when we render CoT/analysis messages in inference. #}
277
+ {%- set future_final_message = namespace(found=false) %}
278
+ {%- for future_message in loop_messages[loop.index:] %}
279
+ {%- if future_message.role == 'assistant' and "tool_calls" not in future_message %}
280
+ {%- set future_final_message.found = true %}
281
+ {%- endif %}
282
+ {%- endfor %}
283
+ {#- We assume max 1 tool call per message, and so we infer the tool call name #}
284
+ {#- in "tool" messages from the most recent assistant tool call name #}
285
+ {%- set tool_call = message.tool_calls[0] %}
286
+ {%- if tool_call.function %}
287
+ {%- set tool_call = tool_call.function %}
288
+ {%- endif %}
289
+ {%- if message.content and message.thinking %}
290
+ {{- raise_exception("Cannot pass both content and thinking in an assistant message with tool calls! Put the analysis message in one or the other, but not both.") }}
291
+ {%- elif message.content and not future_final_message.found %}
292
+ {{- "<|start|>assistant<|channel|>analysis<|message|>" + message.content + "<|end|>" }}
293
+ {%- elif message.thinking and not future_final_message.found %}
294
+ {{- "<|start|>assistant<|channel|>analysis<|message|>" + message.thinking + "<|end|>" }}
295
+ {%- endif %}
296
+ {{- "<|start|>assistant to=" }}
297
+ {{- "functions." + tool_call.name + "<|channel|>commentary " }}
298
+ {{- (tool_call.content_type if tool_call.content_type is defined else "json") + "<|message|>" }}
299
+ {{- tool_call.arguments|tojson }}
300
+ {{- "<|call|>" }}
301
+ {%- set last_tool_call.name = tool_call.name %}
302
+ {%- elif loop.last and not add_generation_prompt %}
303
+ {#- Only render the CoT if the final turn is an assistant turn and add_generation_prompt is false #}
304
+ {#- This is a situation that should only occur in training, never in inference. #}
305
+ {%- if "thinking" in message %}
306
+ {{- "<|start|>assistant<|channel|>analysis<|message|>" + message.thinking + "<|end|>" }}
307
+ {%- endif %}
308
+ {#- <|return|> indicates the end of generation, but <|end|> does not #}
309
+ {#- <|return|> should never be an input to the model, but we include it as the final token #}
310
+ {#- when training, so the model learns to emit it. #}
311
+ {{- "<|start|>assistant<|channel|>final<|message|>" + message.content + "<|return|>" }}
312
+ {%- else %}
313
+ {#- CoT is dropped during all previous turns, so we never render it for inference #}
314
+ {{- "<|start|>assistant<|channel|>final<|message|>" + message.content + "<|end|>" }}
315
+ {%- set last_tool_call.name = none %}
316
+ {%- endif %}
317
+ {%- elif message.role == 'tool' -%}
318
+ {%- if last_tool_call.name is none %}
319
+ {{- raise_exception("Message has tool role, but there was no previous assistant message with a tool call!") }}
320
+ {%- endif %}
321
+ {{- "<|start|>functions." + last_tool_call.name }}
322
+ {{- " to=assistant<|channel|>commentary<|message|>" + message.content|tojson + "<|end|>" }}
323
+ {%- elif message.role == 'user' -%}
324
+ {{- "<|start|>user<|message|>" + message.content + "<|end|>" }}
325
+ {%- endif -%}
326
+ {%- endfor -%}
327
+
328
+ {#- Generation prompt #}
329
+ {%- if add_generation_prompt -%}
330
+ <|start|>assistant
331
+ {%- endif -%}
checkpoint-1250/config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "GptOssForCausalLM"
4
+ ],
5
+ "attention_bias": true,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 199998,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 200002,
10
+ "experts_per_token": 4,
11
+ "head_dim": 64,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 2880,
14
+ "initial_context_length": 4096,
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 2880,
17
+ "layer_types": [
18
+ "sliding_attention",
19
+ "full_attention",
20
+ "sliding_attention",
21
+ "full_attention",
22
+ "sliding_attention",
23
+ "full_attention",
24
+ "sliding_attention",
25
+ "full_attention",
26
+ "sliding_attention",
27
+ "full_attention",
28
+ "sliding_attention",
29
+ "full_attention",
30
+ "sliding_attention",
31
+ "full_attention",
32
+ "sliding_attention",
33
+ "full_attention",
34
+ "sliding_attention",
35
+ "full_attention",
36
+ "sliding_attention",
37
+ "full_attention",
38
+ "sliding_attention",
39
+ "full_attention",
40
+ "sliding_attention",
41
+ "full_attention"
42
+ ],
43
+ "max_position_embeddings": 131072,
44
+ "model_type": "gpt_oss",
45
+ "num_attention_heads": 64,
46
+ "num_experts_per_tok": 4,
47
+ "num_hidden_layers": 24,
48
+ "num_key_value_heads": 8,
49
+ "num_local_experts": 32,
50
+ "output_router_logits": false,
51
+ "pad_token_id": 199999,
52
+ "rms_norm_eps": 1e-05,
53
+ "rope_scaling": {
54
+ "beta_fast": 32.0,
55
+ "beta_slow": 1.0,
56
+ "factor": 32.0,
57
+ "original_max_position_embeddings": 4096,
58
+ "rope_type": "yarn",
59
+ "truncate": false
60
+ },
61
+ "rope_theta": 150000,
62
+ "router_aux_loss_coef": 0.9,
63
+ "sliding_window": 128,
64
+ "swiglu_limit": 7.0,
65
+ "tie_word_embeddings": false,
66
+ "transformers_version": "4.57.1",
67
+ "use_cache": false,
68
+ "vocab_size": 201088
69
+ }
checkpoint-1250/generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 199998,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 200002,
6
+ 199999,
7
+ 200012
8
+ ],
9
+ "pad_token_id": 199999,
10
+ "transformers_version": "4.57.1"
11
+ }
checkpoint-1250/model-00001-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc729189133161b61e287dd7bdec06afc2b7f77ed1820ca53eda0553fa742230
3
+ size 4504304664
checkpoint-1250/model-00002-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4708b3da84cd7d559f5cd2eb9d0d13466ae1ccd90fdee8295d9ec3d8748582c
3
+ size 4939127656
checkpoint-1250/model-00003-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96031fa4efe1e880b2f17ac4352c3a2d251129d8140d122b9e8649a2d45c51d2
3
+ size 4939127656
checkpoint-1250/model-00004-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46e221c089ddcbacd6f4495a2dfccea2206cb8b21795ea868415125e9fb8139e
3
+ size 4939127680
checkpoint-1250/model-00005-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f223ae981c1c8f4925334b8e0954ce78a613b0602d3eba28ea73f0146c9d73f3
3
+ size 4939127704
checkpoint-1250/model-00006-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf0e3a0ccf931369e3ebe77f1f6d0a1781d739d6cb1dfed7c54fedda3f94fe18
3
+ size 4939127704
checkpoint-1250/model-00007-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51a3ad356e1a0d63de63f78975caf1a129d7d29bc00a7614de652aff7a4c57b0
3
+ size 4939127704
checkpoint-1250/model-00008-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07a2d79388f1f6f1bebac3d03216b6ac47c2bf498640ab54ac8c15beb629847f
3
+ size 4939127704
checkpoint-1250/model-00009-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eda29cdc86d75937066cde98574795f08a96ebe34b22e126e1cb252ec494df4
3
+ size 2751362856
checkpoint-1250/model.safetensors.index.json ADDED
@@ -0,0 +1,419 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_parameters": 20914757184,
4
+ "total_size": 41829514368
5
+ },
6
+ "weight_map": {
7
+ "lm_head.weight": "model-00009-of-00009.safetensors",
8
+ "model.embed_tokens.weight": "model-00001-of-00009.safetensors",
9
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00009.safetensors",
10
+ "model.layers.0.mlp.experts.down_proj": "model-00001-of-00009.safetensors",
11
+ "model.layers.0.mlp.experts.down_proj_bias": "model-00001-of-00009.safetensors",
12
+ "model.layers.0.mlp.experts.gate_up_proj": "model-00001-of-00009.safetensors",
13
+ "model.layers.0.mlp.experts.gate_up_proj_bias": "model-00001-of-00009.safetensors",
14
+ "model.layers.0.mlp.router.bias": "model-00001-of-00009.safetensors",
15
+ "model.layers.0.mlp.router.weight": "model-00001-of-00009.safetensors",
16
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00009.safetensors",
17
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00009.safetensors",
18
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00009.safetensors",
19
+ "model.layers.0.self_attn.o_proj.bias": "model-00001-of-00009.safetensors",
20
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00009.safetensors",
21
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00009.safetensors",
22
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00009.safetensors",
23
+ "model.layers.0.self_attn.sinks": "model-00001-of-00009.safetensors",
24
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00009.safetensors",
25
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00009.safetensors",
26
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00009.safetensors",
27
+ "model.layers.1.mlp.experts.down_proj": "model-00001-of-00009.safetensors",
28
+ "model.layers.1.mlp.experts.down_proj_bias": "model-00001-of-00009.safetensors",
29
+ "model.layers.1.mlp.experts.gate_up_proj": "model-00001-of-00009.safetensors",
30
+ "model.layers.1.mlp.experts.gate_up_proj_bias": "model-00001-of-00009.safetensors",
31
+ "model.layers.1.mlp.router.bias": "model-00001-of-00009.safetensors",
32
+ "model.layers.1.mlp.router.weight": "model-00001-of-00009.safetensors",
33
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00009.safetensors",
34
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00009.safetensors",
35
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00009.safetensors",
36
+ "model.layers.1.self_attn.o_proj.bias": "model-00001-of-00009.safetensors",
37
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00009.safetensors",
38
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00009.safetensors",
39
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00009.safetensors",
40
+ "model.layers.1.self_attn.sinks": "model-00001-of-00009.safetensors",
41
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00009.safetensors",
42
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00009.safetensors",
43
+ "model.layers.10.input_layernorm.weight": "model-00004-of-00009.safetensors",
44
+ "model.layers.10.mlp.experts.down_proj": "model-00004-of-00009.safetensors",
45
+ "model.layers.10.mlp.experts.down_proj_bias": "model-00004-of-00009.safetensors",
46
+ "model.layers.10.mlp.experts.gate_up_proj": "model-00004-of-00009.safetensors",
47
+ "model.layers.10.mlp.experts.gate_up_proj_bias": "model-00004-of-00009.safetensors",
48
+ "model.layers.10.mlp.router.bias": "model-00004-of-00009.safetensors",
49
+ "model.layers.10.mlp.router.weight": "model-00004-of-00009.safetensors",
50
+ "model.layers.10.post_attention_layernorm.weight": "model-00004-of-00009.safetensors",
51
+ "model.layers.10.self_attn.k_proj.bias": "model-00004-of-00009.safetensors",
52
+ "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00009.safetensors",
53
+ "model.layers.10.self_attn.o_proj.bias": "model-00004-of-00009.safetensors",
54
+ "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00009.safetensors",
55
+ "model.layers.10.self_attn.q_proj.bias": "model-00004-of-00009.safetensors",
56
+ "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00009.safetensors",
57
+ "model.layers.10.self_attn.sinks": "model-00004-of-00009.safetensors",
58
+ "model.layers.10.self_attn.v_proj.bias": "model-00004-of-00009.safetensors",
59
+ "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00009.safetensors",
60
+ "model.layers.11.input_layernorm.weight": "model-00005-of-00009.safetensors",
61
+ "model.layers.11.mlp.experts.down_proj": "model-00005-of-00009.safetensors",
62
+ "model.layers.11.mlp.experts.down_proj_bias": "model-00005-of-00009.safetensors",
63
+ "model.layers.11.mlp.experts.gate_up_proj": "model-00005-of-00009.safetensors",
64
+ "model.layers.11.mlp.experts.gate_up_proj_bias": "model-00005-of-00009.safetensors",
65
+ "model.layers.11.mlp.router.bias": "model-00004-of-00009.safetensors",
66
+ "model.layers.11.mlp.router.weight": "model-00004-of-00009.safetensors",
67
+ "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00009.safetensors",
68
+ "model.layers.11.self_attn.k_proj.bias": "model-00004-of-00009.safetensors",
69
+ "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00009.safetensors",
70
+ "model.layers.11.self_attn.o_proj.bias": "model-00004-of-00009.safetensors",
71
+ "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00009.safetensors",
72
+ "model.layers.11.self_attn.q_proj.bias": "model-00004-of-00009.safetensors",
73
+ "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00009.safetensors",
74
+ "model.layers.11.self_attn.sinks": "model-00004-of-00009.safetensors",
75
+ "model.layers.11.self_attn.v_proj.bias": "model-00004-of-00009.safetensors",
76
+ "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00009.safetensors",
77
+ "model.layers.12.input_layernorm.weight": "model-00005-of-00009.safetensors",
78
+ "model.layers.12.mlp.experts.down_proj": "model-00005-of-00009.safetensors",
79
+ "model.layers.12.mlp.experts.down_proj_bias": "model-00005-of-00009.safetensors",
80
+ "model.layers.12.mlp.experts.gate_up_proj": "model-00005-of-00009.safetensors",
81
+ "model.layers.12.mlp.experts.gate_up_proj_bias": "model-00005-of-00009.safetensors",
82
+ "model.layers.12.mlp.router.bias": "model-00005-of-00009.safetensors",
83
+ "model.layers.12.mlp.router.weight": "model-00005-of-00009.safetensors",
84
+ "model.layers.12.post_attention_layernorm.weight": "model-00005-of-00009.safetensors",
85
+ "model.layers.12.self_attn.k_proj.bias": "model-00005-of-00009.safetensors",
86
+ "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00009.safetensors",
87
+ "model.layers.12.self_attn.o_proj.bias": "model-00005-of-00009.safetensors",
88
+ "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00009.safetensors",
89
+ "model.layers.12.self_attn.q_proj.bias": "model-00005-of-00009.safetensors",
90
+ "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00009.safetensors",
91
+ "model.layers.12.self_attn.sinks": "model-00005-of-00009.safetensors",
92
+ "model.layers.12.self_attn.v_proj.bias": "model-00005-of-00009.safetensors",
93
+ "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00009.safetensors",
94
+ "model.layers.13.input_layernorm.weight": "model-00005-of-00009.safetensors",
95
+ "model.layers.13.mlp.experts.down_proj": "model-00005-of-00009.safetensors",
96
+ "model.layers.13.mlp.experts.down_proj_bias": "model-00005-of-00009.safetensors",
97
+ "model.layers.13.mlp.experts.gate_up_proj": "model-00005-of-00009.safetensors",
98
+ "model.layers.13.mlp.experts.gate_up_proj_bias": "model-00005-of-00009.safetensors",
99
+ "model.layers.13.mlp.router.bias": "model-00005-of-00009.safetensors",
100
+ "model.layers.13.mlp.router.weight": "model-00005-of-00009.safetensors",
101
+ "model.layers.13.post_attention_layernorm.weight": "model-00005-of-00009.safetensors",
102
+ "model.layers.13.self_attn.k_proj.bias": "model-00005-of-00009.safetensors",
103
+ "model.layers.13.self_attn.k_proj.weight": "model-00005-of-00009.safetensors",
104
+ "model.layers.13.self_attn.o_proj.bias": "model-00005-of-00009.safetensors",
105
+ "model.layers.13.self_attn.o_proj.weight": "model-00005-of-00009.safetensors",
106
+ "model.layers.13.self_attn.q_proj.bias": "model-00005-of-00009.safetensors",
107
+ "model.layers.13.self_attn.q_proj.weight": "model-00005-of-00009.safetensors",
108
+ "model.layers.13.self_attn.sinks": "model-00005-of-00009.safetensors",
109
+ "model.layers.13.self_attn.v_proj.bias": "model-00005-of-00009.safetensors",
110
+ "model.layers.13.self_attn.v_proj.weight": "model-00005-of-00009.safetensors",
111
+ "model.layers.14.input_layernorm.weight": "model-00006-of-00009.safetensors",
112
+ "model.layers.14.mlp.experts.down_proj": "model-00006-of-00009.safetensors",
113
+ "model.layers.14.mlp.experts.down_proj_bias": "model-00006-of-00009.safetensors",
114
+ "model.layers.14.mlp.experts.gate_up_proj": "model-00006-of-00009.safetensors",
115
+ "model.layers.14.mlp.experts.gate_up_proj_bias": "model-00006-of-00009.safetensors",
116
+ "model.layers.14.mlp.router.bias": "model-00005-of-00009.safetensors",
117
+ "model.layers.14.mlp.router.weight": "model-00005-of-00009.safetensors",
118
+ "model.layers.14.post_attention_layernorm.weight": "model-00006-of-00009.safetensors",
119
+ "model.layers.14.self_attn.k_proj.bias": "model-00005-of-00009.safetensors",
120
+ "model.layers.14.self_attn.k_proj.weight": "model-00005-of-00009.safetensors",
121
+ "model.layers.14.self_attn.o_proj.bias": "model-00005-of-00009.safetensors",
122
+ "model.layers.14.self_attn.o_proj.weight": "model-00005-of-00009.safetensors",
123
+ "model.layers.14.self_attn.q_proj.bias": "model-00005-of-00009.safetensors",
124
+ "model.layers.14.self_attn.q_proj.weight": "model-00005-of-00009.safetensors",
125
+ "model.layers.14.self_attn.sinks": "model-00005-of-00009.safetensors",
126
+ "model.layers.14.self_attn.v_proj.bias": "model-00005-of-00009.safetensors",
127
+ "model.layers.14.self_attn.v_proj.weight": "model-00005-of-00009.safetensors",
128
+ "model.layers.15.input_layernorm.weight": "model-00006-of-00009.safetensors",
129
+ "model.layers.15.mlp.experts.down_proj": "model-00006-of-00009.safetensors",
130
+ "model.layers.15.mlp.experts.down_proj_bias": "model-00006-of-00009.safetensors",
131
+ "model.layers.15.mlp.experts.gate_up_proj": "model-00006-of-00009.safetensors",
132
+ "model.layers.15.mlp.experts.gate_up_proj_bias": "model-00006-of-00009.safetensors",
133
+ "model.layers.15.mlp.router.bias": "model-00006-of-00009.safetensors",
134
+ "model.layers.15.mlp.router.weight": "model-00006-of-00009.safetensors",
135
+ "model.layers.15.post_attention_layernorm.weight": "model-00006-of-00009.safetensors",
136
+ "model.layers.15.self_attn.k_proj.bias": "model-00006-of-00009.safetensors",
137
+ "model.layers.15.self_attn.k_proj.weight": "model-00006-of-00009.safetensors",
138
+ "model.layers.15.self_attn.o_proj.bias": "model-00006-of-00009.safetensors",
139
+ "model.layers.15.self_attn.o_proj.weight": "model-00006-of-00009.safetensors",
140
+ "model.layers.15.self_attn.q_proj.bias": "model-00006-of-00009.safetensors",
141
+ "model.layers.15.self_attn.q_proj.weight": "model-00006-of-00009.safetensors",
142
+ "model.layers.15.self_attn.sinks": "model-00006-of-00009.safetensors",
143
+ "model.layers.15.self_attn.v_proj.bias": "model-00006-of-00009.safetensors",
144
+ "model.layers.15.self_attn.v_proj.weight": "model-00006-of-00009.safetensors",
145
+ "model.layers.16.input_layernorm.weight": "model-00006-of-00009.safetensors",
146
+ "model.layers.16.mlp.experts.down_proj": "model-00006-of-00009.safetensors",
147
+ "model.layers.16.mlp.experts.down_proj_bias": "model-00006-of-00009.safetensors",
148
+ "model.layers.16.mlp.experts.gate_up_proj": "model-00006-of-00009.safetensors",
149
+ "model.layers.16.mlp.experts.gate_up_proj_bias": "model-00006-of-00009.safetensors",
150
+ "model.layers.16.mlp.router.bias": "model-00006-of-00009.safetensors",
151
+ "model.layers.16.mlp.router.weight": "model-00006-of-00009.safetensors",
152
+ "model.layers.16.post_attention_layernorm.weight": "model-00006-of-00009.safetensors",
153
+ "model.layers.16.self_attn.k_proj.bias": "model-00006-of-00009.safetensors",
154
+ "model.layers.16.self_attn.k_proj.weight": "model-00006-of-00009.safetensors",
155
+ "model.layers.16.self_attn.o_proj.bias": "model-00006-of-00009.safetensors",
156
+ "model.layers.16.self_attn.o_proj.weight": "model-00006-of-00009.safetensors",
157
+ "model.layers.16.self_attn.q_proj.bias": "model-00006-of-00009.safetensors",
158
+ "model.layers.16.self_attn.q_proj.weight": "model-00006-of-00009.safetensors",
159
+ "model.layers.16.self_attn.sinks": "model-00006-of-00009.safetensors",
160
+ "model.layers.16.self_attn.v_proj.bias": "model-00006-of-00009.safetensors",
161
+ "model.layers.16.self_attn.v_proj.weight": "model-00006-of-00009.safetensors",
162
+ "model.layers.17.input_layernorm.weight": "model-00007-of-00009.safetensors",
163
+ "model.layers.17.mlp.experts.down_proj": "model-00007-of-00009.safetensors",
164
+ "model.layers.17.mlp.experts.down_proj_bias": "model-00007-of-00009.safetensors",
165
+ "model.layers.17.mlp.experts.gate_up_proj": "model-00007-of-00009.safetensors",
166
+ "model.layers.17.mlp.experts.gate_up_proj_bias": "model-00007-of-00009.safetensors",
167
+ "model.layers.17.mlp.router.bias": "model-00006-of-00009.safetensors",
168
+ "model.layers.17.mlp.router.weight": "model-00006-of-00009.safetensors",
169
+ "model.layers.17.post_attention_layernorm.weight": "model-00007-of-00009.safetensors",
170
+ "model.layers.17.self_attn.k_proj.bias": "model-00006-of-00009.safetensors",
171
+ "model.layers.17.self_attn.k_proj.weight": "model-00006-of-00009.safetensors",
172
+ "model.layers.17.self_attn.o_proj.bias": "model-00006-of-00009.safetensors",
173
+ "model.layers.17.self_attn.o_proj.weight": "model-00006-of-00009.safetensors",
174
+ "model.layers.17.self_attn.q_proj.bias": "model-00006-of-00009.safetensors",
175
+ "model.layers.17.self_attn.q_proj.weight": "model-00006-of-00009.safetensors",
176
+ "model.layers.17.self_attn.sinks": "model-00006-of-00009.safetensors",
177
+ "model.layers.17.self_attn.v_proj.bias": "model-00006-of-00009.safetensors",
178
+ "model.layers.17.self_attn.v_proj.weight": "model-00006-of-00009.safetensors",
179
+ "model.layers.18.input_layernorm.weight": "model-00007-of-00009.safetensors",
180
+ "model.layers.18.mlp.experts.down_proj": "model-00007-of-00009.safetensors",
181
+ "model.layers.18.mlp.experts.down_proj_bias": "model-00007-of-00009.safetensors",
182
+ "model.layers.18.mlp.experts.gate_up_proj": "model-00007-of-00009.safetensors",
183
+ "model.layers.18.mlp.experts.gate_up_proj_bias": "model-00007-of-00009.safetensors",
184
+ "model.layers.18.mlp.router.bias": "model-00007-of-00009.safetensors",
185
+ "model.layers.18.mlp.router.weight": "model-00007-of-00009.safetensors",
186
+ "model.layers.18.post_attention_layernorm.weight": "model-00007-of-00009.safetensors",
187
+ "model.layers.18.self_attn.k_proj.bias": "model-00007-of-00009.safetensors",
188
+ "model.layers.18.self_attn.k_proj.weight": "model-00007-of-00009.safetensors",
189
+ "model.layers.18.self_attn.o_proj.bias": "model-00007-of-00009.safetensors",
190
+ "model.layers.18.self_attn.o_proj.weight": "model-00007-of-00009.safetensors",
191
+ "model.layers.18.self_attn.q_proj.bias": "model-00007-of-00009.safetensors",
192
+ "model.layers.18.self_attn.q_proj.weight": "model-00007-of-00009.safetensors",
193
+ "model.layers.18.self_attn.sinks": "model-00007-of-00009.safetensors",
194
+ "model.layers.18.self_attn.v_proj.bias": "model-00007-of-00009.safetensors",
195
+ "model.layers.18.self_attn.v_proj.weight": "model-00007-of-00009.safetensors",
196
+ "model.layers.19.input_layernorm.weight": "model-00007-of-00009.safetensors",
197
+ "model.layers.19.mlp.experts.down_proj": "model-00007-of-00009.safetensors",
198
+ "model.layers.19.mlp.experts.down_proj_bias": "model-00007-of-00009.safetensors",
199
+ "model.layers.19.mlp.experts.gate_up_proj": "model-00007-of-00009.safetensors",
200
+ "model.layers.19.mlp.experts.gate_up_proj_bias": "model-00007-of-00009.safetensors",
201
+ "model.layers.19.mlp.router.bias": "model-00007-of-00009.safetensors",
202
+ "model.layers.19.mlp.router.weight": "model-00007-of-00009.safetensors",
203
+ "model.layers.19.post_attention_layernorm.weight": "model-00007-of-00009.safetensors",
204
+ "model.layers.19.self_attn.k_proj.bias": "model-00007-of-00009.safetensors",
205
+ "model.layers.19.self_attn.k_proj.weight": "model-00007-of-00009.safetensors",
206
+ "model.layers.19.self_attn.o_proj.bias": "model-00007-of-00009.safetensors",
207
+ "model.layers.19.self_attn.o_proj.weight": "model-00007-of-00009.safetensors",
208
+ "model.layers.19.self_attn.q_proj.bias": "model-00007-of-00009.safetensors",
209
+ "model.layers.19.self_attn.q_proj.weight": "model-00007-of-00009.safetensors",
210
+ "model.layers.19.self_attn.sinks": "model-00007-of-00009.safetensors",
211
+ "model.layers.19.self_attn.v_proj.bias": "model-00007-of-00009.safetensors",
212
+ "model.layers.19.self_attn.v_proj.weight": "model-00007-of-00009.safetensors",
213
+ "model.layers.2.input_layernorm.weight": "model-00002-of-00009.safetensors",
214
+ "model.layers.2.mlp.experts.down_proj": "model-00002-of-00009.safetensors",
215
+ "model.layers.2.mlp.experts.down_proj_bias": "model-00002-of-00009.safetensors",
216
+ "model.layers.2.mlp.experts.gate_up_proj": "model-00002-of-00009.safetensors",
217
+ "model.layers.2.mlp.experts.gate_up_proj_bias": "model-00002-of-00009.safetensors",
218
+ "model.layers.2.mlp.router.bias": "model-00001-of-00009.safetensors",
219
+ "model.layers.2.mlp.router.weight": "model-00001-of-00009.safetensors",
220
+ "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00009.safetensors",
221
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00009.safetensors",
222
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00009.safetensors",
223
+ "model.layers.2.self_attn.o_proj.bias": "model-00001-of-00009.safetensors",
224
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00009.safetensors",
225
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00009.safetensors",
226
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00009.safetensors",
227
+ "model.layers.2.self_attn.sinks": "model-00001-of-00009.safetensors",
228
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00009.safetensors",
229
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00009.safetensors",
230
+ "model.layers.20.input_layernorm.weight": "model-00008-of-00009.safetensors",
231
+ "model.layers.20.mlp.experts.down_proj": "model-00008-of-00009.safetensors",
232
+ "model.layers.20.mlp.experts.down_proj_bias": "model-00008-of-00009.safetensors",
233
+ "model.layers.20.mlp.experts.gate_up_proj": "model-00008-of-00009.safetensors",
234
+ "model.layers.20.mlp.experts.gate_up_proj_bias": "model-00008-of-00009.safetensors",
235
+ "model.layers.20.mlp.router.bias": "model-00007-of-00009.safetensors",
236
+ "model.layers.20.mlp.router.weight": "model-00007-of-00009.safetensors",
237
+ "model.layers.20.post_attention_layernorm.weight": "model-00008-of-00009.safetensors",
238
+ "model.layers.20.self_attn.k_proj.bias": "model-00007-of-00009.safetensors",
239
+ "model.layers.20.self_attn.k_proj.weight": "model-00007-of-00009.safetensors",
240
+ "model.layers.20.self_attn.o_proj.bias": "model-00007-of-00009.safetensors",
241
+ "model.layers.20.self_attn.o_proj.weight": "model-00007-of-00009.safetensors",
242
+ "model.layers.20.self_attn.q_proj.bias": "model-00007-of-00009.safetensors",
243
+ "model.layers.20.self_attn.q_proj.weight": "model-00007-of-00009.safetensors",
244
+ "model.layers.20.self_attn.sinks": "model-00007-of-00009.safetensors",
245
+ "model.layers.20.self_attn.v_proj.bias": "model-00007-of-00009.safetensors",
246
+ "model.layers.20.self_attn.v_proj.weight": "model-00007-of-00009.safetensors",
247
+ "model.layers.21.input_layernorm.weight": "model-00008-of-00009.safetensors",
248
+ "model.layers.21.mlp.experts.down_proj": "model-00008-of-00009.safetensors",
249
+ "model.layers.21.mlp.experts.down_proj_bias": "model-00008-of-00009.safetensors",
250
+ "model.layers.21.mlp.experts.gate_up_proj": "model-00008-of-00009.safetensors",
251
+ "model.layers.21.mlp.experts.gate_up_proj_bias": "model-00008-of-00009.safetensors",
252
+ "model.layers.21.mlp.router.bias": "model-00008-of-00009.safetensors",
253
+ "model.layers.21.mlp.router.weight": "model-00008-of-00009.safetensors",
254
+ "model.layers.21.post_attention_layernorm.weight": "model-00008-of-00009.safetensors",
255
+ "model.layers.21.self_attn.k_proj.bias": "model-00008-of-00009.safetensors",
256
+ "model.layers.21.self_attn.k_proj.weight": "model-00008-of-00009.safetensors",
257
+ "model.layers.21.self_attn.o_proj.bias": "model-00008-of-00009.safetensors",
258
+ "model.layers.21.self_attn.o_proj.weight": "model-00008-of-00009.safetensors",
259
+ "model.layers.21.self_attn.q_proj.bias": "model-00008-of-00009.safetensors",
260
+ "model.layers.21.self_attn.q_proj.weight": "model-00008-of-00009.safetensors",
261
+ "model.layers.21.self_attn.sinks": "model-00008-of-00009.safetensors",
262
+ "model.layers.21.self_attn.v_proj.bias": "model-00008-of-00009.safetensors",
263
+ "model.layers.21.self_attn.v_proj.weight": "model-00008-of-00009.safetensors",
264
+ "model.layers.22.input_layernorm.weight": "model-00008-of-00009.safetensors",
265
+ "model.layers.22.mlp.experts.down_proj": "model-00008-of-00009.safetensors",
266
+ "model.layers.22.mlp.experts.down_proj_bias": "model-00008-of-00009.safetensors",
267
+ "model.layers.22.mlp.experts.gate_up_proj": "model-00008-of-00009.safetensors",
268
+ "model.layers.22.mlp.experts.gate_up_proj_bias": "model-00008-of-00009.safetensors",
269
+ "model.layers.22.mlp.router.bias": "model-00008-of-00009.safetensors",
270
+ "model.layers.22.mlp.router.weight": "model-00008-of-00009.safetensors",
271
+ "model.layers.22.post_attention_layernorm.weight": "model-00008-of-00009.safetensors",
272
+ "model.layers.22.self_attn.k_proj.bias": "model-00008-of-00009.safetensors",
273
+ "model.layers.22.self_attn.k_proj.weight": "model-00008-of-00009.safetensors",
274
+ "model.layers.22.self_attn.o_proj.bias": "model-00008-of-00009.safetensors",
275
+ "model.layers.22.self_attn.o_proj.weight": "model-00008-of-00009.safetensors",
276
+ "model.layers.22.self_attn.q_proj.bias": "model-00008-of-00009.safetensors",
277
+ "model.layers.22.self_attn.q_proj.weight": "model-00008-of-00009.safetensors",
278
+ "model.layers.22.self_attn.sinks": "model-00008-of-00009.safetensors",
279
+ "model.layers.22.self_attn.v_proj.bias": "model-00008-of-00009.safetensors",
280
+ "model.layers.22.self_attn.v_proj.weight": "model-00008-of-00009.safetensors",
281
+ "model.layers.23.input_layernorm.weight": "model-00009-of-00009.safetensors",
282
+ "model.layers.23.mlp.experts.down_proj": "model-00009-of-00009.safetensors",
283
+ "model.layers.23.mlp.experts.down_proj_bias": "model-00009-of-00009.safetensors",
284
+ "model.layers.23.mlp.experts.gate_up_proj": "model-00009-of-00009.safetensors",
285
+ "model.layers.23.mlp.experts.gate_up_proj_bias": "model-00009-of-00009.safetensors",
286
+ "model.layers.23.mlp.router.bias": "model-00008-of-00009.safetensors",
287
+ "model.layers.23.mlp.router.weight": "model-00008-of-00009.safetensors",
288
+ "model.layers.23.post_attention_layernorm.weight": "model-00009-of-00009.safetensors",
289
+ "model.layers.23.self_attn.k_proj.bias": "model-00008-of-00009.safetensors",
290
+ "model.layers.23.self_attn.k_proj.weight": "model-00008-of-00009.safetensors",
291
+ "model.layers.23.self_attn.o_proj.bias": "model-00008-of-00009.safetensors",
292
+ "model.layers.23.self_attn.o_proj.weight": "model-00008-of-00009.safetensors",
293
+ "model.layers.23.self_attn.q_proj.bias": "model-00008-of-00009.safetensors",
294
+ "model.layers.23.self_attn.q_proj.weight": "model-00008-of-00009.safetensors",
295
+ "model.layers.23.self_attn.sinks": "model-00008-of-00009.safetensors",
296
+ "model.layers.23.self_attn.v_proj.bias": "model-00008-of-00009.safetensors",
297
+ "model.layers.23.self_attn.v_proj.weight": "model-00008-of-00009.safetensors",
298
+ "model.layers.3.input_layernorm.weight": "model-00002-of-00009.safetensors",
299
+ "model.layers.3.mlp.experts.down_proj": "model-00002-of-00009.safetensors",
300
+ "model.layers.3.mlp.experts.down_proj_bias": "model-00002-of-00009.safetensors",
301
+ "model.layers.3.mlp.experts.gate_up_proj": "model-00002-of-00009.safetensors",
302
+ "model.layers.3.mlp.experts.gate_up_proj_bias": "model-00002-of-00009.safetensors",
303
+ "model.layers.3.mlp.router.bias": "model-00002-of-00009.safetensors",
304
+ "model.layers.3.mlp.router.weight": "model-00002-of-00009.safetensors",
305
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00009.safetensors",
306
+ "model.layers.3.self_attn.k_proj.bias": "model-00002-of-00009.safetensors",
307
+ "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00009.safetensors",
308
+ "model.layers.3.self_attn.o_proj.bias": "model-00002-of-00009.safetensors",
309
+ "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00009.safetensors",
310
+ "model.layers.3.self_attn.q_proj.bias": "model-00002-of-00009.safetensors",
311
+ "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00009.safetensors",
312
+ "model.layers.3.self_attn.sinks": "model-00002-of-00009.safetensors",
313
+ "model.layers.3.self_attn.v_proj.bias": "model-00002-of-00009.safetensors",
314
+ "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00009.safetensors",
315
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00009.safetensors",
316
+ "model.layers.4.mlp.experts.down_proj": "model-00002-of-00009.safetensors",
317
+ "model.layers.4.mlp.experts.down_proj_bias": "model-00002-of-00009.safetensors",
318
+ "model.layers.4.mlp.experts.gate_up_proj": "model-00002-of-00009.safetensors",
319
+ "model.layers.4.mlp.experts.gate_up_proj_bias": "model-00002-of-00009.safetensors",
320
+ "model.layers.4.mlp.router.bias": "model-00002-of-00009.safetensors",
321
+ "model.layers.4.mlp.router.weight": "model-00002-of-00009.safetensors",
322
+ "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00009.safetensors",
323
+ "model.layers.4.self_attn.k_proj.bias": "model-00002-of-00009.safetensors",
324
+ "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00009.safetensors",
325
+ "model.layers.4.self_attn.o_proj.bias": "model-00002-of-00009.safetensors",
326
+ "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00009.safetensors",
327
+ "model.layers.4.self_attn.q_proj.bias": "model-00002-of-00009.safetensors",
328
+ "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00009.safetensors",
329
+ "model.layers.4.self_attn.sinks": "model-00002-of-00009.safetensors",
330
+ "model.layers.4.self_attn.v_proj.bias": "model-00002-of-00009.safetensors",
331
+ "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00009.safetensors",
332
+ "model.layers.5.input_layernorm.weight": "model-00003-of-00009.safetensors",
333
+ "model.layers.5.mlp.experts.down_proj": "model-00003-of-00009.safetensors",
334
+ "model.layers.5.mlp.experts.down_proj_bias": "model-00003-of-00009.safetensors",
335
+ "model.layers.5.mlp.experts.gate_up_proj": "model-00003-of-00009.safetensors",
336
+ "model.layers.5.mlp.experts.gate_up_proj_bias": "model-00003-of-00009.safetensors",
337
+ "model.layers.5.mlp.router.bias": "model-00002-of-00009.safetensors",
338
+ "model.layers.5.mlp.router.weight": "model-00002-of-00009.safetensors",
339
+ "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00009.safetensors",
340
+ "model.layers.5.self_attn.k_proj.bias": "model-00002-of-00009.safetensors",
341
+ "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00009.safetensors",
342
+ "model.layers.5.self_attn.o_proj.bias": "model-00002-of-00009.safetensors",
343
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00009.safetensors",
344
+ "model.layers.5.self_attn.q_proj.bias": "model-00002-of-00009.safetensors",
345
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00009.safetensors",
346
+ "model.layers.5.self_attn.sinks": "model-00002-of-00009.safetensors",
347
+ "model.layers.5.self_attn.v_proj.bias": "model-00002-of-00009.safetensors",
348
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00009.safetensors",
349
+ "model.layers.6.input_layernorm.weight": "model-00003-of-00009.safetensors",
350
+ "model.layers.6.mlp.experts.down_proj": "model-00003-of-00009.safetensors",
351
+ "model.layers.6.mlp.experts.down_proj_bias": "model-00003-of-00009.safetensors",
352
+ "model.layers.6.mlp.experts.gate_up_proj": "model-00003-of-00009.safetensors",
353
+ "model.layers.6.mlp.experts.gate_up_proj_bias": "model-00003-of-00009.safetensors",
354
+ "model.layers.6.mlp.router.bias": "model-00003-of-00009.safetensors",
355
+ "model.layers.6.mlp.router.weight": "model-00003-of-00009.safetensors",
356
+ "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00009.safetensors",
357
+ "model.layers.6.self_attn.k_proj.bias": "model-00003-of-00009.safetensors",
358
+ "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00009.safetensors",
359
+ "model.layers.6.self_attn.o_proj.bias": "model-00003-of-00009.safetensors",
360
+ "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00009.safetensors",
361
+ "model.layers.6.self_attn.q_proj.bias": "model-00003-of-00009.safetensors",
362
+ "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00009.safetensors",
363
+ "model.layers.6.self_attn.sinks": "model-00003-of-00009.safetensors",
364
+ "model.layers.6.self_attn.v_proj.bias": "model-00003-of-00009.safetensors",
365
+ "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00009.safetensors",
366
+ "model.layers.7.input_layernorm.weight": "model-00003-of-00009.safetensors",
367
+ "model.layers.7.mlp.experts.down_proj": "model-00003-of-00009.safetensors",
368
+ "model.layers.7.mlp.experts.down_proj_bias": "model-00003-of-00009.safetensors",
369
+ "model.layers.7.mlp.experts.gate_up_proj": "model-00003-of-00009.safetensors",
370
+ "model.layers.7.mlp.experts.gate_up_proj_bias": "model-00003-of-00009.safetensors",
371
+ "model.layers.7.mlp.router.bias": "model-00003-of-00009.safetensors",
372
+ "model.layers.7.mlp.router.weight": "model-00003-of-00009.safetensors",
373
+ "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00009.safetensors",
374
+ "model.layers.7.self_attn.k_proj.bias": "model-00003-of-00009.safetensors",
375
+ "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00009.safetensors",
376
+ "model.layers.7.self_attn.o_proj.bias": "model-00003-of-00009.safetensors",
377
+ "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00009.safetensors",
378
+ "model.layers.7.self_attn.q_proj.bias": "model-00003-of-00009.safetensors",
379
+ "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00009.safetensors",
380
+ "model.layers.7.self_attn.sinks": "model-00003-of-00009.safetensors",
381
+ "model.layers.7.self_attn.v_proj.bias": "model-00003-of-00009.safetensors",
382
+ "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00009.safetensors",
383
+ "model.layers.8.input_layernorm.weight": "model-00004-of-00009.safetensors",
384
+ "model.layers.8.mlp.experts.down_proj": "model-00004-of-00009.safetensors",
385
+ "model.layers.8.mlp.experts.down_proj_bias": "model-00004-of-00009.safetensors",
386
+ "model.layers.8.mlp.experts.gate_up_proj": "model-00004-of-00009.safetensors",
387
+ "model.layers.8.mlp.experts.gate_up_proj_bias": "model-00004-of-00009.safetensors",
388
+ "model.layers.8.mlp.router.bias": "model-00003-of-00009.safetensors",
389
+ "model.layers.8.mlp.router.weight": "model-00003-of-00009.safetensors",
390
+ "model.layers.8.post_attention_layernorm.weight": "model-00004-of-00009.safetensors",
391
+ "model.layers.8.self_attn.k_proj.bias": "model-00003-of-00009.safetensors",
392
+ "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00009.safetensors",
393
+ "model.layers.8.self_attn.o_proj.bias": "model-00003-of-00009.safetensors",
394
+ "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00009.safetensors",
395
+ "model.layers.8.self_attn.q_proj.bias": "model-00003-of-00009.safetensors",
396
+ "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00009.safetensors",
397
+ "model.layers.8.self_attn.sinks": "model-00003-of-00009.safetensors",
398
+ "model.layers.8.self_attn.v_proj.bias": "model-00003-of-00009.safetensors",
399
+ "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00009.safetensors",
400
+ "model.layers.9.input_layernorm.weight": "model-00004-of-00009.safetensors",
401
+ "model.layers.9.mlp.experts.down_proj": "model-00004-of-00009.safetensors",
402
+ "model.layers.9.mlp.experts.down_proj_bias": "model-00004-of-00009.safetensors",
403
+ "model.layers.9.mlp.experts.gate_up_proj": "model-00004-of-00009.safetensors",
404
+ "model.layers.9.mlp.experts.gate_up_proj_bias": "model-00004-of-00009.safetensors",
405
+ "model.layers.9.mlp.router.bias": "model-00004-of-00009.safetensors",
406
+ "model.layers.9.mlp.router.weight": "model-00004-of-00009.safetensors",
407
+ "model.layers.9.post_attention_layernorm.weight": "model-00004-of-00009.safetensors",
408
+ "model.layers.9.self_attn.k_proj.bias": "model-00004-of-00009.safetensors",
409
+ "model.layers.9.self_attn.k_proj.weight": "model-00004-of-00009.safetensors",
410
+ "model.layers.9.self_attn.o_proj.bias": "model-00004-of-00009.safetensors",
411
+ "model.layers.9.self_attn.o_proj.weight": "model-00004-of-00009.safetensors",
412
+ "model.layers.9.self_attn.q_proj.bias": "model-00004-of-00009.safetensors",
413
+ "model.layers.9.self_attn.q_proj.weight": "model-00004-of-00009.safetensors",
414
+ "model.layers.9.self_attn.sinks": "model-00004-of-00009.safetensors",
415
+ "model.layers.9.self_attn.v_proj.bias": "model-00004-of-00009.safetensors",
416
+ "model.layers.9.self_attn.v_proj.weight": "model-00004-of-00009.safetensors",
417
+ "model.norm.weight": "model-00009-of-00009.safetensors"
418
+ }
419
+ }
checkpoint-1250/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e4466beaa56ab10637e9ff5d265e2d020747c591c11818d1acb0155aae394f7c
3
+ size 2794414739
checkpoint-1250/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61c19bab1174704a4a4441475683bf1270277af15d2e2c95e964789128e482c4
3
+ size 14645
checkpoint-1250/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ad1f24bf2246b019e5c4979ace691aa118b7de7cc84ddd0aa9224068c0079d8
3
+ size 1465
checkpoint-1250/special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|return|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
checkpoint-1250/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0614fe83cadab421296e664e1f48f4261fa8fef6e03e63bb75c20f38e37d07d3
3
+ size 27868174
checkpoint-1250/tokenizer_config.json ADDED
@@ -0,0 +1,183 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "199998": {
4
+ "content": "<|startoftext|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "199999": {
12
+ "content": "<|endoftext|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "200000": {
20
+ "content": "<|reserved_200000|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "200001": {
28
+ "content": "<|reserved_200001|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "200002": {
36
+ "content": "<|return|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "200003": {
44
+ "content": "<|constrain|>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "200004": {
52
+ "content": "<|reserved_200004|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "200005": {
60
+ "content": "<|channel|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "200006": {
68
+ "content": "<|start|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "200007": {
76
+ "content": "<|end|>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "200008": {
84
+ "content": "<|message|>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "200009": {
92
+ "content": "<|reserved_200009|>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "200010": {
100
+ "content": "<|reserved_200010|>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "200011": {
108
+ "content": "<|reserved_200011|>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "200012": {
116
+ "content": "<|call|>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "200013": {
124
+ "content": "<|reserved_200013|>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "200014": {
132
+ "content": "<|reserved_200014|>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "200015": {
140
+ "content": "<|reserved_200015|>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "200016": {
148
+ "content": "<|reserved_200016|>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "200017": {
156
+ "content": "<|reserved_200017|>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "200018": {
164
+ "content": "<|endofprompt|>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ }
171
+ },
172
+ "bos_token": "<|startoftext|>",
173
+ "clean_up_tokenization_spaces": false,
174
+ "eos_token": "<|return|>",
175
+ "extra_special_tokens": {},
176
+ "model_input_names": [
177
+ "input_ids",
178
+ "attention_mask"
179
+ ],
180
+ "model_max_length": 1000000000000000019884624838656,
181
+ "pad_token": "<|endoftext|>",
182
+ "tokenizer_class": "PreTrainedTokenizerFast"
183
+ }
checkpoint-1250/trainer_state.json ADDED
@@ -0,0 +1,916 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1250,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.0008,
14
+ "grad_norm": 11.4375,
15
+ "learning_rate": 0.0,
16
+ "loss": 3.8345,
17
+ "step": 1
18
+ },
19
+ {
20
+ "epoch": 0.008,
21
+ "grad_norm": 11.125,
22
+ "learning_rate": 2.25e-06,
23
+ "loss": 3.3934,
24
+ "step": 10
25
+ },
26
+ {
27
+ "epoch": 0.016,
28
+ "grad_norm": 9.5,
29
+ "learning_rate": 4.75e-06,
30
+ "loss": 3.4002,
31
+ "step": 20
32
+ },
33
+ {
34
+ "epoch": 0.024,
35
+ "grad_norm": 9.4375,
36
+ "learning_rate": 7.25e-06,
37
+ "loss": 3.4003,
38
+ "step": 30
39
+ },
40
+ {
41
+ "epoch": 0.032,
42
+ "grad_norm": 11.1875,
43
+ "learning_rate": 9.750000000000002e-06,
44
+ "loss": 3.374,
45
+ "step": 40
46
+ },
47
+ {
48
+ "epoch": 0.04,
49
+ "grad_norm": 10.625,
50
+ "learning_rate": 1.225e-05,
51
+ "loss": 3.4678,
52
+ "step": 50
53
+ },
54
+ {
55
+ "epoch": 0.048,
56
+ "grad_norm": 11.0,
57
+ "learning_rate": 1.475e-05,
58
+ "loss": 3.4331,
59
+ "step": 60
60
+ },
61
+ {
62
+ "epoch": 0.056,
63
+ "grad_norm": 10.6875,
64
+ "learning_rate": 1.725e-05,
65
+ "loss": 3.4722,
66
+ "step": 70
67
+ },
68
+ {
69
+ "epoch": 0.064,
70
+ "grad_norm": 10.25,
71
+ "learning_rate": 1.9750000000000002e-05,
72
+ "loss": 3.2413,
73
+ "step": 80
74
+ },
75
+ {
76
+ "epoch": 0.072,
77
+ "grad_norm": 10.8125,
78
+ "learning_rate": 2.2250000000000002e-05,
79
+ "loss": 3.1804,
80
+ "step": 90
81
+ },
82
+ {
83
+ "epoch": 0.08,
84
+ "grad_norm": 10.1875,
85
+ "learning_rate": 2.4750000000000002e-05,
86
+ "loss": 3.1778,
87
+ "step": 100
88
+ },
89
+ {
90
+ "epoch": 0.088,
91
+ "grad_norm": 8.4375,
92
+ "learning_rate": 2.725e-05,
93
+ "loss": 3.2479,
94
+ "step": 110
95
+ },
96
+ {
97
+ "epoch": 0.096,
98
+ "grad_norm": 11.0,
99
+ "learning_rate": 2.975e-05,
100
+ "loss": 3.2259,
101
+ "step": 120
102
+ },
103
+ {
104
+ "epoch": 0.104,
105
+ "grad_norm": 9.5625,
106
+ "learning_rate": 3.2250000000000005e-05,
107
+ "loss": 3.0845,
108
+ "step": 130
109
+ },
110
+ {
111
+ "epoch": 0.112,
112
+ "grad_norm": 8.9375,
113
+ "learning_rate": 3.475e-05,
114
+ "loss": 3.1694,
115
+ "step": 140
116
+ },
117
+ {
118
+ "epoch": 0.12,
119
+ "grad_norm": 8.4375,
120
+ "learning_rate": 3.7250000000000004e-05,
121
+ "loss": 3.1736,
122
+ "step": 150
123
+ },
124
+ {
125
+ "epoch": 0.128,
126
+ "grad_norm": 9.5,
127
+ "learning_rate": 3.9750000000000004e-05,
128
+ "loss": 3.035,
129
+ "step": 160
130
+ },
131
+ {
132
+ "epoch": 0.136,
133
+ "grad_norm": 9.8125,
134
+ "learning_rate": 4.2250000000000004e-05,
135
+ "loss": 3.1456,
136
+ "step": 170
137
+ },
138
+ {
139
+ "epoch": 0.144,
140
+ "grad_norm": 9.25,
141
+ "learning_rate": 4.4750000000000004e-05,
142
+ "loss": 3.0917,
143
+ "step": 180
144
+ },
145
+ {
146
+ "epoch": 0.152,
147
+ "grad_norm": 9.1875,
148
+ "learning_rate": 4.7249999999999997e-05,
149
+ "loss": 2.9948,
150
+ "step": 190
151
+ },
152
+ {
153
+ "epoch": 0.16,
154
+ "grad_norm": 8.5625,
155
+ "learning_rate": 4.975e-05,
156
+ "loss": 3.0667,
157
+ "step": 200
158
+ },
159
+ {
160
+ "epoch": 0.168,
161
+ "grad_norm": 8.125,
162
+ "learning_rate": 4.957142857142857e-05,
163
+ "loss": 2.9987,
164
+ "step": 210
165
+ },
166
+ {
167
+ "epoch": 0.176,
168
+ "grad_norm": 7.875,
169
+ "learning_rate": 4.90952380952381e-05,
170
+ "loss": 2.9686,
171
+ "step": 220
172
+ },
173
+ {
174
+ "epoch": 0.184,
175
+ "grad_norm": 9.1875,
176
+ "learning_rate": 4.861904761904762e-05,
177
+ "loss": 3.0708,
178
+ "step": 230
179
+ },
180
+ {
181
+ "epoch": 0.192,
182
+ "grad_norm": 7.90625,
183
+ "learning_rate": 4.8142857142857147e-05,
184
+ "loss": 3.0676,
185
+ "step": 240
186
+ },
187
+ {
188
+ "epoch": 0.2,
189
+ "grad_norm": 9.375,
190
+ "learning_rate": 4.766666666666667e-05,
191
+ "loss": 2.8833,
192
+ "step": 250
193
+ },
194
+ {
195
+ "epoch": 0.208,
196
+ "grad_norm": 8.75,
197
+ "learning_rate": 4.719047619047619e-05,
198
+ "loss": 3.0558,
199
+ "step": 260
200
+ },
201
+ {
202
+ "epoch": 0.216,
203
+ "grad_norm": 9.375,
204
+ "learning_rate": 4.671428571428571e-05,
205
+ "loss": 3.0003,
206
+ "step": 270
207
+ },
208
+ {
209
+ "epoch": 0.224,
210
+ "grad_norm": 8.25,
211
+ "learning_rate": 4.623809523809524e-05,
212
+ "loss": 2.9514,
213
+ "step": 280
214
+ },
215
+ {
216
+ "epoch": 0.232,
217
+ "grad_norm": 7.15625,
218
+ "learning_rate": 4.5761904761904765e-05,
219
+ "loss": 2.984,
220
+ "step": 290
221
+ },
222
+ {
223
+ "epoch": 0.24,
224
+ "grad_norm": 8.625,
225
+ "learning_rate": 4.528571428571429e-05,
226
+ "loss": 3.0273,
227
+ "step": 300
228
+ },
229
+ {
230
+ "epoch": 0.248,
231
+ "grad_norm": 7.75,
232
+ "learning_rate": 4.480952380952381e-05,
233
+ "loss": 2.9437,
234
+ "step": 310
235
+ },
236
+ {
237
+ "epoch": 0.256,
238
+ "grad_norm": 9.1875,
239
+ "learning_rate": 4.433333333333334e-05,
240
+ "loss": 3.0913,
241
+ "step": 320
242
+ },
243
+ {
244
+ "epoch": 0.264,
245
+ "grad_norm": 8.5625,
246
+ "learning_rate": 4.385714285714286e-05,
247
+ "loss": 2.9659,
248
+ "step": 330
249
+ },
250
+ {
251
+ "epoch": 0.272,
252
+ "grad_norm": 8.0625,
253
+ "learning_rate": 4.338095238095238e-05,
254
+ "loss": 2.854,
255
+ "step": 340
256
+ },
257
+ {
258
+ "epoch": 0.28,
259
+ "grad_norm": 7.0,
260
+ "learning_rate": 4.290476190476191e-05,
261
+ "loss": 2.8936,
262
+ "step": 350
263
+ },
264
+ {
265
+ "epoch": 0.288,
266
+ "grad_norm": 8.125,
267
+ "learning_rate": 4.242857142857143e-05,
268
+ "loss": 2.9805,
269
+ "step": 360
270
+ },
271
+ {
272
+ "epoch": 0.296,
273
+ "grad_norm": 8.5,
274
+ "learning_rate": 4.1952380952380956e-05,
275
+ "loss": 2.9814,
276
+ "step": 370
277
+ },
278
+ {
279
+ "epoch": 0.304,
280
+ "grad_norm": 7.25,
281
+ "learning_rate": 4.147619047619048e-05,
282
+ "loss": 2.8731,
283
+ "step": 380
284
+ },
285
+ {
286
+ "epoch": 0.312,
287
+ "grad_norm": 7.4375,
288
+ "learning_rate": 4.1e-05,
289
+ "loss": 2.8186,
290
+ "step": 390
291
+ },
292
+ {
293
+ "epoch": 0.32,
294
+ "grad_norm": 8.0,
295
+ "learning_rate": 4.052380952380952e-05,
296
+ "loss": 3.048,
297
+ "step": 400
298
+ },
299
+ {
300
+ "epoch": 0.328,
301
+ "grad_norm": 8.5,
302
+ "learning_rate": 4.004761904761905e-05,
303
+ "loss": 2.8923,
304
+ "step": 410
305
+ },
306
+ {
307
+ "epoch": 0.336,
308
+ "grad_norm": 7.96875,
309
+ "learning_rate": 3.9571428571428574e-05,
310
+ "loss": 2.8443,
311
+ "step": 420
312
+ },
313
+ {
314
+ "epoch": 0.344,
315
+ "grad_norm": 7.75,
316
+ "learning_rate": 3.9095238095238096e-05,
317
+ "loss": 3.0182,
318
+ "step": 430
319
+ },
320
+ {
321
+ "epoch": 0.352,
322
+ "grad_norm": 7.625,
323
+ "learning_rate": 3.861904761904762e-05,
324
+ "loss": 2.849,
325
+ "step": 440
326
+ },
327
+ {
328
+ "epoch": 0.36,
329
+ "grad_norm": 8.125,
330
+ "learning_rate": 3.814285714285715e-05,
331
+ "loss": 2.8413,
332
+ "step": 450
333
+ },
334
+ {
335
+ "epoch": 0.368,
336
+ "grad_norm": 7.6875,
337
+ "learning_rate": 3.766666666666667e-05,
338
+ "loss": 2.9321,
339
+ "step": 460
340
+ },
341
+ {
342
+ "epoch": 0.376,
343
+ "grad_norm": 7.15625,
344
+ "learning_rate": 3.719047619047619e-05,
345
+ "loss": 2.9628,
346
+ "step": 470
347
+ },
348
+ {
349
+ "epoch": 0.384,
350
+ "grad_norm": 8.0625,
351
+ "learning_rate": 3.671428571428572e-05,
352
+ "loss": 3.0175,
353
+ "step": 480
354
+ },
355
+ {
356
+ "epoch": 0.392,
357
+ "grad_norm": 7.375,
358
+ "learning_rate": 3.6238095238095236e-05,
359
+ "loss": 2.9446,
360
+ "step": 490
361
+ },
362
+ {
363
+ "epoch": 0.4,
364
+ "grad_norm": 7.625,
365
+ "learning_rate": 3.5761904761904765e-05,
366
+ "loss": 2.8095,
367
+ "step": 500
368
+ },
369
+ {
370
+ "epoch": 0.408,
371
+ "grad_norm": 9.25,
372
+ "learning_rate": 3.528571428571429e-05,
373
+ "loss": 2.9347,
374
+ "step": 510
375
+ },
376
+ {
377
+ "epoch": 0.416,
378
+ "grad_norm": 7.875,
379
+ "learning_rate": 3.480952380952381e-05,
380
+ "loss": 2.7751,
381
+ "step": 520
382
+ },
383
+ {
384
+ "epoch": 0.424,
385
+ "grad_norm": 8.5625,
386
+ "learning_rate": 3.433333333333333e-05,
387
+ "loss": 2.8434,
388
+ "step": 530
389
+ },
390
+ {
391
+ "epoch": 0.432,
392
+ "grad_norm": 7.71875,
393
+ "learning_rate": 3.385714285714286e-05,
394
+ "loss": 2.869,
395
+ "step": 540
396
+ },
397
+ {
398
+ "epoch": 0.44,
399
+ "grad_norm": 6.75,
400
+ "learning_rate": 3.338095238095238e-05,
401
+ "loss": 2.8466,
402
+ "step": 550
403
+ },
404
+ {
405
+ "epoch": 0.448,
406
+ "grad_norm": 7.4375,
407
+ "learning_rate": 3.2904761904761906e-05,
408
+ "loss": 2.7286,
409
+ "step": 560
410
+ },
411
+ {
412
+ "epoch": 0.456,
413
+ "grad_norm": 8.4375,
414
+ "learning_rate": 3.242857142857143e-05,
415
+ "loss": 2.9655,
416
+ "step": 570
417
+ },
418
+ {
419
+ "epoch": 0.464,
420
+ "grad_norm": 7.3125,
421
+ "learning_rate": 3.195238095238096e-05,
422
+ "loss": 2.7955,
423
+ "step": 580
424
+ },
425
+ {
426
+ "epoch": 0.472,
427
+ "grad_norm": 9.25,
428
+ "learning_rate": 3.147619047619048e-05,
429
+ "loss": 2.8469,
430
+ "step": 590
431
+ },
432
+ {
433
+ "epoch": 0.48,
434
+ "grad_norm": 8.25,
435
+ "learning_rate": 3.1e-05,
436
+ "loss": 2.9423,
437
+ "step": 600
438
+ },
439
+ {
440
+ "epoch": 0.488,
441
+ "grad_norm": 8.0625,
442
+ "learning_rate": 3.0523809523809524e-05,
443
+ "loss": 2.9848,
444
+ "step": 610
445
+ },
446
+ {
447
+ "epoch": 0.496,
448
+ "grad_norm": 7.25,
449
+ "learning_rate": 3.0047619047619046e-05,
450
+ "loss": 2.8185,
451
+ "step": 620
452
+ },
453
+ {
454
+ "epoch": 0.504,
455
+ "grad_norm": 8.3125,
456
+ "learning_rate": 2.957142857142857e-05,
457
+ "loss": 2.7036,
458
+ "step": 630
459
+ },
460
+ {
461
+ "epoch": 0.512,
462
+ "grad_norm": 7.5625,
463
+ "learning_rate": 2.9095238095238097e-05,
464
+ "loss": 2.6809,
465
+ "step": 640
466
+ },
467
+ {
468
+ "epoch": 0.52,
469
+ "grad_norm": 8.3125,
470
+ "learning_rate": 2.8619047619047623e-05,
471
+ "loss": 2.8011,
472
+ "step": 650
473
+ },
474
+ {
475
+ "epoch": 0.528,
476
+ "grad_norm": 7.40625,
477
+ "learning_rate": 2.814285714285714e-05,
478
+ "loss": 2.7213,
479
+ "step": 660
480
+ },
481
+ {
482
+ "epoch": 0.536,
483
+ "grad_norm": 7.15625,
484
+ "learning_rate": 2.7666666666666667e-05,
485
+ "loss": 2.8342,
486
+ "step": 670
487
+ },
488
+ {
489
+ "epoch": 0.544,
490
+ "grad_norm": 6.625,
491
+ "learning_rate": 2.7190476190476193e-05,
492
+ "loss": 2.7829,
493
+ "step": 680
494
+ },
495
+ {
496
+ "epoch": 0.552,
497
+ "grad_norm": 7.5625,
498
+ "learning_rate": 2.6714285714285715e-05,
499
+ "loss": 2.7949,
500
+ "step": 690
501
+ },
502
+ {
503
+ "epoch": 0.56,
504
+ "grad_norm": 7.5625,
505
+ "learning_rate": 2.623809523809524e-05,
506
+ "loss": 2.7692,
507
+ "step": 700
508
+ },
509
+ {
510
+ "epoch": 0.568,
511
+ "grad_norm": 8.25,
512
+ "learning_rate": 2.5761904761904766e-05,
513
+ "loss": 2.8251,
514
+ "step": 710
515
+ },
516
+ {
517
+ "epoch": 0.576,
518
+ "grad_norm": 7.15625,
519
+ "learning_rate": 2.5285714285714285e-05,
520
+ "loss": 2.7056,
521
+ "step": 720
522
+ },
523
+ {
524
+ "epoch": 0.584,
525
+ "grad_norm": 8.8125,
526
+ "learning_rate": 2.480952380952381e-05,
527
+ "loss": 2.8116,
528
+ "step": 730
529
+ },
530
+ {
531
+ "epoch": 0.592,
532
+ "grad_norm": 8.125,
533
+ "learning_rate": 2.4333333333333336e-05,
534
+ "loss": 2.6844,
535
+ "step": 740
536
+ },
537
+ {
538
+ "epoch": 0.6,
539
+ "grad_norm": 8.5625,
540
+ "learning_rate": 2.385714285714286e-05,
541
+ "loss": 2.7657,
542
+ "step": 750
543
+ },
544
+ {
545
+ "epoch": 0.608,
546
+ "grad_norm": 8.1875,
547
+ "learning_rate": 2.338095238095238e-05,
548
+ "loss": 2.8522,
549
+ "step": 760
550
+ },
551
+ {
552
+ "epoch": 0.616,
553
+ "grad_norm": 7.71875,
554
+ "learning_rate": 2.2904761904761906e-05,
555
+ "loss": 2.8368,
556
+ "step": 770
557
+ },
558
+ {
559
+ "epoch": 0.624,
560
+ "grad_norm": 7.71875,
561
+ "learning_rate": 2.242857142857143e-05,
562
+ "loss": 2.6978,
563
+ "step": 780
564
+ },
565
+ {
566
+ "epoch": 0.632,
567
+ "grad_norm": 7.25,
568
+ "learning_rate": 2.195238095238095e-05,
569
+ "loss": 2.8267,
570
+ "step": 790
571
+ },
572
+ {
573
+ "epoch": 0.64,
574
+ "grad_norm": 8.1875,
575
+ "learning_rate": 2.1476190476190477e-05,
576
+ "loss": 2.7188,
577
+ "step": 800
578
+ },
579
+ {
580
+ "epoch": 0.648,
581
+ "grad_norm": 7.78125,
582
+ "learning_rate": 2.1e-05,
583
+ "loss": 2.8553,
584
+ "step": 810
585
+ },
586
+ {
587
+ "epoch": 0.656,
588
+ "grad_norm": 7.25,
589
+ "learning_rate": 2.0523809523809524e-05,
590
+ "loss": 2.675,
591
+ "step": 820
592
+ },
593
+ {
594
+ "epoch": 0.664,
595
+ "grad_norm": 6.9375,
596
+ "learning_rate": 2.004761904761905e-05,
597
+ "loss": 2.7901,
598
+ "step": 830
599
+ },
600
+ {
601
+ "epoch": 0.672,
602
+ "grad_norm": 8.4375,
603
+ "learning_rate": 1.9571428571428572e-05,
604
+ "loss": 2.6719,
605
+ "step": 840
606
+ },
607
+ {
608
+ "epoch": 0.68,
609
+ "grad_norm": 7.8125,
610
+ "learning_rate": 1.9095238095238098e-05,
611
+ "loss": 2.6436,
612
+ "step": 850
613
+ },
614
+ {
615
+ "epoch": 0.688,
616
+ "grad_norm": 7.71875,
617
+ "learning_rate": 1.861904761904762e-05,
618
+ "loss": 2.8045,
619
+ "step": 860
620
+ },
621
+ {
622
+ "epoch": 0.696,
623
+ "grad_norm": 8.75,
624
+ "learning_rate": 1.8142857142857146e-05,
625
+ "loss": 2.802,
626
+ "step": 870
627
+ },
628
+ {
629
+ "epoch": 0.704,
630
+ "grad_norm": 8.0,
631
+ "learning_rate": 1.7666666666666668e-05,
632
+ "loss": 2.807,
633
+ "step": 880
634
+ },
635
+ {
636
+ "epoch": 0.712,
637
+ "grad_norm": 8.75,
638
+ "learning_rate": 1.719047619047619e-05,
639
+ "loss": 2.7425,
640
+ "step": 890
641
+ },
642
+ {
643
+ "epoch": 0.72,
644
+ "grad_norm": 7.40625,
645
+ "learning_rate": 1.6714285714285716e-05,
646
+ "loss": 2.8515,
647
+ "step": 900
648
+ },
649
+ {
650
+ "epoch": 0.728,
651
+ "grad_norm": 7.0625,
652
+ "learning_rate": 1.6238095238095238e-05,
653
+ "loss": 2.7694,
654
+ "step": 910
655
+ },
656
+ {
657
+ "epoch": 0.736,
658
+ "grad_norm": 8.0,
659
+ "learning_rate": 1.5761904761904764e-05,
660
+ "loss": 2.6311,
661
+ "step": 920
662
+ },
663
+ {
664
+ "epoch": 0.744,
665
+ "grad_norm": 6.6875,
666
+ "learning_rate": 1.5285714285714286e-05,
667
+ "loss": 2.6884,
668
+ "step": 930
669
+ },
670
+ {
671
+ "epoch": 0.752,
672
+ "grad_norm": 6.9375,
673
+ "learning_rate": 1.480952380952381e-05,
674
+ "loss": 2.8331,
675
+ "step": 940
676
+ },
677
+ {
678
+ "epoch": 0.76,
679
+ "grad_norm": 7.75,
680
+ "learning_rate": 1.4333333333333334e-05,
681
+ "loss": 2.7617,
682
+ "step": 950
683
+ },
684
+ {
685
+ "epoch": 0.768,
686
+ "grad_norm": 7.1875,
687
+ "learning_rate": 1.3857142857142858e-05,
688
+ "loss": 2.6287,
689
+ "step": 960
690
+ },
691
+ {
692
+ "epoch": 0.776,
693
+ "grad_norm": 7.71875,
694
+ "learning_rate": 1.3380952380952383e-05,
695
+ "loss": 2.8381,
696
+ "step": 970
697
+ },
698
+ {
699
+ "epoch": 0.784,
700
+ "grad_norm": 7.625,
701
+ "learning_rate": 1.2904761904761906e-05,
702
+ "loss": 2.6136,
703
+ "step": 980
704
+ },
705
+ {
706
+ "epoch": 0.792,
707
+ "grad_norm": 7.40625,
708
+ "learning_rate": 1.242857142857143e-05,
709
+ "loss": 2.7682,
710
+ "step": 990
711
+ },
712
+ {
713
+ "epoch": 0.8,
714
+ "grad_norm": 7.71875,
715
+ "learning_rate": 1.1952380952380952e-05,
716
+ "loss": 2.6878,
717
+ "step": 1000
718
+ },
719
+ {
720
+ "epoch": 0.808,
721
+ "grad_norm": 8.0,
722
+ "learning_rate": 1.1476190476190476e-05,
723
+ "loss": 2.5709,
724
+ "step": 1010
725
+ },
726
+ {
727
+ "epoch": 0.816,
728
+ "grad_norm": 8.0625,
729
+ "learning_rate": 1.1000000000000001e-05,
730
+ "loss": 2.7941,
731
+ "step": 1020
732
+ },
733
+ {
734
+ "epoch": 0.824,
735
+ "grad_norm": 8.1875,
736
+ "learning_rate": 1.0523809523809525e-05,
737
+ "loss": 2.6739,
738
+ "step": 1030
739
+ },
740
+ {
741
+ "epoch": 0.832,
742
+ "grad_norm": 7.59375,
743
+ "learning_rate": 1.004761904761905e-05,
744
+ "loss": 2.6953,
745
+ "step": 1040
746
+ },
747
+ {
748
+ "epoch": 0.84,
749
+ "grad_norm": 7.3125,
750
+ "learning_rate": 9.571428571428572e-06,
751
+ "loss": 2.6107,
752
+ "step": 1050
753
+ },
754
+ {
755
+ "epoch": 0.848,
756
+ "grad_norm": 7.59375,
757
+ "learning_rate": 9.095238095238095e-06,
758
+ "loss": 2.7805,
759
+ "step": 1060
760
+ },
761
+ {
762
+ "epoch": 0.856,
763
+ "grad_norm": 9.0625,
764
+ "learning_rate": 8.61904761904762e-06,
765
+ "loss": 2.7856,
766
+ "step": 1070
767
+ },
768
+ {
769
+ "epoch": 0.864,
770
+ "grad_norm": 7.75,
771
+ "learning_rate": 8.142857142857143e-06,
772
+ "loss": 2.6762,
773
+ "step": 1080
774
+ },
775
+ {
776
+ "epoch": 0.872,
777
+ "grad_norm": 7.625,
778
+ "learning_rate": 7.666666666666667e-06,
779
+ "loss": 2.7825,
780
+ "step": 1090
781
+ },
782
+ {
783
+ "epoch": 0.88,
784
+ "grad_norm": 7.0,
785
+ "learning_rate": 7.19047619047619e-06,
786
+ "loss": 2.7893,
787
+ "step": 1100
788
+ },
789
+ {
790
+ "epoch": 0.888,
791
+ "grad_norm": 6.9375,
792
+ "learning_rate": 6.714285714285714e-06,
793
+ "loss": 2.7058,
794
+ "step": 1110
795
+ },
796
+ {
797
+ "epoch": 0.896,
798
+ "grad_norm": 7.03125,
799
+ "learning_rate": 6.238095238095239e-06,
800
+ "loss": 2.7899,
801
+ "step": 1120
802
+ },
803
+ {
804
+ "epoch": 0.904,
805
+ "grad_norm": 6.9375,
806
+ "learning_rate": 5.761904761904762e-06,
807
+ "loss": 2.6251,
808
+ "step": 1130
809
+ },
810
+ {
811
+ "epoch": 0.912,
812
+ "grad_norm": 6.84375,
813
+ "learning_rate": 5.285714285714286e-06,
814
+ "loss": 2.7011,
815
+ "step": 1140
816
+ },
817
+ {
818
+ "epoch": 0.92,
819
+ "grad_norm": 6.78125,
820
+ "learning_rate": 4.80952380952381e-06,
821
+ "loss": 2.6375,
822
+ "step": 1150
823
+ },
824
+ {
825
+ "epoch": 0.928,
826
+ "grad_norm": 8.5625,
827
+ "learning_rate": 4.333333333333334e-06,
828
+ "loss": 2.7994,
829
+ "step": 1160
830
+ },
831
+ {
832
+ "epoch": 0.936,
833
+ "grad_norm": 6.59375,
834
+ "learning_rate": 3.857142857142857e-06,
835
+ "loss": 2.7145,
836
+ "step": 1170
837
+ },
838
+ {
839
+ "epoch": 0.944,
840
+ "grad_norm": 7.0625,
841
+ "learning_rate": 3.3809523809523814e-06,
842
+ "loss": 2.6465,
843
+ "step": 1180
844
+ },
845
+ {
846
+ "epoch": 0.952,
847
+ "grad_norm": 8.3125,
848
+ "learning_rate": 2.904761904761905e-06,
849
+ "loss": 2.7466,
850
+ "step": 1190
851
+ },
852
+ {
853
+ "epoch": 0.96,
854
+ "grad_norm": 7.84375,
855
+ "learning_rate": 2.428571428571429e-06,
856
+ "loss": 2.76,
857
+ "step": 1200
858
+ },
859
+ {
860
+ "epoch": 0.968,
861
+ "grad_norm": 7.78125,
862
+ "learning_rate": 1.9523809523809523e-06,
863
+ "loss": 2.7507,
864
+ "step": 1210
865
+ },
866
+ {
867
+ "epoch": 0.976,
868
+ "grad_norm": 7.4375,
869
+ "learning_rate": 1.4761904761904762e-06,
870
+ "loss": 2.7197,
871
+ "step": 1220
872
+ },
873
+ {
874
+ "epoch": 0.984,
875
+ "grad_norm": 7.96875,
876
+ "learning_rate": 1.0000000000000002e-06,
877
+ "loss": 2.7261,
878
+ "step": 1230
879
+ },
880
+ {
881
+ "epoch": 0.992,
882
+ "grad_norm": 5.90625,
883
+ "learning_rate": 5.238095238095238e-07,
884
+ "loss": 2.6812,
885
+ "step": 1240
886
+ },
887
+ {
888
+ "epoch": 1.0,
889
+ "grad_norm": 8.125,
890
+ "learning_rate": 4.761904761904762e-08,
891
+ "loss": 2.6935,
892
+ "step": 1250
893
+ }
894
+ ],
895
+ "logging_steps": 10,
896
+ "max_steps": 1250,
897
+ "num_input_tokens_seen": 0,
898
+ "num_train_epochs": 1,
899
+ "save_steps": 500,
900
+ "stateful_callbacks": {
901
+ "TrainerControl": {
902
+ "args": {
903
+ "should_epoch_stop": false,
904
+ "should_evaluate": false,
905
+ "should_log": false,
906
+ "should_save": true,
907
+ "should_training_stop": true
908
+ },
909
+ "attributes": {}
910
+ }
911
+ },
912
+ "total_flos": 1.443927710242921e+17,
913
+ "train_batch_size": 2,
914
+ "trial_name": null,
915
+ "trial_params": null
916
+ }
checkpoint-1250/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7bfa58ce9748715a5711511426ed685dd546e89891028b84e29bedc4291ade9
3
+ size 5841
config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "GptOssForCausalLM"
4
+ ],
5
+ "attention_bias": true,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 199998,
8
+ "dtype": "bfloat16",
9
+ "eos_token_id": 200002,
10
+ "experts_per_token": 4,
11
+ "head_dim": 64,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 2880,
14
+ "initial_context_length": 4096,
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 2880,
17
+ "layer_types": [
18
+ "sliding_attention",
19
+ "full_attention",
20
+ "sliding_attention",
21
+ "full_attention",
22
+ "sliding_attention",
23
+ "full_attention",
24
+ "sliding_attention",
25
+ "full_attention",
26
+ "sliding_attention",
27
+ "full_attention",
28
+ "sliding_attention",
29
+ "full_attention",
30
+ "sliding_attention",
31
+ "full_attention",
32
+ "sliding_attention",
33
+ "full_attention",
34
+ "sliding_attention",
35
+ "full_attention",
36
+ "sliding_attention",
37
+ "full_attention",
38
+ "sliding_attention",
39
+ "full_attention",
40
+ "sliding_attention",
41
+ "full_attention"
42
+ ],
43
+ "max_position_embeddings": 131072,
44
+ "model_type": "gpt_oss",
45
+ "num_attention_heads": 64,
46
+ "num_experts_per_tok": 4,
47
+ "num_hidden_layers": 24,
48
+ "num_key_value_heads": 8,
49
+ "num_local_experts": 32,
50
+ "output_router_logits": false,
51
+ "pad_token_id": 199999,
52
+ "rms_norm_eps": 1e-05,
53
+ "rope_scaling": {
54
+ "beta_fast": 32.0,
55
+ "beta_slow": 1.0,
56
+ "factor": 32.0,
57
+ "original_max_position_embeddings": 4096,
58
+ "rope_type": "yarn",
59
+ "truncate": false
60
+ },
61
+ "rope_theta": 150000,
62
+ "router_aux_loss_coef": 0.9,
63
+ "sliding_window": 128,
64
+ "swiglu_limit": 7.0,
65
+ "tie_word_embeddings": false,
66
+ "transformers_version": "4.57.1",
67
+ "use_cache": false,
68
+ "vocab_size": 201088
69
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 199998,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 200002,
6
+ 199999,
7
+ 200012
8
+ ],
9
+ "pad_token_id": 199999,
10
+ "transformers_version": "4.57.1"
11
+ }
logs/events.out.tfevents.1763377278.idealistic-pudu-from-uranus-fa7e.10674.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccf40292de8fef165b1991f4da308ccf49c0bf864d5bcb0f5725b6e7ed67955d
3
+ size 5999
logs/events.out.tfevents.1763377589.idealistic-pudu-from-uranus-fa7e.11740.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d89dbdf99bc27167a894217e0a77b19fe1c893499fe9e3803a4602661a928cf
3
+ size 5999
logs/events.out.tfevents.1763377820.idealistic-pudu-from-uranus-fa7e.12716.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e81f935854e495713952658a7b2cbbe0e60ba4d72d0ac25c1b6cb5465b315d66
3
+ size 6000
logs/events.out.tfevents.1763377925.idealistic-pudu-from-uranus-fa7e.13187.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:992703685fa84a0079cdd95164b7dc3c47661631bee013f71313ec930f2fa9b1
3
+ size 6207
logs/events.out.tfevents.1763378147.idealistic-pudu-from-uranus-fa7e.14024.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:665e62bf2af7ef70f55c4289454ca80d1fda2ae476f46a775fc72ee0286aa461
3
+ size 6207
logs/events.out.tfevents.1763378228.idealistic-pudu-from-uranus-fa7e.14429.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71e605388bcafaf25ce434a0ceb99de7b3a37c10b4eb491f342db94047ae2965
3
+ size 32888
model-00001-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc729189133161b61e287dd7bdec06afc2b7f77ed1820ca53eda0553fa742230
3
+ size 4504304664
model-00002-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4708b3da84cd7d559f5cd2eb9d0d13466ae1ccd90fdee8295d9ec3d8748582c
3
+ size 4939127656
model-00003-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96031fa4efe1e880b2f17ac4352c3a2d251129d8140d122b9e8649a2d45c51d2
3
+ size 4939127656
model-00004-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46e221c089ddcbacd6f4495a2dfccea2206cb8b21795ea868415125e9fb8139e
3
+ size 4939127680
model-00005-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f223ae981c1c8f4925334b8e0954ce78a613b0602d3eba28ea73f0146c9d73f3
3
+ size 4939127704
model-00006-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf0e3a0ccf931369e3ebe77f1f6d0a1781d739d6cb1dfed7c54fedda3f94fe18
3
+ size 4939127704
model-00007-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51a3ad356e1a0d63de63f78975caf1a129d7d29bc00a7614de652aff7a4c57b0
3
+ size 4939127704
model-00008-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07a2d79388f1f6f1bebac3d03216b6ac47c2bf498640ab54ac8c15beb629847f
3
+ size 4939127704
model-00009-of-00009.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eda29cdc86d75937066cde98574795f08a96ebe34b22e126e1cb252ec494df4
3
+ size 2751362856
model.safetensors.index.json ADDED
@@ -0,0 +1,419 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_parameters": 20914757184,
4
+ "total_size": 41829514368
5
+ },
6
+ "weight_map": {
7
+ "lm_head.weight": "model-00009-of-00009.safetensors",
8
+ "model.embed_tokens.weight": "model-00001-of-00009.safetensors",
9
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00009.safetensors",
10
+ "model.layers.0.mlp.experts.down_proj": "model-00001-of-00009.safetensors",
11
+ "model.layers.0.mlp.experts.down_proj_bias": "model-00001-of-00009.safetensors",
12
+ "model.layers.0.mlp.experts.gate_up_proj": "model-00001-of-00009.safetensors",
13
+ "model.layers.0.mlp.experts.gate_up_proj_bias": "model-00001-of-00009.safetensors",
14
+ "model.layers.0.mlp.router.bias": "model-00001-of-00009.safetensors",
15
+ "model.layers.0.mlp.router.weight": "model-00001-of-00009.safetensors",
16
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00009.safetensors",
17
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00009.safetensors",
18
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00009.safetensors",
19
+ "model.layers.0.self_attn.o_proj.bias": "model-00001-of-00009.safetensors",
20
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00009.safetensors",
21
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00009.safetensors",
22
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00009.safetensors",
23
+ "model.layers.0.self_attn.sinks": "model-00001-of-00009.safetensors",
24
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00009.safetensors",
25
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00009.safetensors",
26
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00009.safetensors",
27
+ "model.layers.1.mlp.experts.down_proj": "model-00001-of-00009.safetensors",
28
+ "model.layers.1.mlp.experts.down_proj_bias": "model-00001-of-00009.safetensors",
29
+ "model.layers.1.mlp.experts.gate_up_proj": "model-00001-of-00009.safetensors",
30
+ "model.layers.1.mlp.experts.gate_up_proj_bias": "model-00001-of-00009.safetensors",
31
+ "model.layers.1.mlp.router.bias": "model-00001-of-00009.safetensors",
32
+ "model.layers.1.mlp.router.weight": "model-00001-of-00009.safetensors",
33
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00009.safetensors",
34
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00009.safetensors",
35
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00009.safetensors",
36
+ "model.layers.1.self_attn.o_proj.bias": "model-00001-of-00009.safetensors",
37
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00009.safetensors",
38
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00009.safetensors",
39
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00009.safetensors",
40
+ "model.layers.1.self_attn.sinks": "model-00001-of-00009.safetensors",
41
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00009.safetensors",
42
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00009.safetensors",
43
+ "model.layers.10.input_layernorm.weight": "model-00004-of-00009.safetensors",
44
+ "model.layers.10.mlp.experts.down_proj": "model-00004-of-00009.safetensors",
45
+ "model.layers.10.mlp.experts.down_proj_bias": "model-00004-of-00009.safetensors",
46
+ "model.layers.10.mlp.experts.gate_up_proj": "model-00004-of-00009.safetensors",
47
+ "model.layers.10.mlp.experts.gate_up_proj_bias": "model-00004-of-00009.safetensors",
48
+ "model.layers.10.mlp.router.bias": "model-00004-of-00009.safetensors",
49
+ "model.layers.10.mlp.router.weight": "model-00004-of-00009.safetensors",
50
+ "model.layers.10.post_attention_layernorm.weight": "model-00004-of-00009.safetensors",
51
+ "model.layers.10.self_attn.k_proj.bias": "model-00004-of-00009.safetensors",
52
+ "model.layers.10.self_attn.k_proj.weight": "model-00004-of-00009.safetensors",
53
+ "model.layers.10.self_attn.o_proj.bias": "model-00004-of-00009.safetensors",
54
+ "model.layers.10.self_attn.o_proj.weight": "model-00004-of-00009.safetensors",
55
+ "model.layers.10.self_attn.q_proj.bias": "model-00004-of-00009.safetensors",
56
+ "model.layers.10.self_attn.q_proj.weight": "model-00004-of-00009.safetensors",
57
+ "model.layers.10.self_attn.sinks": "model-00004-of-00009.safetensors",
58
+ "model.layers.10.self_attn.v_proj.bias": "model-00004-of-00009.safetensors",
59
+ "model.layers.10.self_attn.v_proj.weight": "model-00004-of-00009.safetensors",
60
+ "model.layers.11.input_layernorm.weight": "model-00005-of-00009.safetensors",
61
+ "model.layers.11.mlp.experts.down_proj": "model-00005-of-00009.safetensors",
62
+ "model.layers.11.mlp.experts.down_proj_bias": "model-00005-of-00009.safetensors",
63
+ "model.layers.11.mlp.experts.gate_up_proj": "model-00005-of-00009.safetensors",
64
+ "model.layers.11.mlp.experts.gate_up_proj_bias": "model-00005-of-00009.safetensors",
65
+ "model.layers.11.mlp.router.bias": "model-00004-of-00009.safetensors",
66
+ "model.layers.11.mlp.router.weight": "model-00004-of-00009.safetensors",
67
+ "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00009.safetensors",
68
+ "model.layers.11.self_attn.k_proj.bias": "model-00004-of-00009.safetensors",
69
+ "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00009.safetensors",
70
+ "model.layers.11.self_attn.o_proj.bias": "model-00004-of-00009.safetensors",
71
+ "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00009.safetensors",
72
+ "model.layers.11.self_attn.q_proj.bias": "model-00004-of-00009.safetensors",
73
+ "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00009.safetensors",
74
+ "model.layers.11.self_attn.sinks": "model-00004-of-00009.safetensors",
75
+ "model.layers.11.self_attn.v_proj.bias": "model-00004-of-00009.safetensors",
76
+ "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00009.safetensors",
77
+ "model.layers.12.input_layernorm.weight": "model-00005-of-00009.safetensors",
78
+ "model.layers.12.mlp.experts.down_proj": "model-00005-of-00009.safetensors",
79
+ "model.layers.12.mlp.experts.down_proj_bias": "model-00005-of-00009.safetensors",
80
+ "model.layers.12.mlp.experts.gate_up_proj": "model-00005-of-00009.safetensors",
81
+ "model.layers.12.mlp.experts.gate_up_proj_bias": "model-00005-of-00009.safetensors",
82
+ "model.layers.12.mlp.router.bias": "model-00005-of-00009.safetensors",
83
+ "model.layers.12.mlp.router.weight": "model-00005-of-00009.safetensors",
84
+ "model.layers.12.post_attention_layernorm.weight": "model-00005-of-00009.safetensors",
85
+ "model.layers.12.self_attn.k_proj.bias": "model-00005-of-00009.safetensors",
86
+ "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00009.safetensors",
87
+ "model.layers.12.self_attn.o_proj.bias": "model-00005-of-00009.safetensors",
88
+ "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00009.safetensors",
89
+ "model.layers.12.self_attn.q_proj.bias": "model-00005-of-00009.safetensors",
90
+ "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00009.safetensors",
91
+ "model.layers.12.self_attn.sinks": "model-00005-of-00009.safetensors",
92
+ "model.layers.12.self_attn.v_proj.bias": "model-00005-of-00009.safetensors",
93
+ "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00009.safetensors",
94
+ "model.layers.13.input_layernorm.weight": "model-00005-of-00009.safetensors",
95
+ "model.layers.13.mlp.experts.down_proj": "model-00005-of-00009.safetensors",
96
+ "model.layers.13.mlp.experts.down_proj_bias": "model-00005-of-00009.safetensors",
97
+ "model.layers.13.mlp.experts.gate_up_proj": "model-00005-of-00009.safetensors",
98
+ "model.layers.13.mlp.experts.gate_up_proj_bias": "model-00005-of-00009.safetensors",
99
+ "model.layers.13.mlp.router.bias": "model-00005-of-00009.safetensors",
100
+ "model.layers.13.mlp.router.weight": "model-00005-of-00009.safetensors",
101
+ "model.layers.13.post_attention_layernorm.weight": "model-00005-of-00009.safetensors",
102
+ "model.layers.13.self_attn.k_proj.bias": "model-00005-of-00009.safetensors",
103
+ "model.layers.13.self_attn.k_proj.weight": "model-00005-of-00009.safetensors",
104
+ "model.layers.13.self_attn.o_proj.bias": "model-00005-of-00009.safetensors",
105
+ "model.layers.13.self_attn.o_proj.weight": "model-00005-of-00009.safetensors",
106
+ "model.layers.13.self_attn.q_proj.bias": "model-00005-of-00009.safetensors",
107
+ "model.layers.13.self_attn.q_proj.weight": "model-00005-of-00009.safetensors",
108
+ "model.layers.13.self_attn.sinks": "model-00005-of-00009.safetensors",
109
+ "model.layers.13.self_attn.v_proj.bias": "model-00005-of-00009.safetensors",
110
+ "model.layers.13.self_attn.v_proj.weight": "model-00005-of-00009.safetensors",
111
+ "model.layers.14.input_layernorm.weight": "model-00006-of-00009.safetensors",
112
+ "model.layers.14.mlp.experts.down_proj": "model-00006-of-00009.safetensors",
113
+ "model.layers.14.mlp.experts.down_proj_bias": "model-00006-of-00009.safetensors",
114
+ "model.layers.14.mlp.experts.gate_up_proj": "model-00006-of-00009.safetensors",
115
+ "model.layers.14.mlp.experts.gate_up_proj_bias": "model-00006-of-00009.safetensors",
116
+ "model.layers.14.mlp.router.bias": "model-00005-of-00009.safetensors",
117
+ "model.layers.14.mlp.router.weight": "model-00005-of-00009.safetensors",
118
+ "model.layers.14.post_attention_layernorm.weight": "model-00006-of-00009.safetensors",
119
+ "model.layers.14.self_attn.k_proj.bias": "model-00005-of-00009.safetensors",
120
+ "model.layers.14.self_attn.k_proj.weight": "model-00005-of-00009.safetensors",
121
+ "model.layers.14.self_attn.o_proj.bias": "model-00005-of-00009.safetensors",
122
+ "model.layers.14.self_attn.o_proj.weight": "model-00005-of-00009.safetensors",
123
+ "model.layers.14.self_attn.q_proj.bias": "model-00005-of-00009.safetensors",
124
+ "model.layers.14.self_attn.q_proj.weight": "model-00005-of-00009.safetensors",
125
+ "model.layers.14.self_attn.sinks": "model-00005-of-00009.safetensors",
126
+ "model.layers.14.self_attn.v_proj.bias": "model-00005-of-00009.safetensors",
127
+ "model.layers.14.self_attn.v_proj.weight": "model-00005-of-00009.safetensors",
128
+ "model.layers.15.input_layernorm.weight": "model-00006-of-00009.safetensors",
129
+ "model.layers.15.mlp.experts.down_proj": "model-00006-of-00009.safetensors",
130
+ "model.layers.15.mlp.experts.down_proj_bias": "model-00006-of-00009.safetensors",
131
+ "model.layers.15.mlp.experts.gate_up_proj": "model-00006-of-00009.safetensors",
132
+ "model.layers.15.mlp.experts.gate_up_proj_bias": "model-00006-of-00009.safetensors",
133
+ "model.layers.15.mlp.router.bias": "model-00006-of-00009.safetensors",
134
+ "model.layers.15.mlp.router.weight": "model-00006-of-00009.safetensors",
135
+ "model.layers.15.post_attention_layernorm.weight": "model-00006-of-00009.safetensors",
136
+ "model.layers.15.self_attn.k_proj.bias": "model-00006-of-00009.safetensors",
137
+ "model.layers.15.self_attn.k_proj.weight": "model-00006-of-00009.safetensors",
138
+ "model.layers.15.self_attn.o_proj.bias": "model-00006-of-00009.safetensors",
139
+ "model.layers.15.self_attn.o_proj.weight": "model-00006-of-00009.safetensors",
140
+ "model.layers.15.self_attn.q_proj.bias": "model-00006-of-00009.safetensors",
141
+ "model.layers.15.self_attn.q_proj.weight": "model-00006-of-00009.safetensors",
142
+ "model.layers.15.self_attn.sinks": "model-00006-of-00009.safetensors",
143
+ "model.layers.15.self_attn.v_proj.bias": "model-00006-of-00009.safetensors",
144
+ "model.layers.15.self_attn.v_proj.weight": "model-00006-of-00009.safetensors",
145
+ "model.layers.16.input_layernorm.weight": "model-00006-of-00009.safetensors",
146
+ "model.layers.16.mlp.experts.down_proj": "model-00006-of-00009.safetensors",
147
+ "model.layers.16.mlp.experts.down_proj_bias": "model-00006-of-00009.safetensors",
148
+ "model.layers.16.mlp.experts.gate_up_proj": "model-00006-of-00009.safetensors",
149
+ "model.layers.16.mlp.experts.gate_up_proj_bias": "model-00006-of-00009.safetensors",
150
+ "model.layers.16.mlp.router.bias": "model-00006-of-00009.safetensors",
151
+ "model.layers.16.mlp.router.weight": "model-00006-of-00009.safetensors",
152
+ "model.layers.16.post_attention_layernorm.weight": "model-00006-of-00009.safetensors",
153
+ "model.layers.16.self_attn.k_proj.bias": "model-00006-of-00009.safetensors",
154
+ "model.layers.16.self_attn.k_proj.weight": "model-00006-of-00009.safetensors",
155
+ "model.layers.16.self_attn.o_proj.bias": "model-00006-of-00009.safetensors",
156
+ "model.layers.16.self_attn.o_proj.weight": "model-00006-of-00009.safetensors",
157
+ "model.layers.16.self_attn.q_proj.bias": "model-00006-of-00009.safetensors",
158
+ "model.layers.16.self_attn.q_proj.weight": "model-00006-of-00009.safetensors",
159
+ "model.layers.16.self_attn.sinks": "model-00006-of-00009.safetensors",
160
+ "model.layers.16.self_attn.v_proj.bias": "model-00006-of-00009.safetensors",
161
+ "model.layers.16.self_attn.v_proj.weight": "model-00006-of-00009.safetensors",
162
+ "model.layers.17.input_layernorm.weight": "model-00007-of-00009.safetensors",
163
+ "model.layers.17.mlp.experts.down_proj": "model-00007-of-00009.safetensors",
164
+ "model.layers.17.mlp.experts.down_proj_bias": "model-00007-of-00009.safetensors",
165
+ "model.layers.17.mlp.experts.gate_up_proj": "model-00007-of-00009.safetensors",
166
+ "model.layers.17.mlp.experts.gate_up_proj_bias": "model-00007-of-00009.safetensors",
167
+ "model.layers.17.mlp.router.bias": "model-00006-of-00009.safetensors",
168
+ "model.layers.17.mlp.router.weight": "model-00006-of-00009.safetensors",
169
+ "model.layers.17.post_attention_layernorm.weight": "model-00007-of-00009.safetensors",
170
+ "model.layers.17.self_attn.k_proj.bias": "model-00006-of-00009.safetensors",
171
+ "model.layers.17.self_attn.k_proj.weight": "model-00006-of-00009.safetensors",
172
+ "model.layers.17.self_attn.o_proj.bias": "model-00006-of-00009.safetensors",
173
+ "model.layers.17.self_attn.o_proj.weight": "model-00006-of-00009.safetensors",
174
+ "model.layers.17.self_attn.q_proj.bias": "model-00006-of-00009.safetensors",
175
+ "model.layers.17.self_attn.q_proj.weight": "model-00006-of-00009.safetensors",
176
+ "model.layers.17.self_attn.sinks": "model-00006-of-00009.safetensors",
177
+ "model.layers.17.self_attn.v_proj.bias": "model-00006-of-00009.safetensors",
178
+ "model.layers.17.self_attn.v_proj.weight": "model-00006-of-00009.safetensors",
179
+ "model.layers.18.input_layernorm.weight": "model-00007-of-00009.safetensors",
180
+ "model.layers.18.mlp.experts.down_proj": "model-00007-of-00009.safetensors",
181
+ "model.layers.18.mlp.experts.down_proj_bias": "model-00007-of-00009.safetensors",
182
+ "model.layers.18.mlp.experts.gate_up_proj": "model-00007-of-00009.safetensors",
183
+ "model.layers.18.mlp.experts.gate_up_proj_bias": "model-00007-of-00009.safetensors",
184
+ "model.layers.18.mlp.router.bias": "model-00007-of-00009.safetensors",
185
+ "model.layers.18.mlp.router.weight": "model-00007-of-00009.safetensors",
186
+ "model.layers.18.post_attention_layernorm.weight": "model-00007-of-00009.safetensors",
187
+ "model.layers.18.self_attn.k_proj.bias": "model-00007-of-00009.safetensors",
188
+ "model.layers.18.self_attn.k_proj.weight": "model-00007-of-00009.safetensors",
189
+ "model.layers.18.self_attn.o_proj.bias": "model-00007-of-00009.safetensors",
190
+ "model.layers.18.self_attn.o_proj.weight": "model-00007-of-00009.safetensors",
191
+ "model.layers.18.self_attn.q_proj.bias": "model-00007-of-00009.safetensors",
192
+ "model.layers.18.self_attn.q_proj.weight": "model-00007-of-00009.safetensors",
193
+ "model.layers.18.self_attn.sinks": "model-00007-of-00009.safetensors",
194
+ "model.layers.18.self_attn.v_proj.bias": "model-00007-of-00009.safetensors",
195
+ "model.layers.18.self_attn.v_proj.weight": "model-00007-of-00009.safetensors",
196
+ "model.layers.19.input_layernorm.weight": "model-00007-of-00009.safetensors",
197
+ "model.layers.19.mlp.experts.down_proj": "model-00007-of-00009.safetensors",
198
+ "model.layers.19.mlp.experts.down_proj_bias": "model-00007-of-00009.safetensors",
199
+ "model.layers.19.mlp.experts.gate_up_proj": "model-00007-of-00009.safetensors",
200
+ "model.layers.19.mlp.experts.gate_up_proj_bias": "model-00007-of-00009.safetensors",
201
+ "model.layers.19.mlp.router.bias": "model-00007-of-00009.safetensors",
202
+ "model.layers.19.mlp.router.weight": "model-00007-of-00009.safetensors",
203
+ "model.layers.19.post_attention_layernorm.weight": "model-00007-of-00009.safetensors",
204
+ "model.layers.19.self_attn.k_proj.bias": "model-00007-of-00009.safetensors",
205
+ "model.layers.19.self_attn.k_proj.weight": "model-00007-of-00009.safetensors",
206
+ "model.layers.19.self_attn.o_proj.bias": "model-00007-of-00009.safetensors",
207
+ "model.layers.19.self_attn.o_proj.weight": "model-00007-of-00009.safetensors",
208
+ "model.layers.19.self_attn.q_proj.bias": "model-00007-of-00009.safetensors",
209
+ "model.layers.19.self_attn.q_proj.weight": "model-00007-of-00009.safetensors",
210
+ "model.layers.19.self_attn.sinks": "model-00007-of-00009.safetensors",
211
+ "model.layers.19.self_attn.v_proj.bias": "model-00007-of-00009.safetensors",
212
+ "model.layers.19.self_attn.v_proj.weight": "model-00007-of-00009.safetensors",
213
+ "model.layers.2.input_layernorm.weight": "model-00002-of-00009.safetensors",
214
+ "model.layers.2.mlp.experts.down_proj": "model-00002-of-00009.safetensors",
215
+ "model.layers.2.mlp.experts.down_proj_bias": "model-00002-of-00009.safetensors",
216
+ "model.layers.2.mlp.experts.gate_up_proj": "model-00002-of-00009.safetensors",
217
+ "model.layers.2.mlp.experts.gate_up_proj_bias": "model-00002-of-00009.safetensors",
218
+ "model.layers.2.mlp.router.bias": "model-00001-of-00009.safetensors",
219
+ "model.layers.2.mlp.router.weight": "model-00001-of-00009.safetensors",
220
+ "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00009.safetensors",
221
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00009.safetensors",
222
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00009.safetensors",
223
+ "model.layers.2.self_attn.o_proj.bias": "model-00001-of-00009.safetensors",
224
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00009.safetensors",
225
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00009.safetensors",
226
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00009.safetensors",
227
+ "model.layers.2.self_attn.sinks": "model-00001-of-00009.safetensors",
228
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00009.safetensors",
229
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00009.safetensors",
230
+ "model.layers.20.input_layernorm.weight": "model-00008-of-00009.safetensors",
231
+ "model.layers.20.mlp.experts.down_proj": "model-00008-of-00009.safetensors",
232
+ "model.layers.20.mlp.experts.down_proj_bias": "model-00008-of-00009.safetensors",
233
+ "model.layers.20.mlp.experts.gate_up_proj": "model-00008-of-00009.safetensors",
234
+ "model.layers.20.mlp.experts.gate_up_proj_bias": "model-00008-of-00009.safetensors",
235
+ "model.layers.20.mlp.router.bias": "model-00007-of-00009.safetensors",
236
+ "model.layers.20.mlp.router.weight": "model-00007-of-00009.safetensors",
237
+ "model.layers.20.post_attention_layernorm.weight": "model-00008-of-00009.safetensors",
238
+ "model.layers.20.self_attn.k_proj.bias": "model-00007-of-00009.safetensors",
239
+ "model.layers.20.self_attn.k_proj.weight": "model-00007-of-00009.safetensors",
240
+ "model.layers.20.self_attn.o_proj.bias": "model-00007-of-00009.safetensors",
241
+ "model.layers.20.self_attn.o_proj.weight": "model-00007-of-00009.safetensors",
242
+ "model.layers.20.self_attn.q_proj.bias": "model-00007-of-00009.safetensors",
243
+ "model.layers.20.self_attn.q_proj.weight": "model-00007-of-00009.safetensors",
244
+ "model.layers.20.self_attn.sinks": "model-00007-of-00009.safetensors",
245
+ "model.layers.20.self_attn.v_proj.bias": "model-00007-of-00009.safetensors",
246
+ "model.layers.20.self_attn.v_proj.weight": "model-00007-of-00009.safetensors",
247
+ "model.layers.21.input_layernorm.weight": "model-00008-of-00009.safetensors",
248
+ "model.layers.21.mlp.experts.down_proj": "model-00008-of-00009.safetensors",
249
+ "model.layers.21.mlp.experts.down_proj_bias": "model-00008-of-00009.safetensors",
250
+ "model.layers.21.mlp.experts.gate_up_proj": "model-00008-of-00009.safetensors",
251
+ "model.layers.21.mlp.experts.gate_up_proj_bias": "model-00008-of-00009.safetensors",
252
+ "model.layers.21.mlp.router.bias": "model-00008-of-00009.safetensors",
253
+ "model.layers.21.mlp.router.weight": "model-00008-of-00009.safetensors",
254
+ "model.layers.21.post_attention_layernorm.weight": "model-00008-of-00009.safetensors",
255
+ "model.layers.21.self_attn.k_proj.bias": "model-00008-of-00009.safetensors",
256
+ "model.layers.21.self_attn.k_proj.weight": "model-00008-of-00009.safetensors",
257
+ "model.layers.21.self_attn.o_proj.bias": "model-00008-of-00009.safetensors",
258
+ "model.layers.21.self_attn.o_proj.weight": "model-00008-of-00009.safetensors",
259
+ "model.layers.21.self_attn.q_proj.bias": "model-00008-of-00009.safetensors",
260
+ "model.layers.21.self_attn.q_proj.weight": "model-00008-of-00009.safetensors",
261
+ "model.layers.21.self_attn.sinks": "model-00008-of-00009.safetensors",
262
+ "model.layers.21.self_attn.v_proj.bias": "model-00008-of-00009.safetensors",
263
+ "model.layers.21.self_attn.v_proj.weight": "model-00008-of-00009.safetensors",
264
+ "model.layers.22.input_layernorm.weight": "model-00008-of-00009.safetensors",
265
+ "model.layers.22.mlp.experts.down_proj": "model-00008-of-00009.safetensors",
266
+ "model.layers.22.mlp.experts.down_proj_bias": "model-00008-of-00009.safetensors",
267
+ "model.layers.22.mlp.experts.gate_up_proj": "model-00008-of-00009.safetensors",
268
+ "model.layers.22.mlp.experts.gate_up_proj_bias": "model-00008-of-00009.safetensors",
269
+ "model.layers.22.mlp.router.bias": "model-00008-of-00009.safetensors",
270
+ "model.layers.22.mlp.router.weight": "model-00008-of-00009.safetensors",
271
+ "model.layers.22.post_attention_layernorm.weight": "model-00008-of-00009.safetensors",
272
+ "model.layers.22.self_attn.k_proj.bias": "model-00008-of-00009.safetensors",
273
+ "model.layers.22.self_attn.k_proj.weight": "model-00008-of-00009.safetensors",
274
+ "model.layers.22.self_attn.o_proj.bias": "model-00008-of-00009.safetensors",
275
+ "model.layers.22.self_attn.o_proj.weight": "model-00008-of-00009.safetensors",
276
+ "model.layers.22.self_attn.q_proj.bias": "model-00008-of-00009.safetensors",
277
+ "model.layers.22.self_attn.q_proj.weight": "model-00008-of-00009.safetensors",
278
+ "model.layers.22.self_attn.sinks": "model-00008-of-00009.safetensors",
279
+ "model.layers.22.self_attn.v_proj.bias": "model-00008-of-00009.safetensors",
280
+ "model.layers.22.self_attn.v_proj.weight": "model-00008-of-00009.safetensors",
281
+ "model.layers.23.input_layernorm.weight": "model-00009-of-00009.safetensors",
282
+ "model.layers.23.mlp.experts.down_proj": "model-00009-of-00009.safetensors",
283
+ "model.layers.23.mlp.experts.down_proj_bias": "model-00009-of-00009.safetensors",
284
+ "model.layers.23.mlp.experts.gate_up_proj": "model-00009-of-00009.safetensors",
285
+ "model.layers.23.mlp.experts.gate_up_proj_bias": "model-00009-of-00009.safetensors",
286
+ "model.layers.23.mlp.router.bias": "model-00008-of-00009.safetensors",
287
+ "model.layers.23.mlp.router.weight": "model-00008-of-00009.safetensors",
288
+ "model.layers.23.post_attention_layernorm.weight": "model-00009-of-00009.safetensors",
289
+ "model.layers.23.self_attn.k_proj.bias": "model-00008-of-00009.safetensors",
290
+ "model.layers.23.self_attn.k_proj.weight": "model-00008-of-00009.safetensors",
291
+ "model.layers.23.self_attn.o_proj.bias": "model-00008-of-00009.safetensors",
292
+ "model.layers.23.self_attn.o_proj.weight": "model-00008-of-00009.safetensors",
293
+ "model.layers.23.self_attn.q_proj.bias": "model-00008-of-00009.safetensors",
294
+ "model.layers.23.self_attn.q_proj.weight": "model-00008-of-00009.safetensors",
295
+ "model.layers.23.self_attn.sinks": "model-00008-of-00009.safetensors",
296
+ "model.layers.23.self_attn.v_proj.bias": "model-00008-of-00009.safetensors",
297
+ "model.layers.23.self_attn.v_proj.weight": "model-00008-of-00009.safetensors",
298
+ "model.layers.3.input_layernorm.weight": "model-00002-of-00009.safetensors",
299
+ "model.layers.3.mlp.experts.down_proj": "model-00002-of-00009.safetensors",
300
+ "model.layers.3.mlp.experts.down_proj_bias": "model-00002-of-00009.safetensors",
301
+ "model.layers.3.mlp.experts.gate_up_proj": "model-00002-of-00009.safetensors",
302
+ "model.layers.3.mlp.experts.gate_up_proj_bias": "model-00002-of-00009.safetensors",
303
+ "model.layers.3.mlp.router.bias": "model-00002-of-00009.safetensors",
304
+ "model.layers.3.mlp.router.weight": "model-00002-of-00009.safetensors",
305
+ "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00009.safetensors",
306
+ "model.layers.3.self_attn.k_proj.bias": "model-00002-of-00009.safetensors",
307
+ "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00009.safetensors",
308
+ "model.layers.3.self_attn.o_proj.bias": "model-00002-of-00009.safetensors",
309
+ "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00009.safetensors",
310
+ "model.layers.3.self_attn.q_proj.bias": "model-00002-of-00009.safetensors",
311
+ "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00009.safetensors",
312
+ "model.layers.3.self_attn.sinks": "model-00002-of-00009.safetensors",
313
+ "model.layers.3.self_attn.v_proj.bias": "model-00002-of-00009.safetensors",
314
+ "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00009.safetensors",
315
+ "model.layers.4.input_layernorm.weight": "model-00002-of-00009.safetensors",
316
+ "model.layers.4.mlp.experts.down_proj": "model-00002-of-00009.safetensors",
317
+ "model.layers.4.mlp.experts.down_proj_bias": "model-00002-of-00009.safetensors",
318
+ "model.layers.4.mlp.experts.gate_up_proj": "model-00002-of-00009.safetensors",
319
+ "model.layers.4.mlp.experts.gate_up_proj_bias": "model-00002-of-00009.safetensors",
320
+ "model.layers.4.mlp.router.bias": "model-00002-of-00009.safetensors",
321
+ "model.layers.4.mlp.router.weight": "model-00002-of-00009.safetensors",
322
+ "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00009.safetensors",
323
+ "model.layers.4.self_attn.k_proj.bias": "model-00002-of-00009.safetensors",
324
+ "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00009.safetensors",
325
+ "model.layers.4.self_attn.o_proj.bias": "model-00002-of-00009.safetensors",
326
+ "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00009.safetensors",
327
+ "model.layers.4.self_attn.q_proj.bias": "model-00002-of-00009.safetensors",
328
+ "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00009.safetensors",
329
+ "model.layers.4.self_attn.sinks": "model-00002-of-00009.safetensors",
330
+ "model.layers.4.self_attn.v_proj.bias": "model-00002-of-00009.safetensors",
331
+ "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00009.safetensors",
332
+ "model.layers.5.input_layernorm.weight": "model-00003-of-00009.safetensors",
333
+ "model.layers.5.mlp.experts.down_proj": "model-00003-of-00009.safetensors",
334
+ "model.layers.5.mlp.experts.down_proj_bias": "model-00003-of-00009.safetensors",
335
+ "model.layers.5.mlp.experts.gate_up_proj": "model-00003-of-00009.safetensors",
336
+ "model.layers.5.mlp.experts.gate_up_proj_bias": "model-00003-of-00009.safetensors",
337
+ "model.layers.5.mlp.router.bias": "model-00002-of-00009.safetensors",
338
+ "model.layers.5.mlp.router.weight": "model-00002-of-00009.safetensors",
339
+ "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00009.safetensors",
340
+ "model.layers.5.self_attn.k_proj.bias": "model-00002-of-00009.safetensors",
341
+ "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00009.safetensors",
342
+ "model.layers.5.self_attn.o_proj.bias": "model-00002-of-00009.safetensors",
343
+ "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00009.safetensors",
344
+ "model.layers.5.self_attn.q_proj.bias": "model-00002-of-00009.safetensors",
345
+ "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00009.safetensors",
346
+ "model.layers.5.self_attn.sinks": "model-00002-of-00009.safetensors",
347
+ "model.layers.5.self_attn.v_proj.bias": "model-00002-of-00009.safetensors",
348
+ "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00009.safetensors",
349
+ "model.layers.6.input_layernorm.weight": "model-00003-of-00009.safetensors",
350
+ "model.layers.6.mlp.experts.down_proj": "model-00003-of-00009.safetensors",
351
+ "model.layers.6.mlp.experts.down_proj_bias": "model-00003-of-00009.safetensors",
352
+ "model.layers.6.mlp.experts.gate_up_proj": "model-00003-of-00009.safetensors",
353
+ "model.layers.6.mlp.experts.gate_up_proj_bias": "model-00003-of-00009.safetensors",
354
+ "model.layers.6.mlp.router.bias": "model-00003-of-00009.safetensors",
355
+ "model.layers.6.mlp.router.weight": "model-00003-of-00009.safetensors",
356
+ "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00009.safetensors",
357
+ "model.layers.6.self_attn.k_proj.bias": "model-00003-of-00009.safetensors",
358
+ "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00009.safetensors",
359
+ "model.layers.6.self_attn.o_proj.bias": "model-00003-of-00009.safetensors",
360
+ "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00009.safetensors",
361
+ "model.layers.6.self_attn.q_proj.bias": "model-00003-of-00009.safetensors",
362
+ "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00009.safetensors",
363
+ "model.layers.6.self_attn.sinks": "model-00003-of-00009.safetensors",
364
+ "model.layers.6.self_attn.v_proj.bias": "model-00003-of-00009.safetensors",
365
+ "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00009.safetensors",
366
+ "model.layers.7.input_layernorm.weight": "model-00003-of-00009.safetensors",
367
+ "model.layers.7.mlp.experts.down_proj": "model-00003-of-00009.safetensors",
368
+ "model.layers.7.mlp.experts.down_proj_bias": "model-00003-of-00009.safetensors",
369
+ "model.layers.7.mlp.experts.gate_up_proj": "model-00003-of-00009.safetensors",
370
+ "model.layers.7.mlp.experts.gate_up_proj_bias": "model-00003-of-00009.safetensors",
371
+ "model.layers.7.mlp.router.bias": "model-00003-of-00009.safetensors",
372
+ "model.layers.7.mlp.router.weight": "model-00003-of-00009.safetensors",
373
+ "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00009.safetensors",
374
+ "model.layers.7.self_attn.k_proj.bias": "model-00003-of-00009.safetensors",
375
+ "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00009.safetensors",
376
+ "model.layers.7.self_attn.o_proj.bias": "model-00003-of-00009.safetensors",
377
+ "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00009.safetensors",
378
+ "model.layers.7.self_attn.q_proj.bias": "model-00003-of-00009.safetensors",
379
+ "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00009.safetensors",
380
+ "model.layers.7.self_attn.sinks": "model-00003-of-00009.safetensors",
381
+ "model.layers.7.self_attn.v_proj.bias": "model-00003-of-00009.safetensors",
382
+ "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00009.safetensors",
383
+ "model.layers.8.input_layernorm.weight": "model-00004-of-00009.safetensors",
384
+ "model.layers.8.mlp.experts.down_proj": "model-00004-of-00009.safetensors",
385
+ "model.layers.8.mlp.experts.down_proj_bias": "model-00004-of-00009.safetensors",
386
+ "model.layers.8.mlp.experts.gate_up_proj": "model-00004-of-00009.safetensors",
387
+ "model.layers.8.mlp.experts.gate_up_proj_bias": "model-00004-of-00009.safetensors",
388
+ "model.layers.8.mlp.router.bias": "model-00003-of-00009.safetensors",
389
+ "model.layers.8.mlp.router.weight": "model-00003-of-00009.safetensors",
390
+ "model.layers.8.post_attention_layernorm.weight": "model-00004-of-00009.safetensors",
391
+ "model.layers.8.self_attn.k_proj.bias": "model-00003-of-00009.safetensors",
392
+ "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00009.safetensors",
393
+ "model.layers.8.self_attn.o_proj.bias": "model-00003-of-00009.safetensors",
394
+ "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00009.safetensors",
395
+ "model.layers.8.self_attn.q_proj.bias": "model-00003-of-00009.safetensors",
396
+ "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00009.safetensors",
397
+ "model.layers.8.self_attn.sinks": "model-00003-of-00009.safetensors",
398
+ "model.layers.8.self_attn.v_proj.bias": "model-00003-of-00009.safetensors",
399
+ "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00009.safetensors",
400
+ "model.layers.9.input_layernorm.weight": "model-00004-of-00009.safetensors",
401
+ "model.layers.9.mlp.experts.down_proj": "model-00004-of-00009.safetensors",
402
+ "model.layers.9.mlp.experts.down_proj_bias": "model-00004-of-00009.safetensors",
403
+ "model.layers.9.mlp.experts.gate_up_proj": "model-00004-of-00009.safetensors",
404
+ "model.layers.9.mlp.experts.gate_up_proj_bias": "model-00004-of-00009.safetensors",
405
+ "model.layers.9.mlp.router.bias": "model-00004-of-00009.safetensors",
406
+ "model.layers.9.mlp.router.weight": "model-00004-of-00009.safetensors",
407
+ "model.layers.9.post_attention_layernorm.weight": "model-00004-of-00009.safetensors",
408
+ "model.layers.9.self_attn.k_proj.bias": "model-00004-of-00009.safetensors",
409
+ "model.layers.9.self_attn.k_proj.weight": "model-00004-of-00009.safetensors",
410
+ "model.layers.9.self_attn.o_proj.bias": "model-00004-of-00009.safetensors",
411
+ "model.layers.9.self_attn.o_proj.weight": "model-00004-of-00009.safetensors",
412
+ "model.layers.9.self_attn.q_proj.bias": "model-00004-of-00009.safetensors",
413
+ "model.layers.9.self_attn.q_proj.weight": "model-00004-of-00009.safetensors",
414
+ "model.layers.9.self_attn.sinks": "model-00004-of-00009.safetensors",
415
+ "model.layers.9.self_attn.v_proj.bias": "model-00004-of-00009.safetensors",
416
+ "model.layers.9.self_attn.v_proj.weight": "model-00004-of-00009.safetensors",
417
+ "model.norm.weight": "model-00009-of-00009.safetensors"
418
+ }
419
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|return|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|endoftext|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0614fe83cadab421296e664e1f48f4261fa8fef6e03e63bb75c20f38e37d07d3
3
+ size 27868174
tokenizer_config.json ADDED
@@ -0,0 +1,183 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "199998": {
4
+ "content": "<|startoftext|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "199999": {
12
+ "content": "<|endoftext|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "200000": {
20
+ "content": "<|reserved_200000|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "200001": {
28
+ "content": "<|reserved_200001|>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "200002": {
36
+ "content": "<|return|>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "200003": {
44
+ "content": "<|constrain|>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "200004": {
52
+ "content": "<|reserved_200004|>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "200005": {
60
+ "content": "<|channel|>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "200006": {
68
+ "content": "<|start|>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "200007": {
76
+ "content": "<|end|>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "200008": {
84
+ "content": "<|message|>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "200009": {
92
+ "content": "<|reserved_200009|>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "200010": {
100
+ "content": "<|reserved_200010|>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "200011": {
108
+ "content": "<|reserved_200011|>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "200012": {
116
+ "content": "<|call|>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "200013": {
124
+ "content": "<|reserved_200013|>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "200014": {
132
+ "content": "<|reserved_200014|>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "200015": {
140
+ "content": "<|reserved_200015|>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "200016": {
148
+ "content": "<|reserved_200016|>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "200017": {
156
+ "content": "<|reserved_200017|>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "200018": {
164
+ "content": "<|endofprompt|>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ }
171
+ },
172
+ "bos_token": "<|startoftext|>",
173
+ "clean_up_tokenization_spaces": false,
174
+ "eos_token": "<|return|>",
175
+ "extra_special_tokens": {},
176
+ "model_input_names": [
177
+ "input_ids",
178
+ "attention_mask"
179
+ ],
180
+ "model_max_length": 1000000000000000019884624838656,
181
+ "pad_token": "<|endoftext|>",
182
+ "tokenizer_class": "PreTrainedTokenizerFast"
183
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7bfa58ce9748715a5711511426ed685dd546e89891028b84e29bedc4291ade9
3
+ size 5841