OpenCircuit commited on
Commit
2b7ac9c
·
verified ·
1 Parent(s): 58afedd

Upload BF-Router v0.5 QLoRA adapter (Qwen3-4B)

Browse files
Files changed (41) hide show
  1. .gitattributes +4 -0
  2. README.md +62 -0
  3. adapter_config.json +46 -0
  4. adapter_model.safetensors +3 -0
  5. chat_template.jinja +61 -0
  6. checkpoint-281/README.md +209 -0
  7. checkpoint-281/adapter_config.json +46 -0
  8. checkpoint-281/adapter_model.safetensors +3 -0
  9. checkpoint-281/chat_template.jinja +61 -0
  10. checkpoint-281/optimizer.pt +3 -0
  11. checkpoint-281/rng_state.pth +3 -0
  12. checkpoint-281/scheduler.pt +3 -0
  13. checkpoint-281/tokenizer.json +3 -0
  14. checkpoint-281/tokenizer_config.json +29 -0
  15. checkpoint-281/trainer_state.json +335 -0
  16. checkpoint-281/training_args.bin +3 -0
  17. checkpoint-562/README.md +209 -0
  18. checkpoint-562/adapter_config.json +46 -0
  19. checkpoint-562/adapter_model.safetensors +3 -0
  20. checkpoint-562/chat_template.jinja +61 -0
  21. checkpoint-562/optimizer.pt +3 -0
  22. checkpoint-562/rng_state.pth +3 -0
  23. checkpoint-562/scheduler.pt +3 -0
  24. checkpoint-562/tokenizer.json +3 -0
  25. checkpoint-562/tokenizer_config.json +29 -0
  26. checkpoint-562/trainer_state.json +626 -0
  27. checkpoint-562/training_args.bin +3 -0
  28. checkpoint-843/README.md +209 -0
  29. checkpoint-843/adapter_config.json +46 -0
  30. checkpoint-843/adapter_model.safetensors +3 -0
  31. checkpoint-843/chat_template.jinja +61 -0
  32. checkpoint-843/optimizer.pt +3 -0
  33. checkpoint-843/rng_state.pth +3 -0
  34. checkpoint-843/scheduler.pt +3 -0
  35. checkpoint-843/tokenizer.json +3 -0
  36. checkpoint-843/tokenizer_config.json +29 -0
  37. checkpoint-843/trainer_state.json +917 -0
  38. checkpoint-843/training_args.bin +3 -0
  39. tokenizer.json +3 -0
  40. tokenizer_config.json +29 -0
  41. training_args.bin +3 -0
.gitattributes CHANGED
@@ -33,3 +33,7 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ checkpoint-281/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ checkpoint-562/tokenizer.json filter=lfs diff=lfs merge=lfs -text
38
+ checkpoint-843/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B-Instruct-2507
3
+ library_name: peft
4
+ model_name: bf-router-v0.5
5
+ tags:
6
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ licence: license
12
+ pipeline_tag: text-generation
13
+ ---
14
+
15
+ # Model Card for bf-router-v0.5
16
+
17
+ This model is a fine-tuned version of [Qwen/Qwen3-4B-Instruct-2507](https://huggingface.co/Qwen/Qwen3-4B-Instruct-2507).
18
+ It has been trained using [TRL](https://github.com/huggingface/trl).
19
+
20
+ ## Quick start
21
+
22
+ ```python
23
+ from transformers import pipeline
24
+
25
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
26
+ generator = pipeline("text-generation", model="None", device="cuda")
27
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
28
+ print(output["generated_text"])
29
+ ```
30
+
31
+ ## Training procedure
32
+
33
+
34
+
35
+
36
+
37
+ This model was trained with SFT.
38
+
39
+ ### Framework versions
40
+
41
+ - PEFT 0.18.1
42
+ - TRL: 0.29.0
43
+ - Transformers: 5.2.0
44
+ - Pytorch: 2.10.0
45
+ - Datasets: 4.6.1
46
+ - Tokenizers: 0.22.2
47
+
48
+ ## Citations
49
+
50
+
51
+
52
+ Cite TRL as:
53
+
54
+ ```bibtex
55
+ @software{vonwerra2020trl,
56
+ title = {{TRL: Transformers Reinforcement Learning}},
57
+ author = {von Werra, Leandro and Belkada, Younes and Tunstall, Lewis and Beeching, Edward and Thrush, Tristan and Lambert, Nathan and Huang, Shengyi and Rasul, Kashif and Gallouédec, Quentin},
58
+ license = {Apache-2.0},
59
+ url = {https://github.com/huggingface/trl},
60
+ year = {2020}
61
+ }
62
+ ```
adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": null,
25
+ "peft_type": "LORA",
26
+ "peft_version": "0.18.1",
27
+ "qalora_group_size": 16,
28
+ "r": 16,
29
+ "rank_pattern": {},
30
+ "revision": null,
31
+ "target_modules": [
32
+ "v_proj",
33
+ "down_proj",
34
+ "up_proj",
35
+ "q_proj",
36
+ "gate_proj",
37
+ "k_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bc1f22f429ab6ab7527f80ca2f343b90e4ea80c39dda2ec83bada6e7924d676
3
+ size 132187888
chat_template.jinja ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- for message in messages %}
18
+ {%- if message.content is string %}
19
+ {%- set content = message.content %}
20
+ {%- else %}
21
+ {%- set content = '' %}
22
+ {%- endif %}
23
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
24
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
25
+ {%- elif message.role == "assistant" %}
26
+ {{- '<|im_start|>' + message.role + '\n' + content }}
27
+ {%- if message.tool_calls %}
28
+ {%- for tool_call in message.tool_calls %}
29
+ {%- if (loop.first and content) or (not loop.first) %}
30
+ {{- '\n' }}
31
+ {%- endif %}
32
+ {%- if tool_call.function %}
33
+ {%- set tool_call = tool_call.function %}
34
+ {%- endif %}
35
+ {{- '<tool_call>\n{"name": "' }}
36
+ {{- tool_call.name }}
37
+ {{- '", "arguments": ' }}
38
+ {%- if tool_call.arguments is string %}
39
+ {{- tool_call.arguments }}
40
+ {%- else %}
41
+ {{- tool_call.arguments | tojson }}
42
+ {%- endif %}
43
+ {{- '}\n</tool_call>' }}
44
+ {%- endfor %}
45
+ {%- endif %}
46
+ {{- '<|im_end|>\n' }}
47
+ {%- elif message.role == "tool" %}
48
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
49
+ {{- '<|im_start|>user' }}
50
+ {%- endif %}
51
+ {{- '\n<tool_response>\n' }}
52
+ {{- content }}
53
+ {{- '\n</tool_response>' }}
54
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
55
+ {{- '<|im_end|>\n' }}
56
+ {%- endif %}
57
+ {%- endif %}
58
+ {%- endfor %}
59
+ {%- if add_generation_prompt %}
60
+ {{- '<|im_start|>assistant\n' }}
61
+ {%- endif %}
checkpoint-281/README.md ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B-Instruct-2507
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ ---
12
+
13
+ # Model Card for Model ID
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ <!-- Provide a longer summary of what this model is. -->
24
+
25
+
26
+
27
+ - **Developed by:** [More Information Needed]
28
+ - **Funded by [optional]:** [More Information Needed]
29
+ - **Shared by [optional]:** [More Information Needed]
30
+ - **Model type:** [More Information Needed]
31
+ - **Language(s) (NLP):** [More Information Needed]
32
+ - **License:** [More Information Needed]
33
+ - **Finetuned from model [optional]:** [More Information Needed]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ <!-- Provide the basic links for the model. -->
38
+
39
+ - **Repository:** [More Information Needed]
40
+ - **Paper [optional]:** [More Information Needed]
41
+ - **Demo [optional]:** [More Information Needed]
42
+
43
+ ## Uses
44
+
45
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
+
47
+ ### Direct Use
48
+
49
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Downstream Use [optional]
54
+
55
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+
57
+ [More Information Needed]
58
+
59
+ ### Out-of-Scope Use
60
+
61
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
74
+
75
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
76
+
77
+ ## How to Get Started with the Model
78
+
79
+ Use the code below to get started with the model.
80
+
81
+ [More Information Needed]
82
+
83
+ ## Training Details
84
+
85
+ ### Training Data
86
+
87
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
+
89
+ [More Information Needed]
90
+
91
+ ### Training Procedure
92
+
93
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
+
95
+ #### Preprocessing [optional]
96
+
97
+ [More Information Needed]
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
103
+
104
+ #### Speeds, Sizes, Times [optional]
105
+
106
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ## Evaluation
111
+
112
+ <!-- This section describes the evaluation protocols and provides the results. -->
113
+
114
+ ### Testing Data, Factors & Metrics
115
+
116
+ #### Testing Data
117
+
118
+ <!-- This should link to a Dataset Card if possible. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Factors
123
+
124
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
125
+
126
+ [More Information Needed]
127
+
128
+ #### Metrics
129
+
130
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
131
+
132
+ [More Information Needed]
133
+
134
+ ### Results
135
+
136
+ [More Information Needed]
137
+
138
+ #### Summary
139
+
140
+
141
+
142
+ ## Model Examination [optional]
143
+
144
+ <!-- Relevant interpretability work for the model goes here -->
145
+
146
+ [More Information Needed]
147
+
148
+ ## Environmental Impact
149
+
150
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
151
+
152
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
153
+
154
+ - **Hardware Type:** [More Information Needed]
155
+ - **Hours used:** [More Information Needed]
156
+ - **Cloud Provider:** [More Information Needed]
157
+ - **Compute Region:** [More Information Needed]
158
+ - **Carbon Emitted:** [More Information Needed]
159
+
160
+ ## Technical Specifications [optional]
161
+
162
+ ### Model Architecture and Objective
163
+
164
+ [More Information Needed]
165
+
166
+ ### Compute Infrastructure
167
+
168
+ [More Information Needed]
169
+
170
+ #### Hardware
171
+
172
+ [More Information Needed]
173
+
174
+ #### Software
175
+
176
+ [More Information Needed]
177
+
178
+ ## Citation [optional]
179
+
180
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
181
+
182
+ **BibTeX:**
183
+
184
+ [More Information Needed]
185
+
186
+ **APA:**
187
+
188
+ [More Information Needed]
189
+
190
+ ## Glossary [optional]
191
+
192
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
193
+
194
+ [More Information Needed]
195
+
196
+ ## More Information [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Authors [optional]
201
+
202
+ [More Information Needed]
203
+
204
+ ## Model Card Contact
205
+
206
+ [More Information Needed]
207
+ ### Framework versions
208
+
209
+ - PEFT 0.18.1
checkpoint-281/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": null,
25
+ "peft_type": "LORA",
26
+ "peft_version": "0.18.1",
27
+ "qalora_group_size": 16,
28
+ "r": 16,
29
+ "rank_pattern": {},
30
+ "revision": null,
31
+ "target_modules": [
32
+ "v_proj",
33
+ "down_proj",
34
+ "up_proj",
35
+ "q_proj",
36
+ "gate_proj",
37
+ "k_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
checkpoint-281/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:581d2f686c971b84c9ab741950e24777fcd2dc0452a11ae7df7b862883fc2804
3
+ size 66127776
checkpoint-281/chat_template.jinja ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- for message in messages %}
18
+ {%- if message.content is string %}
19
+ {%- set content = message.content %}
20
+ {%- else %}
21
+ {%- set content = '' %}
22
+ {%- endif %}
23
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
24
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
25
+ {%- elif message.role == "assistant" %}
26
+ {{- '<|im_start|>' + message.role + '\n' + content }}
27
+ {%- if message.tool_calls %}
28
+ {%- for tool_call in message.tool_calls %}
29
+ {%- if (loop.first and content) or (not loop.first) %}
30
+ {{- '\n' }}
31
+ {%- endif %}
32
+ {%- if tool_call.function %}
33
+ {%- set tool_call = tool_call.function %}
34
+ {%- endif %}
35
+ {{- '<tool_call>\n{"name": "' }}
36
+ {{- tool_call.name }}
37
+ {{- '", "arguments": ' }}
38
+ {%- if tool_call.arguments is string %}
39
+ {{- tool_call.arguments }}
40
+ {%- else %}
41
+ {{- tool_call.arguments | tojson }}
42
+ {%- endif %}
43
+ {{- '}\n</tool_call>' }}
44
+ {%- endfor %}
45
+ {%- endif %}
46
+ {{- '<|im_end|>\n' }}
47
+ {%- elif message.role == "tool" %}
48
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
49
+ {{- '<|im_start|>user' }}
50
+ {%- endif %}
51
+ {{- '\n<tool_response>\n' }}
52
+ {{- content }}
53
+ {{- '\n</tool_response>' }}
54
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
55
+ {{- '<|im_end|>\n' }}
56
+ {%- endif %}
57
+ {%- endif %}
58
+ {%- endfor %}
59
+ {%- if add_generation_prompt %}
60
+ {{- '<|im_start|>assistant\n' }}
61
+ {%- endif %}
checkpoint-281/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:875ffb16cc18b423e5ef0a71bd54a2a49974c68c9ee717b877987df1945c5e51
3
+ size 132552715
checkpoint-281/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95e70b1e3fd60fe41298a7546c09ee7f379684cfb911eba36e8ca1865e2f5d46
3
+ size 14645
checkpoint-281/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aeccf90c5a1ed591a7458cc9cf95f5fb44b5b6e02069bc531935e2d480503332
3
+ size 1465
checkpoint-281/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be75606093db2094d7cd20f3c2f385c212750648bd6ea4fb2bf507a6a4c55506
3
+ size 11422650
checkpoint-281/tokenizer_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": null,
5
+ "clean_up_tokenization_spaces": false,
6
+ "eos_token": "<|im_end|>",
7
+ "errors": "replace",
8
+ "extra_special_tokens": [
9
+ "<|im_start|>",
10
+ "<|im_end|>",
11
+ "<|object_ref_start|>",
12
+ "<|object_ref_end|>",
13
+ "<|box_start|>",
14
+ "<|box_end|>",
15
+ "<|quad_start|>",
16
+ "<|quad_end|>",
17
+ "<|vision_start|>",
18
+ "<|vision_end|>",
19
+ "<|vision_pad|>",
20
+ "<|image_pad|>",
21
+ "<|video_pad|>"
22
+ ],
23
+ "is_local": false,
24
+ "model_max_length": 1010000,
25
+ "pad_token": "<|im_end|>",
26
+ "split_special_tokens": false,
27
+ "tokenizer_class": "Qwen2Tokenizer",
28
+ "unk_token": null
29
+ }
checkpoint-281/trainer_state.json ADDED
@@ -0,0 +1,335 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 281,
3
+ "best_metric": 0.18149949610233307,
4
+ "best_model_checkpoint": "/app/output/bf-router-v0.5/checkpoint-281",
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 281,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "entropy": 1.4157426357269287,
14
+ "epoch": 0.003561887800534283,
15
+ "grad_norm": 1.203125,
16
+ "learning_rate": 0.0,
17
+ "loss": 4.336674213409424,
18
+ "mean_token_accuracy": 0.4680640399456024,
19
+ "num_tokens": 3092.0,
20
+ "step": 1
21
+ },
22
+ {
23
+ "entropy": 1.5884604156017303,
24
+ "epoch": 0.03561887800534283,
25
+ "grad_norm": 0.59375,
26
+ "learning_rate": 6.923076923076924e-05,
27
+ "loss": 3.986017862955729,
28
+ "mean_token_accuracy": 0.4735516524977154,
29
+ "num_tokens": 30650.0,
30
+ "step": 10
31
+ },
32
+ {
33
+ "entropy": 2.2315500736236573,
34
+ "epoch": 0.07123775601068566,
35
+ "grad_norm": 0.26953125,
36
+ "learning_rate": 0.00014615384615384615,
37
+ "loss": 2.16544189453125,
38
+ "mean_token_accuracy": 0.6272616922855377,
39
+ "num_tokens": 61735.0,
40
+ "step": 20
41
+ },
42
+ {
43
+ "entropy": 0.8066075079143047,
44
+ "epoch": 0.10685663401602849,
45
+ "grad_norm": 0.298828125,
46
+ "learning_rate": 0.00019999334629824895,
47
+ "loss": 0.7935717582702637,
48
+ "mean_token_accuracy": 0.857068446278572,
49
+ "num_tokens": 92484.0,
50
+ "step": 30
51
+ },
52
+ {
53
+ "entropy": 0.5340980306267739,
54
+ "epoch": 0.14247551202137132,
55
+ "grad_norm": 0.10693359375,
56
+ "learning_rate": 0.00019987508289735716,
57
+ "loss": 0.5261529922485352,
58
+ "mean_token_accuracy": 0.9004581302404404,
59
+ "num_tokens": 123010.0,
60
+ "step": 40
61
+ },
62
+ {
63
+ "entropy": 0.4001643668860197,
64
+ "epoch": 0.17809439002671415,
65
+ "grad_norm": 0.08251953125,
66
+ "learning_rate": 0.0001996091607179287,
67
+ "loss": 0.3956298351287842,
68
+ "mean_token_accuracy": 0.9209299519658088,
69
+ "num_tokens": 153710.0,
70
+ "step": 50
71
+ },
72
+ {
73
+ "entropy": 0.34828933551907537,
74
+ "epoch": 0.21371326803205698,
75
+ "grad_norm": 0.10009765625,
76
+ "learning_rate": 0.00019919597290851538,
77
+ "loss": 0.3648662090301514,
78
+ "mean_token_accuracy": 0.9263768374919892,
79
+ "num_tokens": 184577.0,
80
+ "step": 60
81
+ },
82
+ {
83
+ "entropy": 0.2983129292726517,
84
+ "epoch": 0.2493321460373998,
85
+ "grad_norm": 0.08447265625,
86
+ "learning_rate": 0.00019863613034027224,
87
+ "loss": 0.29642860889434813,
88
+ "mean_token_accuracy": 0.9371836110949516,
89
+ "num_tokens": 215040.0,
90
+ "step": 70
91
+ },
92
+ {
93
+ "entropy": 0.280751570686698,
94
+ "epoch": 0.28495102404274264,
95
+ "grad_norm": 0.099609375,
96
+ "learning_rate": 0.00019793046070382437,
97
+ "loss": 0.2813178300857544,
98
+ "mean_token_accuracy": 0.9388988897204399,
99
+ "num_tokens": 245683.0,
100
+ "step": 80
101
+ },
102
+ {
103
+ "entropy": 0.2960826952010393,
104
+ "epoch": 0.3205699020480855,
105
+ "grad_norm": 0.11962890625,
106
+ "learning_rate": 0.00019708000728558064,
107
+ "loss": 0.31115410327911375,
108
+ "mean_token_accuracy": 0.9335825845599175,
109
+ "num_tokens": 276430.0,
110
+ "step": 90
111
+ },
112
+ {
113
+ "entropy": 0.28416929207742214,
114
+ "epoch": 0.3561887800534283,
115
+ "grad_norm": 0.103515625,
116
+ "learning_rate": 0.00019608602742530283,
117
+ "loss": 0.27938365936279297,
118
+ "mean_token_accuracy": 0.9393619552254677,
119
+ "num_tokens": 307296.0,
120
+ "step": 100
121
+ },
122
+ {
123
+ "entropy": 0.23417662251740695,
124
+ "epoch": 0.39180765805877116,
125
+ "grad_norm": 0.06298828125,
126
+ "learning_rate": 0.00019494999065721108,
127
+ "loss": 0.23734774589538574,
128
+ "mean_token_accuracy": 0.9487677246332169,
129
+ "num_tokens": 337962.0,
130
+ "step": 110
131
+ },
132
+ {
133
+ "entropy": 0.24019648060202597,
134
+ "epoch": 0.42742653606411396,
135
+ "grad_norm": 0.06640625,
136
+ "learning_rate": 0.0001936735765373737,
137
+ "loss": 0.245324969291687,
138
+ "mean_token_accuracy": 0.9457737937569618,
139
+ "num_tokens": 368782.0,
140
+ "step": 120
141
+ },
142
+ {
143
+ "entropy": 0.269666058011353,
144
+ "epoch": 0.4630454140694568,
145
+ "grad_norm": 0.09765625,
146
+ "learning_rate": 0.00019225867216059325,
147
+ "loss": 0.2686375617980957,
148
+ "mean_token_accuracy": 0.9398508563637733,
149
+ "num_tokens": 399329.0,
150
+ "step": 130
151
+ },
152
+ {
153
+ "entropy": 0.18328354582190515,
154
+ "epoch": 0.4986642920747996,
155
+ "grad_norm": 0.06298828125,
156
+ "learning_rate": 0.00019070736937046035,
157
+ "loss": 0.18070420026779174,
158
+ "mean_token_accuracy": 0.9568557634949684,
159
+ "num_tokens": 429759.0,
160
+ "step": 140
161
+ },
162
+ {
163
+ "entropy": 0.21165105439722537,
164
+ "epoch": 0.5342831700801425,
165
+ "grad_norm": 0.0810546875,
166
+ "learning_rate": 0.0001890219616666997,
167
+ "loss": 0.22004930973052977,
168
+ "mean_token_accuracy": 0.9494227185845375,
169
+ "num_tokens": 460409.0,
170
+ "step": 150
171
+ },
172
+ {
173
+ "entropy": 0.23916025627404452,
174
+ "epoch": 0.5699020480854853,
175
+ "grad_norm": 0.07568359375,
176
+ "learning_rate": 0.00018720494081438078,
177
+ "loss": 0.2379377841949463,
178
+ "mean_token_accuracy": 0.9455192387104034,
179
+ "num_tokens": 491344.0,
180
+ "step": 160
181
+ },
182
+ {
183
+ "entropy": 0.2238690486177802,
184
+ "epoch": 0.6055209260908282,
185
+ "grad_norm": 0.08984375,
186
+ "learning_rate": 0.00018525899316000608,
187
+ "loss": 0.2245168685913086,
188
+ "mean_token_accuracy": 0.9485938593745231,
189
+ "num_tokens": 522029.0,
190
+ "step": 170
191
+ },
192
+ {
193
+ "entropy": 0.17032606173306702,
194
+ "epoch": 0.641139804096171,
195
+ "grad_norm": 0.080078125,
196
+ "learning_rate": 0.00018318699565992357,
197
+ "loss": 0.17549347877502441,
198
+ "mean_token_accuracy": 0.9574712902307511,
199
+ "num_tokens": 552306.0,
200
+ "step": 180
201
+ },
202
+ {
203
+ "entropy": 0.21890966054052113,
204
+ "epoch": 0.6767586821015138,
205
+ "grad_norm": 0.0673828125,
206
+ "learning_rate": 0.00018099201162693476,
207
+ "loss": 0.21968293190002441,
208
+ "mean_token_accuracy": 0.9482985377311707,
209
+ "num_tokens": 583181.0,
210
+ "step": 190
211
+ },
212
+ {
213
+ "entropy": 0.20646683853119613,
214
+ "epoch": 0.7123775601068566,
215
+ "grad_norm": 0.09765625,
216
+ "learning_rate": 0.00017867728620138708,
217
+ "loss": 0.20968315601348878,
218
+ "mean_token_accuracy": 0.9519409075379371,
219
+ "num_tokens": 613818.0,
220
+ "step": 200
221
+ },
222
+ {
223
+ "entropy": 0.23000301327556372,
224
+ "epoch": 0.7479964381121995,
225
+ "grad_norm": 0.08935546875,
226
+ "learning_rate": 0.00017624624155344626,
227
+ "loss": 0.2383474349975586,
228
+ "mean_token_accuracy": 0.9458372846245766,
229
+ "num_tokens": 644576.0,
230
+ "step": 210
231
+ },
232
+ {
233
+ "entropy": 0.21420854832977057,
234
+ "epoch": 0.7836153161175423,
235
+ "grad_norm": 0.0966796875,
236
+ "learning_rate": 0.0001737024718236413,
237
+ "loss": 0.21017465591430665,
238
+ "mean_token_accuracy": 0.9511569887399673,
239
+ "num_tokens": 675564.0,
240
+ "step": 220
241
+ },
242
+ {
243
+ "entropy": 0.19280126914381981,
244
+ "epoch": 0.8192341941228851,
245
+ "grad_norm": 0.08154296875,
246
+ "learning_rate": 0.00017104973780916294,
247
+ "loss": 0.1936139941215515,
248
+ "mean_token_accuracy": 0.954449312388897,
249
+ "num_tokens": 706309.0,
250
+ "step": 230
251
+ },
252
+ {
253
+ "entropy": 0.22524890769273043,
254
+ "epoch": 0.8548530721282279,
255
+ "grad_norm": 0.10546875,
256
+ "learning_rate": 0.00016829196140377085,
257
+ "loss": 0.22950620651245118,
258
+ "mean_token_accuracy": 0.9476036429405212,
259
+ "num_tokens": 737135.0,
260
+ "step": 240
261
+ },
262
+ {
263
+ "entropy": 0.18842433094978334,
264
+ "epoch": 0.8904719501335708,
265
+ "grad_norm": 0.087890625,
266
+ "learning_rate": 0.00016543321979953007,
267
+ "loss": 0.19049547910690307,
268
+ "mean_token_accuracy": 0.9543813273310662,
269
+ "num_tokens": 767853.0,
270
+ "step": 250
271
+ },
272
+ {
273
+ "entropy": 0.1991994746029377,
274
+ "epoch": 0.9260908281389136,
275
+ "grad_norm": 0.07763671875,
276
+ "learning_rate": 0.00016247773945894962,
277
+ "loss": 0.19527161121368408,
278
+ "mean_token_accuracy": 0.9535173490643502,
279
+ "num_tokens": 798511.0,
280
+ "step": 260
281
+ },
282
+ {
283
+ "entropy": 0.18196075949817897,
284
+ "epoch": 0.9617097061442564,
285
+ "grad_norm": 0.0771484375,
286
+ "learning_rate": 0.00015942988986643352,
287
+ "loss": 0.19091761112213135,
288
+ "mean_token_accuracy": 0.9535432115197182,
289
+ "num_tokens": 829447.0,
290
+ "step": 270
291
+ },
292
+ {
293
+ "entropy": 0.18307169582694768,
294
+ "epoch": 0.9973285841495992,
295
+ "grad_norm": 0.08837890625,
296
+ "learning_rate": 0.00015629417706828423,
297
+ "loss": 0.1818631410598755,
298
+ "mean_token_accuracy": 0.955574706196785,
299
+ "num_tokens": 860142.0,
300
+ "step": 280
301
+ },
302
+ {
303
+ "epoch": 1.0,
304
+ "eval_entropy": 0.18207345770817276,
305
+ "eval_loss": 0.18149949610233307,
306
+ "eval_mean_token_accuracy": 0.9555717660180221,
307
+ "eval_num_tokens": 862512.0,
308
+ "eval_runtime": 35.3256,
309
+ "eval_samples_per_second": 15.881,
310
+ "eval_steps_per_second": 3.991,
311
+ "step": 281
312
+ }
313
+ ],
314
+ "logging_steps": 10,
315
+ "max_steps": 843,
316
+ "num_input_tokens_seen": 0,
317
+ "num_train_epochs": 3,
318
+ "save_steps": 500,
319
+ "stateful_callbacks": {
320
+ "TrainerControl": {
321
+ "args": {
322
+ "should_epoch_stop": false,
323
+ "should_evaluate": false,
324
+ "should_log": false,
325
+ "should_save": true,
326
+ "should_training_stop": false
327
+ },
328
+ "attributes": {}
329
+ }
330
+ },
331
+ "total_flos": 2.028075887327232e+16,
332
+ "train_batch_size": 4,
333
+ "trial_name": null,
334
+ "trial_params": null
335
+ }
checkpoint-281/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57851b427cb3a4649963b62f1b5d8860aa50ba7ac4abbbb2ebc56acbf65bd694
3
+ size 5649
checkpoint-562/README.md ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B-Instruct-2507
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ ---
12
+
13
+ # Model Card for Model ID
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ <!-- Provide a longer summary of what this model is. -->
24
+
25
+
26
+
27
+ - **Developed by:** [More Information Needed]
28
+ - **Funded by [optional]:** [More Information Needed]
29
+ - **Shared by [optional]:** [More Information Needed]
30
+ - **Model type:** [More Information Needed]
31
+ - **Language(s) (NLP):** [More Information Needed]
32
+ - **License:** [More Information Needed]
33
+ - **Finetuned from model [optional]:** [More Information Needed]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ <!-- Provide the basic links for the model. -->
38
+
39
+ - **Repository:** [More Information Needed]
40
+ - **Paper [optional]:** [More Information Needed]
41
+ - **Demo [optional]:** [More Information Needed]
42
+
43
+ ## Uses
44
+
45
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
+
47
+ ### Direct Use
48
+
49
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Downstream Use [optional]
54
+
55
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+
57
+ [More Information Needed]
58
+
59
+ ### Out-of-Scope Use
60
+
61
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
74
+
75
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
76
+
77
+ ## How to Get Started with the Model
78
+
79
+ Use the code below to get started with the model.
80
+
81
+ [More Information Needed]
82
+
83
+ ## Training Details
84
+
85
+ ### Training Data
86
+
87
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
+
89
+ [More Information Needed]
90
+
91
+ ### Training Procedure
92
+
93
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
+
95
+ #### Preprocessing [optional]
96
+
97
+ [More Information Needed]
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
103
+
104
+ #### Speeds, Sizes, Times [optional]
105
+
106
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ## Evaluation
111
+
112
+ <!-- This section describes the evaluation protocols and provides the results. -->
113
+
114
+ ### Testing Data, Factors & Metrics
115
+
116
+ #### Testing Data
117
+
118
+ <!-- This should link to a Dataset Card if possible. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Factors
123
+
124
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
125
+
126
+ [More Information Needed]
127
+
128
+ #### Metrics
129
+
130
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
131
+
132
+ [More Information Needed]
133
+
134
+ ### Results
135
+
136
+ [More Information Needed]
137
+
138
+ #### Summary
139
+
140
+
141
+
142
+ ## Model Examination [optional]
143
+
144
+ <!-- Relevant interpretability work for the model goes here -->
145
+
146
+ [More Information Needed]
147
+
148
+ ## Environmental Impact
149
+
150
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
151
+
152
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
153
+
154
+ - **Hardware Type:** [More Information Needed]
155
+ - **Hours used:** [More Information Needed]
156
+ - **Cloud Provider:** [More Information Needed]
157
+ - **Compute Region:** [More Information Needed]
158
+ - **Carbon Emitted:** [More Information Needed]
159
+
160
+ ## Technical Specifications [optional]
161
+
162
+ ### Model Architecture and Objective
163
+
164
+ [More Information Needed]
165
+
166
+ ### Compute Infrastructure
167
+
168
+ [More Information Needed]
169
+
170
+ #### Hardware
171
+
172
+ [More Information Needed]
173
+
174
+ #### Software
175
+
176
+ [More Information Needed]
177
+
178
+ ## Citation [optional]
179
+
180
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
181
+
182
+ **BibTeX:**
183
+
184
+ [More Information Needed]
185
+
186
+ **APA:**
187
+
188
+ [More Information Needed]
189
+
190
+ ## Glossary [optional]
191
+
192
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
193
+
194
+ [More Information Needed]
195
+
196
+ ## More Information [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Authors [optional]
201
+
202
+ [More Information Needed]
203
+
204
+ ## Model Card Contact
205
+
206
+ [More Information Needed]
207
+ ### Framework versions
208
+
209
+ - PEFT 0.18.1
checkpoint-562/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": null,
25
+ "peft_type": "LORA",
26
+ "peft_version": "0.18.1",
27
+ "qalora_group_size": 16,
28
+ "r": 16,
29
+ "rank_pattern": {},
30
+ "revision": null,
31
+ "target_modules": [
32
+ "v_proj",
33
+ "down_proj",
34
+ "up_proj",
35
+ "q_proj",
36
+ "gate_proj",
37
+ "k_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
checkpoint-562/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae3fd5d71a6d75e54a11294dc7fb4c6024c2786654ab088f72658e12516606c7
3
+ size 66127776
checkpoint-562/chat_template.jinja ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- for message in messages %}
18
+ {%- if message.content is string %}
19
+ {%- set content = message.content %}
20
+ {%- else %}
21
+ {%- set content = '' %}
22
+ {%- endif %}
23
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
24
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
25
+ {%- elif message.role == "assistant" %}
26
+ {{- '<|im_start|>' + message.role + '\n' + content }}
27
+ {%- if message.tool_calls %}
28
+ {%- for tool_call in message.tool_calls %}
29
+ {%- if (loop.first and content) or (not loop.first) %}
30
+ {{- '\n' }}
31
+ {%- endif %}
32
+ {%- if tool_call.function %}
33
+ {%- set tool_call = tool_call.function %}
34
+ {%- endif %}
35
+ {{- '<tool_call>\n{"name": "' }}
36
+ {{- tool_call.name }}
37
+ {{- '", "arguments": ' }}
38
+ {%- if tool_call.arguments is string %}
39
+ {{- tool_call.arguments }}
40
+ {%- else %}
41
+ {{- tool_call.arguments | tojson }}
42
+ {%- endif %}
43
+ {{- '}\n</tool_call>' }}
44
+ {%- endfor %}
45
+ {%- endif %}
46
+ {{- '<|im_end|>\n' }}
47
+ {%- elif message.role == "tool" %}
48
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
49
+ {{- '<|im_start|>user' }}
50
+ {%- endif %}
51
+ {{- '\n<tool_response>\n' }}
52
+ {{- content }}
53
+ {{- '\n</tool_response>' }}
54
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
55
+ {{- '<|im_end|>\n' }}
56
+ {%- endif %}
57
+ {%- endif %}
58
+ {%- endfor %}
59
+ {%- if add_generation_prompt %}
60
+ {{- '<|im_start|>assistant\n' }}
61
+ {%- endif %}
checkpoint-562/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9367da0e9e0f21638be2669e6d4ae2225b3503820ad3583a3c9b0334bf37c92d
3
+ size 132552715
checkpoint-562/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4313f9d832b6c6d568b3ac0b53ef1b8e83d1072bbf993a8167c50a30e46375a7
3
+ size 14645
checkpoint-562/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d813cca777bf0ff28660547edc526875199f3774d19049acf9916e60c27f3d6
3
+ size 1465
checkpoint-562/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be75606093db2094d7cd20f3c2f385c212750648bd6ea4fb2bf507a6a4c55506
3
+ size 11422650
checkpoint-562/tokenizer_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": null,
5
+ "clean_up_tokenization_spaces": false,
6
+ "eos_token": "<|im_end|>",
7
+ "errors": "replace",
8
+ "extra_special_tokens": [
9
+ "<|im_start|>",
10
+ "<|im_end|>",
11
+ "<|object_ref_start|>",
12
+ "<|object_ref_end|>",
13
+ "<|box_start|>",
14
+ "<|box_end|>",
15
+ "<|quad_start|>",
16
+ "<|quad_end|>",
17
+ "<|vision_start|>",
18
+ "<|vision_end|>",
19
+ "<|vision_pad|>",
20
+ "<|image_pad|>",
21
+ "<|video_pad|>"
22
+ ],
23
+ "is_local": false,
24
+ "model_max_length": 1010000,
25
+ "pad_token": "<|im_end|>",
26
+ "split_special_tokens": false,
27
+ "tokenizer_class": "Qwen2Tokenizer",
28
+ "unk_token": null
29
+ }
checkpoint-562/trainer_state.json ADDED
@@ -0,0 +1,626 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 562,
3
+ "best_metric": 0.15477769076824188,
4
+ "best_model_checkpoint": "/app/output/bf-router-v0.5/checkpoint-562",
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 562,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "entropy": 1.4157426357269287,
14
+ "epoch": 0.003561887800534283,
15
+ "grad_norm": 1.203125,
16
+ "learning_rate": 0.0,
17
+ "loss": 4.336674213409424,
18
+ "mean_token_accuracy": 0.4680640399456024,
19
+ "num_tokens": 3092.0,
20
+ "step": 1
21
+ },
22
+ {
23
+ "entropy": 1.5884604156017303,
24
+ "epoch": 0.03561887800534283,
25
+ "grad_norm": 0.59375,
26
+ "learning_rate": 6.923076923076924e-05,
27
+ "loss": 3.986017862955729,
28
+ "mean_token_accuracy": 0.4735516524977154,
29
+ "num_tokens": 30650.0,
30
+ "step": 10
31
+ },
32
+ {
33
+ "entropy": 2.2315500736236573,
34
+ "epoch": 0.07123775601068566,
35
+ "grad_norm": 0.26953125,
36
+ "learning_rate": 0.00014615384615384615,
37
+ "loss": 2.16544189453125,
38
+ "mean_token_accuracy": 0.6272616922855377,
39
+ "num_tokens": 61735.0,
40
+ "step": 20
41
+ },
42
+ {
43
+ "entropy": 0.8066075079143047,
44
+ "epoch": 0.10685663401602849,
45
+ "grad_norm": 0.298828125,
46
+ "learning_rate": 0.00019999334629824895,
47
+ "loss": 0.7935717582702637,
48
+ "mean_token_accuracy": 0.857068446278572,
49
+ "num_tokens": 92484.0,
50
+ "step": 30
51
+ },
52
+ {
53
+ "entropy": 0.5340980306267739,
54
+ "epoch": 0.14247551202137132,
55
+ "grad_norm": 0.10693359375,
56
+ "learning_rate": 0.00019987508289735716,
57
+ "loss": 0.5261529922485352,
58
+ "mean_token_accuracy": 0.9004581302404404,
59
+ "num_tokens": 123010.0,
60
+ "step": 40
61
+ },
62
+ {
63
+ "entropy": 0.4001643668860197,
64
+ "epoch": 0.17809439002671415,
65
+ "grad_norm": 0.08251953125,
66
+ "learning_rate": 0.0001996091607179287,
67
+ "loss": 0.3956298351287842,
68
+ "mean_token_accuracy": 0.9209299519658088,
69
+ "num_tokens": 153710.0,
70
+ "step": 50
71
+ },
72
+ {
73
+ "entropy": 0.34828933551907537,
74
+ "epoch": 0.21371326803205698,
75
+ "grad_norm": 0.10009765625,
76
+ "learning_rate": 0.00019919597290851538,
77
+ "loss": 0.3648662090301514,
78
+ "mean_token_accuracy": 0.9263768374919892,
79
+ "num_tokens": 184577.0,
80
+ "step": 60
81
+ },
82
+ {
83
+ "entropy": 0.2983129292726517,
84
+ "epoch": 0.2493321460373998,
85
+ "grad_norm": 0.08447265625,
86
+ "learning_rate": 0.00019863613034027224,
87
+ "loss": 0.29642860889434813,
88
+ "mean_token_accuracy": 0.9371836110949516,
89
+ "num_tokens": 215040.0,
90
+ "step": 70
91
+ },
92
+ {
93
+ "entropy": 0.280751570686698,
94
+ "epoch": 0.28495102404274264,
95
+ "grad_norm": 0.099609375,
96
+ "learning_rate": 0.00019793046070382437,
97
+ "loss": 0.2813178300857544,
98
+ "mean_token_accuracy": 0.9388988897204399,
99
+ "num_tokens": 245683.0,
100
+ "step": 80
101
+ },
102
+ {
103
+ "entropy": 0.2960826952010393,
104
+ "epoch": 0.3205699020480855,
105
+ "grad_norm": 0.11962890625,
106
+ "learning_rate": 0.00019708000728558064,
107
+ "loss": 0.31115410327911375,
108
+ "mean_token_accuracy": 0.9335825845599175,
109
+ "num_tokens": 276430.0,
110
+ "step": 90
111
+ },
112
+ {
113
+ "entropy": 0.28416929207742214,
114
+ "epoch": 0.3561887800534283,
115
+ "grad_norm": 0.103515625,
116
+ "learning_rate": 0.00019608602742530283,
117
+ "loss": 0.27938365936279297,
118
+ "mean_token_accuracy": 0.9393619552254677,
119
+ "num_tokens": 307296.0,
120
+ "step": 100
121
+ },
122
+ {
123
+ "entropy": 0.23417662251740695,
124
+ "epoch": 0.39180765805877116,
125
+ "grad_norm": 0.06298828125,
126
+ "learning_rate": 0.00019494999065721108,
127
+ "loss": 0.23734774589538574,
128
+ "mean_token_accuracy": 0.9487677246332169,
129
+ "num_tokens": 337962.0,
130
+ "step": 110
131
+ },
132
+ {
133
+ "entropy": 0.24019648060202597,
134
+ "epoch": 0.42742653606411396,
135
+ "grad_norm": 0.06640625,
136
+ "learning_rate": 0.0001936735765373737,
137
+ "loss": 0.245324969291687,
138
+ "mean_token_accuracy": 0.9457737937569618,
139
+ "num_tokens": 368782.0,
140
+ "step": 120
141
+ },
142
+ {
143
+ "entropy": 0.269666058011353,
144
+ "epoch": 0.4630454140694568,
145
+ "grad_norm": 0.09765625,
146
+ "learning_rate": 0.00019225867216059325,
147
+ "loss": 0.2686375617980957,
148
+ "mean_token_accuracy": 0.9398508563637733,
149
+ "num_tokens": 399329.0,
150
+ "step": 130
151
+ },
152
+ {
153
+ "entropy": 0.18328354582190515,
154
+ "epoch": 0.4986642920747996,
155
+ "grad_norm": 0.06298828125,
156
+ "learning_rate": 0.00019070736937046035,
157
+ "loss": 0.18070420026779174,
158
+ "mean_token_accuracy": 0.9568557634949684,
159
+ "num_tokens": 429759.0,
160
+ "step": 140
161
+ },
162
+ {
163
+ "entropy": 0.21165105439722537,
164
+ "epoch": 0.5342831700801425,
165
+ "grad_norm": 0.0810546875,
166
+ "learning_rate": 0.0001890219616666997,
167
+ "loss": 0.22004930973052977,
168
+ "mean_token_accuracy": 0.9494227185845375,
169
+ "num_tokens": 460409.0,
170
+ "step": 150
171
+ },
172
+ {
173
+ "entropy": 0.23916025627404452,
174
+ "epoch": 0.5699020480854853,
175
+ "grad_norm": 0.07568359375,
176
+ "learning_rate": 0.00018720494081438078,
177
+ "loss": 0.2379377841949463,
178
+ "mean_token_accuracy": 0.9455192387104034,
179
+ "num_tokens": 491344.0,
180
+ "step": 160
181
+ },
182
+ {
183
+ "entropy": 0.2238690486177802,
184
+ "epoch": 0.6055209260908282,
185
+ "grad_norm": 0.08984375,
186
+ "learning_rate": 0.00018525899316000608,
187
+ "loss": 0.2245168685913086,
188
+ "mean_token_accuracy": 0.9485938593745231,
189
+ "num_tokens": 522029.0,
190
+ "step": 170
191
+ },
192
+ {
193
+ "entropy": 0.17032606173306702,
194
+ "epoch": 0.641139804096171,
195
+ "grad_norm": 0.080078125,
196
+ "learning_rate": 0.00018318699565992357,
197
+ "loss": 0.17549347877502441,
198
+ "mean_token_accuracy": 0.9574712902307511,
199
+ "num_tokens": 552306.0,
200
+ "step": 180
201
+ },
202
+ {
203
+ "entropy": 0.21890966054052113,
204
+ "epoch": 0.6767586821015138,
205
+ "grad_norm": 0.0673828125,
206
+ "learning_rate": 0.00018099201162693476,
207
+ "loss": 0.21968293190002441,
208
+ "mean_token_accuracy": 0.9482985377311707,
209
+ "num_tokens": 583181.0,
210
+ "step": 190
211
+ },
212
+ {
213
+ "entropy": 0.20646683853119613,
214
+ "epoch": 0.7123775601068566,
215
+ "grad_norm": 0.09765625,
216
+ "learning_rate": 0.00017867728620138708,
217
+ "loss": 0.20968315601348878,
218
+ "mean_token_accuracy": 0.9519409075379371,
219
+ "num_tokens": 613818.0,
220
+ "step": 200
221
+ },
222
+ {
223
+ "entropy": 0.23000301327556372,
224
+ "epoch": 0.7479964381121995,
225
+ "grad_norm": 0.08935546875,
226
+ "learning_rate": 0.00017624624155344626,
227
+ "loss": 0.2383474349975586,
228
+ "mean_token_accuracy": 0.9458372846245766,
229
+ "num_tokens": 644576.0,
230
+ "step": 210
231
+ },
232
+ {
233
+ "entropy": 0.21420854832977057,
234
+ "epoch": 0.7836153161175423,
235
+ "grad_norm": 0.0966796875,
236
+ "learning_rate": 0.0001737024718236413,
237
+ "loss": 0.21017465591430665,
238
+ "mean_token_accuracy": 0.9511569887399673,
239
+ "num_tokens": 675564.0,
240
+ "step": 220
241
+ },
242
+ {
243
+ "entropy": 0.19280126914381981,
244
+ "epoch": 0.8192341941228851,
245
+ "grad_norm": 0.08154296875,
246
+ "learning_rate": 0.00017104973780916294,
247
+ "loss": 0.1936139941215515,
248
+ "mean_token_accuracy": 0.954449312388897,
249
+ "num_tokens": 706309.0,
250
+ "step": 230
251
+ },
252
+ {
253
+ "entropy": 0.22524890769273043,
254
+ "epoch": 0.8548530721282279,
255
+ "grad_norm": 0.10546875,
256
+ "learning_rate": 0.00016829196140377085,
257
+ "loss": 0.22950620651245118,
258
+ "mean_token_accuracy": 0.9476036429405212,
259
+ "num_tokens": 737135.0,
260
+ "step": 240
261
+ },
262
+ {
263
+ "entropy": 0.18842433094978334,
264
+ "epoch": 0.8904719501335708,
265
+ "grad_norm": 0.087890625,
266
+ "learning_rate": 0.00016543321979953007,
267
+ "loss": 0.19049547910690307,
268
+ "mean_token_accuracy": 0.9543813273310662,
269
+ "num_tokens": 767853.0,
270
+ "step": 250
271
+ },
272
+ {
273
+ "entropy": 0.1991994746029377,
274
+ "epoch": 0.9260908281389136,
275
+ "grad_norm": 0.07763671875,
276
+ "learning_rate": 0.00016247773945894962,
277
+ "loss": 0.19527161121368408,
278
+ "mean_token_accuracy": 0.9535173490643502,
279
+ "num_tokens": 798511.0,
280
+ "step": 260
281
+ },
282
+ {
283
+ "entropy": 0.18196075949817897,
284
+ "epoch": 0.9617097061442564,
285
+ "grad_norm": 0.0771484375,
286
+ "learning_rate": 0.00015942988986643352,
287
+ "loss": 0.19091761112213135,
288
+ "mean_token_accuracy": 0.9535432115197182,
289
+ "num_tokens": 829447.0,
290
+ "step": 270
291
+ },
292
+ {
293
+ "entropy": 0.18307169582694768,
294
+ "epoch": 0.9973285841495992,
295
+ "grad_norm": 0.08837890625,
296
+ "learning_rate": 0.00015629417706828423,
297
+ "loss": 0.1818631410598755,
298
+ "mean_token_accuracy": 0.955574706196785,
299
+ "num_tokens": 860142.0,
300
+ "step": 280
301
+ },
302
+ {
303
+ "epoch": 1.0,
304
+ "eval_entropy": 0.18207345770817276,
305
+ "eval_loss": 0.18149949610233307,
306
+ "eval_mean_token_accuracy": 0.9555717660180221,
307
+ "eval_num_tokens": 862512.0,
308
+ "eval_runtime": 35.3256,
309
+ "eval_samples_per_second": 15.881,
310
+ "eval_steps_per_second": 3.991,
311
+ "step": 281
312
+ },
313
+ {
314
+ "entropy": 0.1545313375118451,
315
+ "epoch": 1.0320569902048085,
316
+ "grad_norm": 0.064453125,
317
+ "learning_rate": 0.00015307523701080768,
318
+ "loss": 0.14675980806350708,
319
+ "mean_token_accuracy": 0.9623546004295349,
320
+ "num_tokens": 889907.0,
321
+ "step": 290
322
+ },
323
+ {
324
+ "entropy": 0.13382284864783286,
325
+ "epoch": 1.0676758682101515,
326
+ "grad_norm": 0.060546875,
327
+ "learning_rate": 0.00014977782868636999,
328
+ "loss": 0.13706474304199218,
329
+ "mean_token_accuracy": 0.9633408606052398,
330
+ "num_tokens": 920639.0,
331
+ "step": 300
332
+ },
333
+ {
334
+ "entropy": 0.16596811451017857,
335
+ "epoch": 1.1032947462154943,
336
+ "grad_norm": 0.0791015625,
337
+ "learning_rate": 0.00014640682709753832,
338
+ "loss": 0.16318607330322266,
339
+ "mean_token_accuracy": 0.9578790530562401,
340
+ "num_tokens": 951323.0,
341
+ "step": 310
342
+ },
343
+ {
344
+ "entropy": 0.1480622159317136,
345
+ "epoch": 1.138913624220837,
346
+ "grad_norm": 0.057861328125,
347
+ "learning_rate": 0.0001429672160497085,
348
+ "loss": 0.1488279938697815,
349
+ "mean_token_accuracy": 0.9609164595603943,
350
+ "num_tokens": 981911.0,
351
+ "step": 320
352
+ },
353
+ {
354
+ "entropy": 0.1655621325597167,
355
+ "epoch": 1.1745325022261799,
356
+ "grad_norm": 0.07275390625,
357
+ "learning_rate": 0.00013946408078287462,
358
+ "loss": 0.16427644491195678,
359
+ "mean_token_accuracy": 0.9576473370194435,
360
+ "num_tokens": 1012535.0,
361
+ "step": 330
362
+ },
363
+ {
364
+ "entropy": 0.18122829273343086,
365
+ "epoch": 1.2101513802315227,
366
+ "grad_norm": 0.06689453125,
367
+ "learning_rate": 0.00013590260045343432,
368
+ "loss": 0.176006543636322,
369
+ "mean_token_accuracy": 0.9551453411579132,
370
+ "num_tokens": 1043503.0,
371
+ "step": 340
372
+ },
373
+ {
374
+ "entropy": 0.15644419118762015,
375
+ "epoch": 1.2457702582368655,
376
+ "grad_norm": 0.06201171875,
377
+ "learning_rate": 0.00013228804047714463,
378
+ "loss": 0.1652477502822876,
379
+ "mean_token_accuracy": 0.9575047269463539,
380
+ "num_tokens": 1074168.0,
381
+ "step": 350
382
+ },
383
+ {
384
+ "entropy": 0.14828295167535543,
385
+ "epoch": 1.2813891362422083,
386
+ "grad_norm": 0.06982421875,
387
+ "learning_rate": 0.00012862574474454928,
388
+ "loss": 0.14066768884658815,
389
+ "mean_token_accuracy": 0.9639188721776009,
390
+ "num_tokens": 1104681.0,
391
+ "step": 360
392
+ },
393
+ {
394
+ "entropy": 0.19175196047872306,
395
+ "epoch": 1.317008014247551,
396
+ "grad_norm": 0.09619140625,
397
+ "learning_rate": 0.0001249211277203859,
398
+ "loss": 0.1917089343070984,
399
+ "mean_token_accuracy": 0.9522068575024605,
400
+ "num_tokens": 1135361.0,
401
+ "step": 370
402
+ },
403
+ {
404
+ "entropy": 0.1644950734451413,
405
+ "epoch": 1.3526268922528941,
406
+ "grad_norm": 0.0849609375,
407
+ "learning_rate": 0.00012117966643865398,
408
+ "loss": 0.16864393949508666,
409
+ "mean_token_accuracy": 0.9572859182953835,
410
+ "num_tokens": 1166125.0,
411
+ "step": 380
412
+ },
413
+ {
414
+ "entropy": 0.16431492734700442,
415
+ "epoch": 1.388245770258237,
416
+ "grad_norm": 0.08056640625,
417
+ "learning_rate": 0.00011740689240517837,
418
+ "loss": 0.1597315788269043,
419
+ "mean_token_accuracy": 0.9589557304978371,
420
+ "num_tokens": 1197232.0,
421
+ "step": 390
422
+ },
423
+ {
424
+ "entropy": 0.15399955678731203,
425
+ "epoch": 1.4238646482635797,
426
+ "grad_norm": 0.07421875,
427
+ "learning_rate": 0.00011360838341963964,
428
+ "loss": 0.15647656917572023,
429
+ "mean_token_accuracy": 0.9591503396630288,
430
+ "num_tokens": 1227801.0,
431
+ "step": 400
432
+ },
433
+ {
434
+ "entropy": 0.15702628958970308,
435
+ "epoch": 1.4594835262689225,
436
+ "grad_norm": 0.0888671875,
437
+ "learning_rate": 0.00010978975532916189,
438
+ "loss": 0.16044070720672607,
439
+ "mean_token_accuracy": 0.958300518989563,
440
+ "num_tokens": 1258629.0,
441
+ "step": 410
442
+ },
443
+ {
444
+ "entropy": 0.12639004811644555,
445
+ "epoch": 1.4951024042742653,
446
+ "grad_norm": 0.0654296875,
447
+ "learning_rate": 0.00010595665372565027,
448
+ "loss": 0.12638626098632813,
449
+ "mean_token_accuracy": 0.9659075498580932,
450
+ "num_tokens": 1289072.0,
451
+ "step": 420
452
+ },
453
+ {
454
+ "entropy": 0.14171069134026765,
455
+ "epoch": 1.5307212822796084,
456
+ "grad_norm": 0.08056640625,
457
+ "learning_rate": 0.00010211474559915233,
458
+ "loss": 0.13931651115417482,
459
+ "mean_token_accuracy": 0.9626615524291993,
460
+ "num_tokens": 1319747.0,
461
+ "step": 430
462
+ },
463
+ {
464
+ "entropy": 0.16094463262706996,
465
+ "epoch": 1.566340160284951,
466
+ "grad_norm": 0.07958984375,
467
+ "learning_rate": 9.826971095958395e-05,
468
+ "loss": 0.15868637561798096,
469
+ "mean_token_accuracy": 0.9590194016695023,
470
+ "num_tokens": 1350664.0,
471
+ "step": 440
472
+ },
473
+ {
474
+ "entropy": 0.15832340456545352,
475
+ "epoch": 1.601959038290294,
476
+ "grad_norm": 0.0673828125,
477
+ "learning_rate": 9.442723443920623e-05,
478
+ "loss": 0.16317789554595946,
479
+ "mean_token_accuracy": 0.9585595563054085,
480
+ "num_tokens": 1381628.0,
481
+ "step": 450
482
+ },
483
+ {
484
+ "entropy": 0.16448966227471828,
485
+ "epoch": 1.6375779162956365,
486
+ "grad_norm": 0.0751953125,
487
+ "learning_rate": 9.059299688826816e-05,
488
+ "loss": 0.16371761560440062,
489
+ "mean_token_accuracy": 0.9589671149849892,
490
+ "num_tokens": 1412506.0,
491
+ "step": 460
492
+ },
493
+ {
494
+ "entropy": 0.15921370945870877,
495
+ "epoch": 1.6731967943009796,
496
+ "grad_norm": 0.0732421875,
497
+ "learning_rate": 8.677266697624138e-05,
498
+ "loss": 0.15553101301193237,
499
+ "mean_token_accuracy": 0.9597976416349411,
500
+ "num_tokens": 1442986.0,
501
+ "step": 470
502
+ },
503
+ {
504
+ "entropy": 0.14739455822855235,
505
+ "epoch": 1.7088156723063224,
506
+ "grad_norm": 0.0751953125,
507
+ "learning_rate": 8.297189281106278e-05,
508
+ "loss": 0.14005433320999144,
509
+ "mean_token_accuracy": 0.9631642028689384,
510
+ "num_tokens": 1473444.0,
511
+ "step": 480
512
+ },
513
+ {
514
+ "entropy": 0.1320339234545827,
515
+ "epoch": 1.7444345503116652,
516
+ "grad_norm": 0.07275390625,
517
+ "learning_rate": 7.919629358877657e-05,
518
+ "loss": 0.13607435226440429,
519
+ "mean_token_accuracy": 0.9643332988023758,
520
+ "num_tokens": 1504098.0,
521
+ "step": 490
522
+ },
523
+ {
524
+ "entropy": 0.12228272054344416,
525
+ "epoch": 1.780053428317008,
526
+ "grad_norm": 0.0888671875,
527
+ "learning_rate": 7.54514512859201e-05,
528
+ "loss": 0.12318435907363892,
529
+ "mean_token_accuracy": 0.9668563097715378,
530
+ "num_tokens": 1534785.0,
531
+ "step": 500
532
+ },
533
+ {
534
+ "entropy": 0.14428242221474646,
535
+ "epoch": 1.8156723063223508,
536
+ "grad_norm": 0.07861328125,
537
+ "learning_rate": 7.174290240693689e-05,
538
+ "loss": 0.14451712369918823,
539
+ "mean_token_accuracy": 0.9615424692630767,
540
+ "num_tokens": 1565500.0,
541
+ "step": 510
542
+ },
543
+ {
544
+ "entropy": 0.1427522897720337,
545
+ "epoch": 1.8512911843276938,
546
+ "grad_norm": 0.056884765625,
547
+ "learning_rate": 6.807612979881661e-05,
548
+ "loss": 0.1388334035873413,
549
+ "mean_token_accuracy": 0.9621888875961304,
550
+ "num_tokens": 1596446.0,
551
+ "step": 520
552
+ },
553
+ {
554
+ "entropy": 0.15157688688486814,
555
+ "epoch": 1.8869100623330364,
556
+ "grad_norm": 0.08349609375,
557
+ "learning_rate": 6.445655454506465e-05,
558
+ "loss": 0.15446548461914061,
559
+ "mean_token_accuracy": 0.9599128782749176,
560
+ "num_tokens": 1627189.0,
561
+ "step": 530
562
+ },
563
+ {
564
+ "entropy": 0.1300442773848772,
565
+ "epoch": 1.9225289403383794,
566
+ "grad_norm": 0.059326171875,
567
+ "learning_rate": 6.0889527950984416e-05,
568
+ "loss": 0.1254338026046753,
569
+ "mean_token_accuracy": 0.9661760002374649,
570
+ "num_tokens": 1657858.0,
571
+ "step": 540
572
+ },
573
+ {
574
+ "entropy": 0.1557828625664115,
575
+ "epoch": 1.9581478183437222,
576
+ "grad_norm": 0.06689453125,
577
+ "learning_rate": 5.738032363212258e-05,
578
+ "loss": 0.16083067655563354,
579
+ "mean_token_accuracy": 0.9595108240842819,
580
+ "num_tokens": 1688776.0,
581
+ "step": 550
582
+ },
583
+ {
584
+ "entropy": 0.13151839561760426,
585
+ "epoch": 1.993766696349065,
586
+ "grad_norm": 0.0810546875,
587
+ "learning_rate": 5.3934129717573165e-05,
588
+ "loss": 0.12607554197311402,
589
+ "mean_token_accuracy": 0.966105441749096,
590
+ "num_tokens": 1719573.0,
591
+ "step": 560
592
+ },
593
+ {
594
+ "epoch": 2.0,
595
+ "eval_entropy": 0.14480842649936676,
596
+ "eval_loss": 0.15477769076824188,
597
+ "eval_mean_token_accuracy": 0.9607759993127052,
598
+ "eval_num_tokens": 1725024.0,
599
+ "eval_runtime": 35.306,
600
+ "eval_samples_per_second": 15.89,
601
+ "eval_steps_per_second": 3.994,
602
+ "step": 562
603
+ }
604
+ ],
605
+ "logging_steps": 10,
606
+ "max_steps": 843,
607
+ "num_input_tokens_seen": 0,
608
+ "num_train_epochs": 3,
609
+ "save_steps": 500,
610
+ "stateful_callbacks": {
611
+ "TrainerControl": {
612
+ "args": {
613
+ "should_epoch_stop": false,
614
+ "should_evaluate": false,
615
+ "should_log": false,
616
+ "should_save": true,
617
+ "should_training_stop": false
618
+ },
619
+ "attributes": {}
620
+ }
621
+ },
622
+ "total_flos": 4.057603725318144e+16,
623
+ "train_batch_size": 4,
624
+ "trial_name": null,
625
+ "trial_params": null
626
+ }
checkpoint-562/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57851b427cb3a4649963b62f1b5d8860aa50ba7ac4abbbb2ebc56acbf65bd694
3
+ size 5649
checkpoint-843/README.md ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen3-4B-Instruct-2507
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:Qwen/Qwen3-4B-Instruct-2507
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ ---
12
+
13
+ # Model Card for Model ID
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ <!-- Provide a longer summary of what this model is. -->
24
+
25
+
26
+
27
+ - **Developed by:** [More Information Needed]
28
+ - **Funded by [optional]:** [More Information Needed]
29
+ - **Shared by [optional]:** [More Information Needed]
30
+ - **Model type:** [More Information Needed]
31
+ - **Language(s) (NLP):** [More Information Needed]
32
+ - **License:** [More Information Needed]
33
+ - **Finetuned from model [optional]:** [More Information Needed]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ <!-- Provide the basic links for the model. -->
38
+
39
+ - **Repository:** [More Information Needed]
40
+ - **Paper [optional]:** [More Information Needed]
41
+ - **Demo [optional]:** [More Information Needed]
42
+
43
+ ## Uses
44
+
45
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
+
47
+ ### Direct Use
48
+
49
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Downstream Use [optional]
54
+
55
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+
57
+ [More Information Needed]
58
+
59
+ ### Out-of-Scope Use
60
+
61
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
74
+
75
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
76
+
77
+ ## How to Get Started with the Model
78
+
79
+ Use the code below to get started with the model.
80
+
81
+ [More Information Needed]
82
+
83
+ ## Training Details
84
+
85
+ ### Training Data
86
+
87
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
+
89
+ [More Information Needed]
90
+
91
+ ### Training Procedure
92
+
93
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
+
95
+ #### Preprocessing [optional]
96
+
97
+ [More Information Needed]
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
103
+
104
+ #### Speeds, Sizes, Times [optional]
105
+
106
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ## Evaluation
111
+
112
+ <!-- This section describes the evaluation protocols and provides the results. -->
113
+
114
+ ### Testing Data, Factors & Metrics
115
+
116
+ #### Testing Data
117
+
118
+ <!-- This should link to a Dataset Card if possible. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Factors
123
+
124
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
125
+
126
+ [More Information Needed]
127
+
128
+ #### Metrics
129
+
130
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
131
+
132
+ [More Information Needed]
133
+
134
+ ### Results
135
+
136
+ [More Information Needed]
137
+
138
+ #### Summary
139
+
140
+
141
+
142
+ ## Model Examination [optional]
143
+
144
+ <!-- Relevant interpretability work for the model goes here -->
145
+
146
+ [More Information Needed]
147
+
148
+ ## Environmental Impact
149
+
150
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
151
+
152
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
153
+
154
+ - **Hardware Type:** [More Information Needed]
155
+ - **Hours used:** [More Information Needed]
156
+ - **Cloud Provider:** [More Information Needed]
157
+ - **Compute Region:** [More Information Needed]
158
+ - **Carbon Emitted:** [More Information Needed]
159
+
160
+ ## Technical Specifications [optional]
161
+
162
+ ### Model Architecture and Objective
163
+
164
+ [More Information Needed]
165
+
166
+ ### Compute Infrastructure
167
+
168
+ [More Information Needed]
169
+
170
+ #### Hardware
171
+
172
+ [More Information Needed]
173
+
174
+ #### Software
175
+
176
+ [More Information Needed]
177
+
178
+ ## Citation [optional]
179
+
180
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
181
+
182
+ **BibTeX:**
183
+
184
+ [More Information Needed]
185
+
186
+ **APA:**
187
+
188
+ [More Information Needed]
189
+
190
+ ## Glossary [optional]
191
+
192
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
193
+
194
+ [More Information Needed]
195
+
196
+ ## More Information [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Authors [optional]
201
+
202
+ [More Information Needed]
203
+
204
+ ## Model Card Contact
205
+
206
+ [More Information Needed]
207
+ ### Framework versions
208
+
209
+ - PEFT 0.18.1
checkpoint-843/adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "Qwen/Qwen3-4B-Instruct-2507",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 32,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": null,
25
+ "peft_type": "LORA",
26
+ "peft_version": "0.18.1",
27
+ "qalora_group_size": 16,
28
+ "r": 16,
29
+ "rank_pattern": {},
30
+ "revision": null,
31
+ "target_modules": [
32
+ "v_proj",
33
+ "down_proj",
34
+ "up_proj",
35
+ "q_proj",
36
+ "gate_proj",
37
+ "k_proj",
38
+ "o_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
checkpoint-843/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fef732bbd8022ce3c2bda83792e299ed5c35b1ae1dd373c9ef1b5ff2fa61c16
3
+ size 66127776
checkpoint-843/chat_template.jinja ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %}
2
+ {{- '<|im_start|>system\n' }}
3
+ {%- if messages[0].role == 'system' %}
4
+ {{- messages[0].content + '\n\n' }}
5
+ {%- endif %}
6
+ {{- "# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within <tools></tools> XML tags:\n<tools>" }}
7
+ {%- for tool in tools %}
8
+ {{- "\n" }}
9
+ {{- tool | tojson }}
10
+ {%- endfor %}
11
+ {{- "\n</tools>\n\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\n<tool_call>\n{\"name\": <function-name>, \"arguments\": <args-json-object>}\n</tool_call><|im_end|>\n" }}
12
+ {%- else %}
13
+ {%- if messages[0].role == 'system' %}
14
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
15
+ {%- endif %}
16
+ {%- endif %}
17
+ {%- for message in messages %}
18
+ {%- if message.content is string %}
19
+ {%- set content = message.content %}
20
+ {%- else %}
21
+ {%- set content = '' %}
22
+ {%- endif %}
23
+ {%- if (message.role == "user") or (message.role == "system" and not loop.first) %}
24
+ {{- '<|im_start|>' + message.role + '\n' + content + '<|im_end|>' + '\n' }}
25
+ {%- elif message.role == "assistant" %}
26
+ {{- '<|im_start|>' + message.role + '\n' + content }}
27
+ {%- if message.tool_calls %}
28
+ {%- for tool_call in message.tool_calls %}
29
+ {%- if (loop.first and content) or (not loop.first) %}
30
+ {{- '\n' }}
31
+ {%- endif %}
32
+ {%- if tool_call.function %}
33
+ {%- set tool_call = tool_call.function %}
34
+ {%- endif %}
35
+ {{- '<tool_call>\n{"name": "' }}
36
+ {{- tool_call.name }}
37
+ {{- '", "arguments": ' }}
38
+ {%- if tool_call.arguments is string %}
39
+ {{- tool_call.arguments }}
40
+ {%- else %}
41
+ {{- tool_call.arguments | tojson }}
42
+ {%- endif %}
43
+ {{- '}\n</tool_call>' }}
44
+ {%- endfor %}
45
+ {%- endif %}
46
+ {{- '<|im_end|>\n' }}
47
+ {%- elif message.role == "tool" %}
48
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
49
+ {{- '<|im_start|>user' }}
50
+ {%- endif %}
51
+ {{- '\n<tool_response>\n' }}
52
+ {{- content }}
53
+ {{- '\n</tool_response>' }}
54
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
55
+ {{- '<|im_end|>\n' }}
56
+ {%- endif %}
57
+ {%- endif %}
58
+ {%- endfor %}
59
+ {%- if add_generation_prompt %}
60
+ {{- '<|im_start|>assistant\n' }}
61
+ {%- endif %}
checkpoint-843/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26bf1ba2744c6b96164983ec84962f35fc51660dafcac46c99a5b1ed5e5d61ea
3
+ size 132552715
checkpoint-843/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d302a10b4bf78a5b33d836d0dba74ecc5c351791675f6c0e73639bbf88acd3df
3
+ size 14645
checkpoint-843/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f3eaaa49ad0633eb0127e8e12aac9b83ab64a686d17ee15c6d799822cc33157
3
+ size 1465
checkpoint-843/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be75606093db2094d7cd20f3c2f385c212750648bd6ea4fb2bf507a6a4c55506
3
+ size 11422650
checkpoint-843/tokenizer_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": null,
5
+ "clean_up_tokenization_spaces": false,
6
+ "eos_token": "<|im_end|>",
7
+ "errors": "replace",
8
+ "extra_special_tokens": [
9
+ "<|im_start|>",
10
+ "<|im_end|>",
11
+ "<|object_ref_start|>",
12
+ "<|object_ref_end|>",
13
+ "<|box_start|>",
14
+ "<|box_end|>",
15
+ "<|quad_start|>",
16
+ "<|quad_end|>",
17
+ "<|vision_start|>",
18
+ "<|vision_end|>",
19
+ "<|vision_pad|>",
20
+ "<|image_pad|>",
21
+ "<|video_pad|>"
22
+ ],
23
+ "is_local": false,
24
+ "model_max_length": 1010000,
25
+ "pad_token": "<|im_end|>",
26
+ "split_special_tokens": false,
27
+ "tokenizer_class": "Qwen2Tokenizer",
28
+ "unk_token": null
29
+ }
checkpoint-843/trainer_state.json ADDED
@@ -0,0 +1,917 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 562,
3
+ "best_metric": 0.15477769076824188,
4
+ "best_model_checkpoint": "/app/output/bf-router-v0.5/checkpoint-562",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 843,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "entropy": 1.4157426357269287,
14
+ "epoch": 0.003561887800534283,
15
+ "grad_norm": 1.203125,
16
+ "learning_rate": 0.0,
17
+ "loss": 4.336674213409424,
18
+ "mean_token_accuracy": 0.4680640399456024,
19
+ "num_tokens": 3092.0,
20
+ "step": 1
21
+ },
22
+ {
23
+ "entropy": 1.5884604156017303,
24
+ "epoch": 0.03561887800534283,
25
+ "grad_norm": 0.59375,
26
+ "learning_rate": 6.923076923076924e-05,
27
+ "loss": 3.986017862955729,
28
+ "mean_token_accuracy": 0.4735516524977154,
29
+ "num_tokens": 30650.0,
30
+ "step": 10
31
+ },
32
+ {
33
+ "entropy": 2.2315500736236573,
34
+ "epoch": 0.07123775601068566,
35
+ "grad_norm": 0.26953125,
36
+ "learning_rate": 0.00014615384615384615,
37
+ "loss": 2.16544189453125,
38
+ "mean_token_accuracy": 0.6272616922855377,
39
+ "num_tokens": 61735.0,
40
+ "step": 20
41
+ },
42
+ {
43
+ "entropy": 0.8066075079143047,
44
+ "epoch": 0.10685663401602849,
45
+ "grad_norm": 0.298828125,
46
+ "learning_rate": 0.00019999334629824895,
47
+ "loss": 0.7935717582702637,
48
+ "mean_token_accuracy": 0.857068446278572,
49
+ "num_tokens": 92484.0,
50
+ "step": 30
51
+ },
52
+ {
53
+ "entropy": 0.5340980306267739,
54
+ "epoch": 0.14247551202137132,
55
+ "grad_norm": 0.10693359375,
56
+ "learning_rate": 0.00019987508289735716,
57
+ "loss": 0.5261529922485352,
58
+ "mean_token_accuracy": 0.9004581302404404,
59
+ "num_tokens": 123010.0,
60
+ "step": 40
61
+ },
62
+ {
63
+ "entropy": 0.4001643668860197,
64
+ "epoch": 0.17809439002671415,
65
+ "grad_norm": 0.08251953125,
66
+ "learning_rate": 0.0001996091607179287,
67
+ "loss": 0.3956298351287842,
68
+ "mean_token_accuracy": 0.9209299519658088,
69
+ "num_tokens": 153710.0,
70
+ "step": 50
71
+ },
72
+ {
73
+ "entropy": 0.34828933551907537,
74
+ "epoch": 0.21371326803205698,
75
+ "grad_norm": 0.10009765625,
76
+ "learning_rate": 0.00019919597290851538,
77
+ "loss": 0.3648662090301514,
78
+ "mean_token_accuracy": 0.9263768374919892,
79
+ "num_tokens": 184577.0,
80
+ "step": 60
81
+ },
82
+ {
83
+ "entropy": 0.2983129292726517,
84
+ "epoch": 0.2493321460373998,
85
+ "grad_norm": 0.08447265625,
86
+ "learning_rate": 0.00019863613034027224,
87
+ "loss": 0.29642860889434813,
88
+ "mean_token_accuracy": 0.9371836110949516,
89
+ "num_tokens": 215040.0,
90
+ "step": 70
91
+ },
92
+ {
93
+ "entropy": 0.280751570686698,
94
+ "epoch": 0.28495102404274264,
95
+ "grad_norm": 0.099609375,
96
+ "learning_rate": 0.00019793046070382437,
97
+ "loss": 0.2813178300857544,
98
+ "mean_token_accuracy": 0.9388988897204399,
99
+ "num_tokens": 245683.0,
100
+ "step": 80
101
+ },
102
+ {
103
+ "entropy": 0.2960826952010393,
104
+ "epoch": 0.3205699020480855,
105
+ "grad_norm": 0.11962890625,
106
+ "learning_rate": 0.00019708000728558064,
107
+ "loss": 0.31115410327911375,
108
+ "mean_token_accuracy": 0.9335825845599175,
109
+ "num_tokens": 276430.0,
110
+ "step": 90
111
+ },
112
+ {
113
+ "entropy": 0.28416929207742214,
114
+ "epoch": 0.3561887800534283,
115
+ "grad_norm": 0.103515625,
116
+ "learning_rate": 0.00019608602742530283,
117
+ "loss": 0.27938365936279297,
118
+ "mean_token_accuracy": 0.9393619552254677,
119
+ "num_tokens": 307296.0,
120
+ "step": 100
121
+ },
122
+ {
123
+ "entropy": 0.23417662251740695,
124
+ "epoch": 0.39180765805877116,
125
+ "grad_norm": 0.06298828125,
126
+ "learning_rate": 0.00019494999065721108,
127
+ "loss": 0.23734774589538574,
128
+ "mean_token_accuracy": 0.9487677246332169,
129
+ "num_tokens": 337962.0,
130
+ "step": 110
131
+ },
132
+ {
133
+ "entropy": 0.24019648060202597,
134
+ "epoch": 0.42742653606411396,
135
+ "grad_norm": 0.06640625,
136
+ "learning_rate": 0.0001936735765373737,
137
+ "loss": 0.245324969291687,
138
+ "mean_token_accuracy": 0.9457737937569618,
139
+ "num_tokens": 368782.0,
140
+ "step": 120
141
+ },
142
+ {
143
+ "entropy": 0.269666058011353,
144
+ "epoch": 0.4630454140694568,
145
+ "grad_norm": 0.09765625,
146
+ "learning_rate": 0.00019225867216059325,
147
+ "loss": 0.2686375617980957,
148
+ "mean_token_accuracy": 0.9398508563637733,
149
+ "num_tokens": 399329.0,
150
+ "step": 130
151
+ },
152
+ {
153
+ "entropy": 0.18328354582190515,
154
+ "epoch": 0.4986642920747996,
155
+ "grad_norm": 0.06298828125,
156
+ "learning_rate": 0.00019070736937046035,
157
+ "loss": 0.18070420026779174,
158
+ "mean_token_accuracy": 0.9568557634949684,
159
+ "num_tokens": 429759.0,
160
+ "step": 140
161
+ },
162
+ {
163
+ "entropy": 0.21165105439722537,
164
+ "epoch": 0.5342831700801425,
165
+ "grad_norm": 0.0810546875,
166
+ "learning_rate": 0.0001890219616666997,
167
+ "loss": 0.22004930973052977,
168
+ "mean_token_accuracy": 0.9494227185845375,
169
+ "num_tokens": 460409.0,
170
+ "step": 150
171
+ },
172
+ {
173
+ "entropy": 0.23916025627404452,
174
+ "epoch": 0.5699020480854853,
175
+ "grad_norm": 0.07568359375,
176
+ "learning_rate": 0.00018720494081438078,
177
+ "loss": 0.2379377841949463,
178
+ "mean_token_accuracy": 0.9455192387104034,
179
+ "num_tokens": 491344.0,
180
+ "step": 160
181
+ },
182
+ {
183
+ "entropy": 0.2238690486177802,
184
+ "epoch": 0.6055209260908282,
185
+ "grad_norm": 0.08984375,
186
+ "learning_rate": 0.00018525899316000608,
187
+ "loss": 0.2245168685913086,
188
+ "mean_token_accuracy": 0.9485938593745231,
189
+ "num_tokens": 522029.0,
190
+ "step": 170
191
+ },
192
+ {
193
+ "entropy": 0.17032606173306702,
194
+ "epoch": 0.641139804096171,
195
+ "grad_norm": 0.080078125,
196
+ "learning_rate": 0.00018318699565992357,
197
+ "loss": 0.17549347877502441,
198
+ "mean_token_accuracy": 0.9574712902307511,
199
+ "num_tokens": 552306.0,
200
+ "step": 180
201
+ },
202
+ {
203
+ "entropy": 0.21890966054052113,
204
+ "epoch": 0.6767586821015138,
205
+ "grad_norm": 0.0673828125,
206
+ "learning_rate": 0.00018099201162693476,
207
+ "loss": 0.21968293190002441,
208
+ "mean_token_accuracy": 0.9482985377311707,
209
+ "num_tokens": 583181.0,
210
+ "step": 190
211
+ },
212
+ {
213
+ "entropy": 0.20646683853119613,
214
+ "epoch": 0.7123775601068566,
215
+ "grad_norm": 0.09765625,
216
+ "learning_rate": 0.00017867728620138708,
217
+ "loss": 0.20968315601348878,
218
+ "mean_token_accuracy": 0.9519409075379371,
219
+ "num_tokens": 613818.0,
220
+ "step": 200
221
+ },
222
+ {
223
+ "entropy": 0.23000301327556372,
224
+ "epoch": 0.7479964381121995,
225
+ "grad_norm": 0.08935546875,
226
+ "learning_rate": 0.00017624624155344626,
227
+ "loss": 0.2383474349975586,
228
+ "mean_token_accuracy": 0.9458372846245766,
229
+ "num_tokens": 644576.0,
230
+ "step": 210
231
+ },
232
+ {
233
+ "entropy": 0.21420854832977057,
234
+ "epoch": 0.7836153161175423,
235
+ "grad_norm": 0.0966796875,
236
+ "learning_rate": 0.0001737024718236413,
237
+ "loss": 0.21017465591430665,
238
+ "mean_token_accuracy": 0.9511569887399673,
239
+ "num_tokens": 675564.0,
240
+ "step": 220
241
+ },
242
+ {
243
+ "entropy": 0.19280126914381981,
244
+ "epoch": 0.8192341941228851,
245
+ "grad_norm": 0.08154296875,
246
+ "learning_rate": 0.00017104973780916294,
247
+ "loss": 0.1936139941215515,
248
+ "mean_token_accuracy": 0.954449312388897,
249
+ "num_tokens": 706309.0,
250
+ "step": 230
251
+ },
252
+ {
253
+ "entropy": 0.22524890769273043,
254
+ "epoch": 0.8548530721282279,
255
+ "grad_norm": 0.10546875,
256
+ "learning_rate": 0.00016829196140377085,
257
+ "loss": 0.22950620651245118,
258
+ "mean_token_accuracy": 0.9476036429405212,
259
+ "num_tokens": 737135.0,
260
+ "step": 240
261
+ },
262
+ {
263
+ "entropy": 0.18842433094978334,
264
+ "epoch": 0.8904719501335708,
265
+ "grad_norm": 0.087890625,
266
+ "learning_rate": 0.00016543321979953007,
267
+ "loss": 0.19049547910690307,
268
+ "mean_token_accuracy": 0.9543813273310662,
269
+ "num_tokens": 767853.0,
270
+ "step": 250
271
+ },
272
+ {
273
+ "entropy": 0.1991994746029377,
274
+ "epoch": 0.9260908281389136,
275
+ "grad_norm": 0.07763671875,
276
+ "learning_rate": 0.00016247773945894962,
277
+ "loss": 0.19527161121368408,
278
+ "mean_token_accuracy": 0.9535173490643502,
279
+ "num_tokens": 798511.0,
280
+ "step": 260
281
+ },
282
+ {
283
+ "entropy": 0.18196075949817897,
284
+ "epoch": 0.9617097061442564,
285
+ "grad_norm": 0.0771484375,
286
+ "learning_rate": 0.00015942988986643352,
287
+ "loss": 0.19091761112213135,
288
+ "mean_token_accuracy": 0.9535432115197182,
289
+ "num_tokens": 829447.0,
290
+ "step": 270
291
+ },
292
+ {
293
+ "entropy": 0.18307169582694768,
294
+ "epoch": 0.9973285841495992,
295
+ "grad_norm": 0.08837890625,
296
+ "learning_rate": 0.00015629417706828423,
297
+ "loss": 0.1818631410598755,
298
+ "mean_token_accuracy": 0.955574706196785,
299
+ "num_tokens": 860142.0,
300
+ "step": 280
301
+ },
302
+ {
303
+ "epoch": 1.0,
304
+ "eval_entropy": 0.18207345770817276,
305
+ "eval_loss": 0.18149949610233307,
306
+ "eval_mean_token_accuracy": 0.9555717660180221,
307
+ "eval_num_tokens": 862512.0,
308
+ "eval_runtime": 35.3256,
309
+ "eval_samples_per_second": 15.881,
310
+ "eval_steps_per_second": 3.991,
311
+ "step": 281
312
+ },
313
+ {
314
+ "entropy": 0.1545313375118451,
315
+ "epoch": 1.0320569902048085,
316
+ "grad_norm": 0.064453125,
317
+ "learning_rate": 0.00015307523701080768,
318
+ "loss": 0.14675980806350708,
319
+ "mean_token_accuracy": 0.9623546004295349,
320
+ "num_tokens": 889907.0,
321
+ "step": 290
322
+ },
323
+ {
324
+ "entropy": 0.13382284864783286,
325
+ "epoch": 1.0676758682101515,
326
+ "grad_norm": 0.060546875,
327
+ "learning_rate": 0.00014977782868636999,
328
+ "loss": 0.13706474304199218,
329
+ "mean_token_accuracy": 0.9633408606052398,
330
+ "num_tokens": 920639.0,
331
+ "step": 300
332
+ },
333
+ {
334
+ "entropy": 0.16596811451017857,
335
+ "epoch": 1.1032947462154943,
336
+ "grad_norm": 0.0791015625,
337
+ "learning_rate": 0.00014640682709753832,
338
+ "loss": 0.16318607330322266,
339
+ "mean_token_accuracy": 0.9578790530562401,
340
+ "num_tokens": 951323.0,
341
+ "step": 310
342
+ },
343
+ {
344
+ "entropy": 0.1480622159317136,
345
+ "epoch": 1.138913624220837,
346
+ "grad_norm": 0.057861328125,
347
+ "learning_rate": 0.0001429672160497085,
348
+ "loss": 0.1488279938697815,
349
+ "mean_token_accuracy": 0.9609164595603943,
350
+ "num_tokens": 981911.0,
351
+ "step": 320
352
+ },
353
+ {
354
+ "entropy": 0.1655621325597167,
355
+ "epoch": 1.1745325022261799,
356
+ "grad_norm": 0.07275390625,
357
+ "learning_rate": 0.00013946408078287462,
358
+ "loss": 0.16427644491195678,
359
+ "mean_token_accuracy": 0.9576473370194435,
360
+ "num_tokens": 1012535.0,
361
+ "step": 330
362
+ },
363
+ {
364
+ "entropy": 0.18122829273343086,
365
+ "epoch": 1.2101513802315227,
366
+ "grad_norm": 0.06689453125,
367
+ "learning_rate": 0.00013590260045343432,
368
+ "loss": 0.176006543636322,
369
+ "mean_token_accuracy": 0.9551453411579132,
370
+ "num_tokens": 1043503.0,
371
+ "step": 340
372
+ },
373
+ {
374
+ "entropy": 0.15644419118762015,
375
+ "epoch": 1.2457702582368655,
376
+ "grad_norm": 0.06201171875,
377
+ "learning_rate": 0.00013228804047714463,
378
+ "loss": 0.1652477502822876,
379
+ "mean_token_accuracy": 0.9575047269463539,
380
+ "num_tokens": 1074168.0,
381
+ "step": 350
382
+ },
383
+ {
384
+ "entropy": 0.14828295167535543,
385
+ "epoch": 1.2813891362422083,
386
+ "grad_norm": 0.06982421875,
387
+ "learning_rate": 0.00012862574474454928,
388
+ "loss": 0.14066768884658815,
389
+ "mean_token_accuracy": 0.9639188721776009,
390
+ "num_tokens": 1104681.0,
391
+ "step": 360
392
+ },
393
+ {
394
+ "entropy": 0.19175196047872306,
395
+ "epoch": 1.317008014247551,
396
+ "grad_norm": 0.09619140625,
397
+ "learning_rate": 0.0001249211277203859,
398
+ "loss": 0.1917089343070984,
399
+ "mean_token_accuracy": 0.9522068575024605,
400
+ "num_tokens": 1135361.0,
401
+ "step": 370
402
+ },
403
+ {
404
+ "entropy": 0.1644950734451413,
405
+ "epoch": 1.3526268922528941,
406
+ "grad_norm": 0.0849609375,
407
+ "learning_rate": 0.00012117966643865398,
408
+ "loss": 0.16864393949508666,
409
+ "mean_token_accuracy": 0.9572859182953835,
410
+ "num_tokens": 1166125.0,
411
+ "step": 380
412
+ },
413
+ {
414
+ "entropy": 0.16431492734700442,
415
+ "epoch": 1.388245770258237,
416
+ "grad_norm": 0.08056640625,
417
+ "learning_rate": 0.00011740689240517837,
418
+ "loss": 0.1597315788269043,
419
+ "mean_token_accuracy": 0.9589557304978371,
420
+ "num_tokens": 1197232.0,
421
+ "step": 390
422
+ },
423
+ {
424
+ "entropy": 0.15399955678731203,
425
+ "epoch": 1.4238646482635797,
426
+ "grad_norm": 0.07421875,
427
+ "learning_rate": 0.00011360838341963964,
428
+ "loss": 0.15647656917572023,
429
+ "mean_token_accuracy": 0.9591503396630288,
430
+ "num_tokens": 1227801.0,
431
+ "step": 400
432
+ },
433
+ {
434
+ "entropy": 0.15702628958970308,
435
+ "epoch": 1.4594835262689225,
436
+ "grad_norm": 0.0888671875,
437
+ "learning_rate": 0.00010978975532916189,
438
+ "loss": 0.16044070720672607,
439
+ "mean_token_accuracy": 0.958300518989563,
440
+ "num_tokens": 1258629.0,
441
+ "step": 410
442
+ },
443
+ {
444
+ "entropy": 0.12639004811644555,
445
+ "epoch": 1.4951024042742653,
446
+ "grad_norm": 0.0654296875,
447
+ "learning_rate": 0.00010595665372565027,
448
+ "loss": 0.12638626098632813,
449
+ "mean_token_accuracy": 0.9659075498580932,
450
+ "num_tokens": 1289072.0,
451
+ "step": 420
452
+ },
453
+ {
454
+ "entropy": 0.14171069134026765,
455
+ "epoch": 1.5307212822796084,
456
+ "grad_norm": 0.08056640625,
457
+ "learning_rate": 0.00010211474559915233,
458
+ "loss": 0.13931651115417482,
459
+ "mean_token_accuracy": 0.9626615524291993,
460
+ "num_tokens": 1319747.0,
461
+ "step": 430
462
+ },
463
+ {
464
+ "entropy": 0.16094463262706996,
465
+ "epoch": 1.566340160284951,
466
+ "grad_norm": 0.07958984375,
467
+ "learning_rate": 9.826971095958395e-05,
468
+ "loss": 0.15868637561798096,
469
+ "mean_token_accuracy": 0.9590194016695023,
470
+ "num_tokens": 1350664.0,
471
+ "step": 440
472
+ },
473
+ {
474
+ "entropy": 0.15832340456545352,
475
+ "epoch": 1.601959038290294,
476
+ "grad_norm": 0.0673828125,
477
+ "learning_rate": 9.442723443920623e-05,
478
+ "loss": 0.16317789554595946,
479
+ "mean_token_accuracy": 0.9585595563054085,
480
+ "num_tokens": 1381628.0,
481
+ "step": 450
482
+ },
483
+ {
484
+ "entropy": 0.16448966227471828,
485
+ "epoch": 1.6375779162956365,
486
+ "grad_norm": 0.0751953125,
487
+ "learning_rate": 9.059299688826816e-05,
488
+ "loss": 0.16371761560440062,
489
+ "mean_token_accuracy": 0.9589671149849892,
490
+ "num_tokens": 1412506.0,
491
+ "step": 460
492
+ },
493
+ {
494
+ "entropy": 0.15921370945870877,
495
+ "epoch": 1.6731967943009796,
496
+ "grad_norm": 0.0732421875,
497
+ "learning_rate": 8.677266697624138e-05,
498
+ "loss": 0.15553101301193237,
499
+ "mean_token_accuracy": 0.9597976416349411,
500
+ "num_tokens": 1442986.0,
501
+ "step": 470
502
+ },
503
+ {
504
+ "entropy": 0.14739455822855235,
505
+ "epoch": 1.7088156723063224,
506
+ "grad_norm": 0.0751953125,
507
+ "learning_rate": 8.297189281106278e-05,
508
+ "loss": 0.14005433320999144,
509
+ "mean_token_accuracy": 0.9631642028689384,
510
+ "num_tokens": 1473444.0,
511
+ "step": 480
512
+ },
513
+ {
514
+ "entropy": 0.1320339234545827,
515
+ "epoch": 1.7444345503116652,
516
+ "grad_norm": 0.07275390625,
517
+ "learning_rate": 7.919629358877657e-05,
518
+ "loss": 0.13607435226440429,
519
+ "mean_token_accuracy": 0.9643332988023758,
520
+ "num_tokens": 1504098.0,
521
+ "step": 490
522
+ },
523
+ {
524
+ "entropy": 0.12228272054344416,
525
+ "epoch": 1.780053428317008,
526
+ "grad_norm": 0.0888671875,
527
+ "learning_rate": 7.54514512859201e-05,
528
+ "loss": 0.12318435907363892,
529
+ "mean_token_accuracy": 0.9668563097715378,
530
+ "num_tokens": 1534785.0,
531
+ "step": 500
532
+ },
533
+ {
534
+ "entropy": 0.14428242221474646,
535
+ "epoch": 1.8156723063223508,
536
+ "grad_norm": 0.07861328125,
537
+ "learning_rate": 7.174290240693689e-05,
538
+ "loss": 0.14451712369918823,
539
+ "mean_token_accuracy": 0.9615424692630767,
540
+ "num_tokens": 1565500.0,
541
+ "step": 510
542
+ },
543
+ {
544
+ "entropy": 0.1427522897720337,
545
+ "epoch": 1.8512911843276938,
546
+ "grad_norm": 0.056884765625,
547
+ "learning_rate": 6.807612979881661e-05,
548
+ "loss": 0.1388334035873413,
549
+ "mean_token_accuracy": 0.9621888875961304,
550
+ "num_tokens": 1596446.0,
551
+ "step": 520
552
+ },
553
+ {
554
+ "entropy": 0.15157688688486814,
555
+ "epoch": 1.8869100623330364,
556
+ "grad_norm": 0.08349609375,
557
+ "learning_rate": 6.445655454506465e-05,
558
+ "loss": 0.15446548461914061,
559
+ "mean_token_accuracy": 0.9599128782749176,
560
+ "num_tokens": 1627189.0,
561
+ "step": 530
562
+ },
563
+ {
564
+ "entropy": 0.1300442773848772,
565
+ "epoch": 1.9225289403383794,
566
+ "grad_norm": 0.059326171875,
567
+ "learning_rate": 6.0889527950984416e-05,
568
+ "loss": 0.1254338026046753,
569
+ "mean_token_accuracy": 0.9661760002374649,
570
+ "num_tokens": 1657858.0,
571
+ "step": 540
572
+ },
573
+ {
574
+ "entropy": 0.1557828625664115,
575
+ "epoch": 1.9581478183437222,
576
+ "grad_norm": 0.06689453125,
577
+ "learning_rate": 5.738032363212258e-05,
578
+ "loss": 0.16083067655563354,
579
+ "mean_token_accuracy": 0.9595108240842819,
580
+ "num_tokens": 1688776.0,
581
+ "step": 550
582
+ },
583
+ {
584
+ "entropy": 0.13151839561760426,
585
+ "epoch": 1.993766696349065,
586
+ "grad_norm": 0.0810546875,
587
+ "learning_rate": 5.3934129717573165e-05,
588
+ "loss": 0.12607554197311402,
589
+ "mean_token_accuracy": 0.966105441749096,
590
+ "num_tokens": 1719573.0,
591
+ "step": 560
592
+ },
593
+ {
594
+ "epoch": 2.0,
595
+ "eval_entropy": 0.14480842649936676,
596
+ "eval_loss": 0.15477769076824188,
597
+ "eval_mean_token_accuracy": 0.9607759993127052,
598
+ "eval_num_tokens": 1725024.0,
599
+ "eval_runtime": 35.306,
600
+ "eval_samples_per_second": 15.89,
601
+ "eval_steps_per_second": 3.994,
602
+ "step": 562
603
+ },
604
+ {
605
+ "entropy": 0.13348992398151985,
606
+ "epoch": 2.0284951024042743,
607
+ "grad_norm": 0.09326171875,
608
+ "learning_rate": 5.0556041179668354e-05,
609
+ "loss": 0.12256227731704712,
610
+ "mean_token_accuracy": 0.9662964267608447,
611
+ "num_tokens": 1749504.0,
612
+ "step": 570
613
+ },
614
+ {
615
+ "entropy": 0.13551258575171232,
616
+ "epoch": 2.064113980409617,
617
+ "grad_norm": 0.0732421875,
618
+ "learning_rate": 4.725105230139465e-05,
619
+ "loss": 0.12811031341552734,
620
+ "mean_token_accuracy": 0.9640967190265656,
621
+ "num_tokens": 1780206.0,
622
+ "step": 580
623
+ },
624
+ {
625
+ "entropy": 0.11903938055038452,
626
+ "epoch": 2.09973285841496,
627
+ "grad_norm": 0.061279296875,
628
+ "learning_rate": 4.402404929267235e-05,
629
+ "loss": 0.1129370927810669,
630
+ "mean_token_accuracy": 0.9682316735386849,
631
+ "num_tokens": 1810985.0,
632
+ "step": 590
633
+ },
634
+ {
635
+ "entropy": 0.11653038449585437,
636
+ "epoch": 2.135351736420303,
637
+ "grad_norm": 0.08056640625,
638
+ "learning_rate": 4.087980306641365e-05,
639
+ "loss": 0.11388391256332397,
640
+ "mean_token_accuracy": 0.9668685078620911,
641
+ "num_tokens": 1841597.0,
642
+ "step": 600
643
+ },
644
+ {
645
+ "entropy": 0.11204307321459055,
646
+ "epoch": 2.1709706144256455,
647
+ "grad_norm": 0.1015625,
648
+ "learning_rate": 3.7822962185039914e-05,
649
+ "loss": 0.11144263744354248,
650
+ "mean_token_accuracy": 0.9687378108501434,
651
+ "num_tokens": 1872199.0,
652
+ "step": 610
653
+ },
654
+ {
655
+ "entropy": 0.12015030086040497,
656
+ "epoch": 2.2065894924309886,
657
+ "grad_norm": 0.0712890625,
658
+ "learning_rate": 3.4858045987886145e-05,
659
+ "loss": 0.11484370231628419,
660
+ "mean_token_accuracy": 0.967641019821167,
661
+ "num_tokens": 1903172.0,
662
+ "step": 620
663
+ },
664
+ {
665
+ "entropy": 0.12499598301947117,
666
+ "epoch": 2.242208370436331,
667
+ "grad_norm": 0.06640625,
668
+ "learning_rate": 3.198943790965332e-05,
669
+ "loss": 0.12357065677642823,
670
+ "mean_token_accuracy": 0.9670094296336174,
671
+ "num_tokens": 1933713.0,
672
+ "step": 630
673
+ },
674
+ {
675
+ "entropy": 0.1296376422047615,
676
+ "epoch": 2.277827248441674,
677
+ "grad_norm": 0.0673828125,
678
+ "learning_rate": 2.9221378999786853e-05,
679
+ "loss": 0.12340972423553467,
680
+ "mean_token_accuracy": 0.96511862128973,
681
+ "num_tokens": 1964556.0,
682
+ "step": 640
683
+ },
684
+ {
685
+ "entropy": 0.12314727567136288,
686
+ "epoch": 2.3134461264470167,
687
+ "grad_norm": 0.080078125,
688
+ "learning_rate": 2.655796165236234e-05,
689
+ "loss": 0.11837244033813477,
690
+ "mean_token_accuracy": 0.9671967878937722,
691
+ "num_tokens": 1995252.0,
692
+ "step": 650
693
+ },
694
+ {
695
+ "entropy": 0.11561896577477455,
696
+ "epoch": 2.3490650044523598,
697
+ "grad_norm": 0.1064453125,
698
+ "learning_rate": 2.400312355574853e-05,
699
+ "loss": 0.11029613018035889,
700
+ "mean_token_accuracy": 0.967219403386116,
701
+ "num_tokens": 2025889.0,
702
+ "step": 660
703
+ },
704
+ {
705
+ "entropy": 0.11878675390034914,
706
+ "epoch": 2.3846838824577024,
707
+ "grad_norm": 0.080078125,
708
+ "learning_rate": 2.1560641870992616e-05,
709
+ "loss": 0.11299164295196533,
710
+ "mean_token_accuracy": 0.9675500631332398,
711
+ "num_tokens": 2056588.0,
712
+ "step": 670
713
+ },
714
+ {
715
+ "entropy": 0.11508079580962657,
716
+ "epoch": 2.4203027604630454,
717
+ "grad_norm": 0.07080078125,
718
+ "learning_rate": 1.9234127647534604e-05,
719
+ "loss": 0.10480635166168213,
720
+ "mean_token_accuracy": 0.9691305905580521,
721
+ "num_tokens": 2087259.0,
722
+ "step": 680
723
+ },
724
+ {
725
+ "entropy": 0.11411711536347866,
726
+ "epoch": 2.4559216384683884,
727
+ "grad_norm": 0.06640625,
728
+ "learning_rate": 1.7027020484506996e-05,
729
+ "loss": 0.11348260641098022,
730
+ "mean_token_accuracy": 0.9682735517621041,
731
+ "num_tokens": 2117728.0,
732
+ "step": 690
733
+ },
734
+ {
735
+ "entropy": 0.11259160749614239,
736
+ "epoch": 2.491540516473731,
737
+ "grad_norm": 0.05224609375,
738
+ "learning_rate": 1.4942583445512103e-05,
739
+ "loss": 0.10898959636688232,
740
+ "mean_token_accuracy": 0.9685835257172585,
741
+ "num_tokens": 2148333.0,
742
+ "step": 700
743
+ },
744
+ {
745
+ "entropy": 0.12499043270945549,
746
+ "epoch": 2.527159394479074,
747
+ "grad_norm": 0.0849609375,
748
+ "learning_rate": 1.2983898234396308e-05,
749
+ "loss": 0.12133818864822388,
750
+ "mean_token_accuracy": 0.9664898782968521,
751
+ "num_tokens": 2178932.0,
752
+ "step": 710
753
+ },
754
+ {
755
+ "entropy": 0.11083998121321201,
756
+ "epoch": 2.5627782724844166,
757
+ "grad_norm": 0.048828125,
758
+ "learning_rate": 1.1153860639152169e-05,
759
+ "loss": 0.10544953346252442,
760
+ "mean_token_accuracy": 0.9696028590202331,
761
+ "num_tokens": 2209610.0,
762
+ "step": 720
763
+ },
764
+ {
765
+ "entropy": 0.11835917010903359,
766
+ "epoch": 2.5983971504897596,
767
+ "grad_norm": 0.08203125,
768
+ "learning_rate": 9.455176250685338e-06,
769
+ "loss": 0.11609755754470825,
770
+ "mean_token_accuracy": 0.9664707094430923,
771
+ "num_tokens": 2240346.0,
772
+ "step": 730
773
+ },
774
+ {
775
+ "entropy": 0.12215895913541316,
776
+ "epoch": 2.634016028495102,
777
+ "grad_norm": 0.0673828125,
778
+ "learning_rate": 7.890356462775373e-06,
779
+ "loss": 0.11725597381591797,
780
+ "mean_token_accuracy": 0.9670157313346863,
781
+ "num_tokens": 2271054.0,
782
+ "step": 740
783
+ },
784
+ {
785
+ "entropy": 0.12366384714841842,
786
+ "epoch": 2.669634906500445,
787
+ "grad_norm": 0.07861328125,
788
+ "learning_rate": 6.461714759144233e-06,
789
+ "loss": 0.1168657660484314,
790
+ "mean_token_accuracy": 0.9668171271681786,
791
+ "num_tokens": 2301971.0,
792
+ "step": 750
793
+ },
794
+ {
795
+ "entropy": 0.10823128782212735,
796
+ "epoch": 2.7052537845057882,
797
+ "grad_norm": 0.05712890625,
798
+ "learning_rate": 5.171363293121901e-06,
799
+ "loss": 0.10103480815887451,
800
+ "mean_token_accuracy": 0.9696328729391098,
801
+ "num_tokens": 2332515.0,
802
+ "step": 760
803
+ },
804
+ {
805
+ "entropy": 0.10126303266733885,
806
+ "epoch": 2.740872662511131,
807
+ "grad_norm": 0.07080078125,
808
+ "learning_rate": 4.021209764965828e-06,
809
+ "loss": 0.09550071954727173,
810
+ "mean_token_accuracy": 0.9718543246388436,
811
+ "num_tokens": 2363176.0,
812
+ "step": 770
813
+ },
814
+ {
815
+ "entropy": 0.10594012532383204,
816
+ "epoch": 2.776491540516474,
817
+ "grad_norm": 0.0625,
818
+ "learning_rate": 3.0129546014508567e-06,
819
+ "loss": 0.1027148962020874,
820
+ "mean_token_accuracy": 0.9698196157813073,
821
+ "num_tokens": 2393823.0,
822
+ "step": 780
823
+ },
824
+ {
825
+ "entropy": 0.11277557052671909,
826
+ "epoch": 2.8121104185218164,
827
+ "grad_norm": 0.057373046875,
828
+ "learning_rate": 2.1480884418993983e-06,
829
+ "loss": 0.10367020368576049,
830
+ "mean_token_accuracy": 0.9688401147723198,
831
+ "num_tokens": 2424575.0,
832
+ "step": 790
833
+ },
834
+ {
835
+ "entropy": 0.13837510608136655,
836
+ "epoch": 2.8477292965271594,
837
+ "grad_norm": 0.0791015625,
838
+ "learning_rate": 1.4278899343687425e-06,
839
+ "loss": 0.13071552515029908,
840
+ "mean_token_accuracy": 0.9625913232564927,
841
+ "num_tokens": 2455901.0,
842
+ "step": 800
843
+ },
844
+ {
845
+ "entropy": 0.12289627455174923,
846
+ "epoch": 2.883348174532502,
847
+ "grad_norm": 0.09423828125,
848
+ "learning_rate": 8.534238452534759e-07,
849
+ "loss": 0.11349356174468994,
850
+ "mean_token_accuracy": 0.9673542976379395,
851
+ "num_tokens": 2486966.0,
852
+ "step": 810
853
+ },
854
+ {
855
+ "entropy": 0.11963035948574544,
856
+ "epoch": 2.918967052537845,
857
+ "grad_norm": 0.05419921875,
858
+ "learning_rate": 4.2553948509802545e-07,
859
+ "loss": 0.11257882118225097,
860
+ "mean_token_accuracy": 0.9673726871609688,
861
+ "num_tokens": 2517590.0,
862
+ "step": 820
863
+ },
864
+ {
865
+ "entropy": 0.13870051857084037,
866
+ "epoch": 2.954585930543188,
867
+ "grad_norm": 0.06298828125,
868
+ "learning_rate": 1.4486945294652776e-07,
869
+ "loss": 0.13519666194915772,
870
+ "mean_token_accuracy": 0.9639795452356339,
871
+ "num_tokens": 2548330.0,
872
+ "step": 830
873
+ },
874
+ {
875
+ "entropy": 0.10831217430531978,
876
+ "epoch": 2.9902048085485307,
877
+ "grad_norm": 0.06494140625,
878
+ "learning_rate": 1.1828701086558092e-08,
879
+ "loss": 0.10172584056854247,
880
+ "mean_token_accuracy": 0.9707534283399581,
881
+ "num_tokens": 2579075.0,
882
+ "step": 840
883
+ },
884
+ {
885
+ "epoch": 3.0,
886
+ "eval_entropy": 0.12990308695651115,
887
+ "eval_loss": 0.15479549765586853,
888
+ "eval_mean_token_accuracy": 0.9612296595641062,
889
+ "eval_num_tokens": 2587536.0,
890
+ "eval_runtime": 35.3111,
891
+ "eval_samples_per_second": 15.887,
892
+ "eval_steps_per_second": 3.993,
893
+ "step": 843
894
+ }
895
+ ],
896
+ "logging_steps": 10,
897
+ "max_steps": 843,
898
+ "num_input_tokens_seen": 0,
899
+ "num_train_epochs": 3,
900
+ "save_steps": 500,
901
+ "stateful_callbacks": {
902
+ "TrainerControl": {
903
+ "args": {
904
+ "should_epoch_stop": false,
905
+ "should_evaluate": false,
906
+ "should_log": false,
907
+ "should_save": true,
908
+ "should_training_stop": true
909
+ },
910
+ "attributes": {}
911
+ }
912
+ },
913
+ "total_flos": 6.085644413841408e+16,
914
+ "train_batch_size": 4,
915
+ "trial_name": null,
916
+ "trial_params": null
917
+ }
checkpoint-843/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57851b427cb3a4649963b62f1b5d8860aa50ba7ac4abbbb2ebc56acbf65bd694
3
+ size 5649
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be75606093db2094d7cd20f3c2f385c212750648bd6ea4fb2bf507a6a4c55506
3
+ size 11422650
tokenizer_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "backend": "tokenizers",
4
+ "bos_token": null,
5
+ "clean_up_tokenization_spaces": false,
6
+ "eos_token": "<|im_end|>",
7
+ "errors": "replace",
8
+ "extra_special_tokens": [
9
+ "<|im_start|>",
10
+ "<|im_end|>",
11
+ "<|object_ref_start|>",
12
+ "<|object_ref_end|>",
13
+ "<|box_start|>",
14
+ "<|box_end|>",
15
+ "<|quad_start|>",
16
+ "<|quad_end|>",
17
+ "<|vision_start|>",
18
+ "<|vision_end|>",
19
+ "<|vision_pad|>",
20
+ "<|image_pad|>",
21
+ "<|video_pad|>"
22
+ ],
23
+ "is_local": false,
24
+ "model_max_length": 1010000,
25
+ "pad_token": "<|im_end|>",
26
+ "split_special_tokens": false,
27
+ "tokenizer_class": "Qwen2Tokenizer",
28
+ "unk_token": null
29
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57851b427cb3a4649963b62f1b5d8860aa50ba7ac4abbbb2ebc56acbf65bd694
3
+ size 5649