clairedhx commited on
Commit
c269315
·
verified ·
1 Parent(s): e7a379b

Upload folder using huggingface_hub

Browse files
Files changed (44) hide show
  1. README.md +62 -0
  2. adapter_config.json +39 -0
  3. adapter_model.safetensors +3 -0
  4. chat_template.jinja +87 -0
  5. checkpoint-1383/README.md +209 -0
  6. checkpoint-1383/adapter_config.json +39 -0
  7. checkpoint-1383/adapter_model.safetensors +3 -0
  8. checkpoint-1383/chat_template.jinja +87 -0
  9. checkpoint-1383/optimizer.pt +3 -0
  10. checkpoint-1383/rng_state.pth +3 -0
  11. checkpoint-1383/scheduler.pt +3 -0
  12. checkpoint-1383/special_tokens_map.json +24 -0
  13. checkpoint-1383/tokenizer.model +3 -0
  14. checkpoint-1383/tokenizer_config.json +0 -0
  15. checkpoint-1383/trainer_state.json +1276 -0
  16. checkpoint-1383/training_args.bin +3 -0
  17. checkpoint-461/README.md +209 -0
  18. checkpoint-461/adapter_config.json +39 -0
  19. checkpoint-461/adapter_model.safetensors +3 -0
  20. checkpoint-461/chat_template.jinja +87 -0
  21. checkpoint-461/optimizer.pt +3 -0
  22. checkpoint-461/rng_state.pth +3 -0
  23. checkpoint-461/scheduler.pt +3 -0
  24. checkpoint-461/special_tokens_map.json +24 -0
  25. checkpoint-461/tokenizer.model +3 -0
  26. checkpoint-461/tokenizer_config.json +0 -0
  27. checkpoint-461/trainer_state.json +448 -0
  28. checkpoint-461/training_args.bin +3 -0
  29. checkpoint-922/README.md +209 -0
  30. checkpoint-922/adapter_config.json +39 -0
  31. checkpoint-922/adapter_model.safetensors +3 -0
  32. checkpoint-922/chat_template.jinja +87 -0
  33. checkpoint-922/optimizer.pt +3 -0
  34. checkpoint-922/rng_state.pth +3 -0
  35. checkpoint-922/scheduler.pt +3 -0
  36. checkpoint-922/special_tokens_map.json +24 -0
  37. checkpoint-922/tokenizer.model +3 -0
  38. checkpoint-922/tokenizer_config.json +0 -0
  39. checkpoint-922/trainer_state.json +862 -0
  40. checkpoint-922/training_args.bin +3 -0
  41. special_tokens_map.json +24 -0
  42. tokenizer.model +3 -0
  43. tokenizer_config.json +0 -0
  44. training_args.bin +3 -0
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Mistral-7B-Instruct-v0.3
3
+ library_name: peft
4
+ model_name: mistral7b_labels2codes_lora
5
+ tags:
6
+ - base_model:adapter:mistralai/Mistral-7B-Instruct-v0.3
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ licence: license
12
+ pipeline_tag: text-generation
13
+ ---
14
+
15
+ # Model Card for mistral7b_labels2codes_lora
16
+
17
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3).
18
+ It has been trained using [TRL](https://github.com/huggingface/trl).
19
+
20
+ ## Quick start
21
+
22
+ ```python
23
+ from transformers import pipeline
24
+
25
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
26
+ generator = pipeline("text-generation", model="None", device="cuda")
27
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
28
+ print(output["generated_text"])
29
+ ```
30
+
31
+ ## Training procedure
32
+
33
+
34
+
35
+
36
+ This model was trained with SFT.
37
+
38
+ ### Framework versions
39
+
40
+ - PEFT 0.17.0
41
+ - TRL: 0.21.0
42
+ - Transformers: 4.55.2
43
+ - Pytorch: 2.7.1
44
+ - Datasets: 3.6.0
45
+ - Tokenizers: 0.21.2
46
+
47
+ ## Citations
48
+
49
+
50
+
51
+ Cite TRL as:
52
+
53
+ ```bibtex
54
+ @misc{vonwerra2022trl,
55
+ title = {{TRL: Transformer Reinforcement Learning}},
56
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
57
+ year = 2020,
58
+ journal = {GitHub repository},
59
+ publisher = {GitHub},
60
+ howpublished = {\url{https://github.com/huggingface/trl}}
61
+ }
62
+ ```
adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj"
32
+ ],
33
+ "target_parameters": null,
34
+ "task_type": "CAUSAL_LM",
35
+ "trainable_token_indices": null,
36
+ "use_dora": false,
37
+ "use_qalora": false,
38
+ "use_rslora": false
39
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1790b7928f19c11eea9b3ae25f2fe1a066a98224c55c01b0bee9e895ec7a9e48
3
+ size 54560368
chat_template.jinja ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if messages[0]["role"] == "system" %}
2
+ {%- set system_message = messages[0]["content"] %}
3
+ {%- set loop_messages = messages[1:] %}
4
+ {%- else %}
5
+ {%- set loop_messages = messages %}
6
+ {%- endif %}
7
+ {%- if not tools is defined %}
8
+ {%- set tools = none %}
9
+ {%- endif %}
10
+ {%- set user_messages = loop_messages | selectattr("role", "equalto", "user") | list %}
11
+
12
+ {#- This block checks for alternating user/assistant messages, skipping tool calling messages #}
13
+ {%- set ns = namespace() %}
14
+ {%- set ns.index = 0 %}
15
+ {%- for message in loop_messages %}
16
+ {%- if not (message.role == "tool" or message.role == "tool_results" or (message.tool_calls is defined and message.tool_calls is not none)) %}
17
+ {%- if (message["role"] == "user") != (ns.index % 2 == 0) %}
18
+ {{- raise_exception("After the optional system message, conversation roles must alternate user/assistant/user/assistant/...") }}
19
+ {%- endif %}
20
+ {%- set ns.index = ns.index + 1 %}
21
+ {%- endif %}
22
+ {%- endfor %}
23
+
24
+ {{- bos_token }}
25
+ {%- for message in loop_messages %}
26
+ {%- if message["role"] == "user" %}
27
+ {%- if tools is not none and (message == user_messages[-1]) %}
28
+ {{- "[AVAILABLE_TOOLS] [" }}
29
+ {%- for tool in tools %}
30
+ {%- set tool = tool.function %}
31
+ {{- '{"type": "function", "function": {' }}
32
+ {%- for key, val in tool.items() if key != "return" %}
33
+ {%- if val is string %}
34
+ {{- '"' + key + '": "' + val + '"' }}
35
+ {%- else %}
36
+ {{- '"' + key + '": ' + val|tojson }}
37
+ {%- endif %}
38
+ {%- if not loop.last %}
39
+ {{- ", " }}
40
+ {%- endif %}
41
+ {%- endfor %}
42
+ {{- "}}" }}
43
+ {%- if not loop.last %}
44
+ {{- ", " }}
45
+ {%- else %}
46
+ {{- "]" }}
47
+ {%- endif %}
48
+ {%- endfor %}
49
+ {{- "[/AVAILABLE_TOOLS]" }}
50
+ {%- endif %}
51
+ {%- if loop.last and system_message is defined %}
52
+ {{- "[INST] " + system_message + "\n\n" + message["content"] + "[/INST]" }}
53
+ {%- else %}
54
+ {{- "[INST] " + message["content"] + "[/INST]" }}
55
+ {%- endif %}
56
+ {%- elif message.tool_calls is defined and message.tool_calls is not none %}
57
+ {{- "[TOOL_CALLS] [" }}
58
+ {%- for tool_call in message.tool_calls %}
59
+ {%- set out = tool_call.function|tojson %}
60
+ {{- out[:-1] }}
61
+ {%- if not tool_call.id is defined or tool_call.id|length != 9 %}
62
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
63
+ {%- endif %}
64
+ {{- ', "id": "' + tool_call.id + '"}' }}
65
+ {%- if not loop.last %}
66
+ {{- ", " }}
67
+ {%- else %}
68
+ {{- "]" + eos_token }}
69
+ {%- endif %}
70
+ {%- endfor %}
71
+ {%- elif message["role"] == "assistant" %}
72
+ {{- " " + message["content"]|trim + eos_token}}
73
+ {%- elif message["role"] == "tool_results" or message["role"] == "tool" %}
74
+ {%- if message.content is defined and message.content.content is defined %}
75
+ {%- set content = message.content.content %}
76
+ {%- else %}
77
+ {%- set content = message.content %}
78
+ {%- endif %}
79
+ {{- '[TOOL_RESULTS] {"content": ' + content|string + ", " }}
80
+ {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}
81
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
82
+ {%- endif %}
83
+ {{- '"call_id": "' + message.tool_call_id + '"}[/TOOL_RESULTS]' }}
84
+ {%- else %}
85
+ {{- raise_exception("Only user and assistant roles are supported, with the exception of an initial optional system message!") }}
86
+ {%- endif %}
87
+ {%- endfor %}
checkpoint-1383/README.md ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Mistral-7B-Instruct-v0.3
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:mistralai/Mistral-7B-Instruct-v0.3
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ ---
12
+
13
+ # Model Card for Model ID
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ <!-- Provide a longer summary of what this model is. -->
24
+
25
+
26
+
27
+ - **Developed by:** [More Information Needed]
28
+ - **Funded by [optional]:** [More Information Needed]
29
+ - **Shared by [optional]:** [More Information Needed]
30
+ - **Model type:** [More Information Needed]
31
+ - **Language(s) (NLP):** [More Information Needed]
32
+ - **License:** [More Information Needed]
33
+ - **Finetuned from model [optional]:** [More Information Needed]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ <!-- Provide the basic links for the model. -->
38
+
39
+ - **Repository:** [More Information Needed]
40
+ - **Paper [optional]:** [More Information Needed]
41
+ - **Demo [optional]:** [More Information Needed]
42
+
43
+ ## Uses
44
+
45
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
+
47
+ ### Direct Use
48
+
49
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Downstream Use [optional]
54
+
55
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+
57
+ [More Information Needed]
58
+
59
+ ### Out-of-Scope Use
60
+
61
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
74
+
75
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
76
+
77
+ ## How to Get Started with the Model
78
+
79
+ Use the code below to get started with the model.
80
+
81
+ [More Information Needed]
82
+
83
+ ## Training Details
84
+
85
+ ### Training Data
86
+
87
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
+
89
+ [More Information Needed]
90
+
91
+ ### Training Procedure
92
+
93
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
+
95
+ #### Preprocessing [optional]
96
+
97
+ [More Information Needed]
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
103
+
104
+ #### Speeds, Sizes, Times [optional]
105
+
106
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ## Evaluation
111
+
112
+ <!-- This section describes the evaluation protocols and provides the results. -->
113
+
114
+ ### Testing Data, Factors & Metrics
115
+
116
+ #### Testing Data
117
+
118
+ <!-- This should link to a Dataset Card if possible. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Factors
123
+
124
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
125
+
126
+ [More Information Needed]
127
+
128
+ #### Metrics
129
+
130
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
131
+
132
+ [More Information Needed]
133
+
134
+ ### Results
135
+
136
+ [More Information Needed]
137
+
138
+ #### Summary
139
+
140
+
141
+
142
+ ## Model Examination [optional]
143
+
144
+ <!-- Relevant interpretability work for the model goes here -->
145
+
146
+ [More Information Needed]
147
+
148
+ ## Environmental Impact
149
+
150
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
151
+
152
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
153
+
154
+ - **Hardware Type:** [More Information Needed]
155
+ - **Hours used:** [More Information Needed]
156
+ - **Cloud Provider:** [More Information Needed]
157
+ - **Compute Region:** [More Information Needed]
158
+ - **Carbon Emitted:** [More Information Needed]
159
+
160
+ ## Technical Specifications [optional]
161
+
162
+ ### Model Architecture and Objective
163
+
164
+ [More Information Needed]
165
+
166
+ ### Compute Infrastructure
167
+
168
+ [More Information Needed]
169
+
170
+ #### Hardware
171
+
172
+ [More Information Needed]
173
+
174
+ #### Software
175
+
176
+ [More Information Needed]
177
+
178
+ ## Citation [optional]
179
+
180
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
181
+
182
+ **BibTeX:**
183
+
184
+ [More Information Needed]
185
+
186
+ **APA:**
187
+
188
+ [More Information Needed]
189
+
190
+ ## Glossary [optional]
191
+
192
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
193
+
194
+ [More Information Needed]
195
+
196
+ ## More Information [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Authors [optional]
201
+
202
+ [More Information Needed]
203
+
204
+ ## Model Card Contact
205
+
206
+ [More Information Needed]
207
+ ### Framework versions
208
+
209
+ - PEFT 0.17.0
checkpoint-1383/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj"
32
+ ],
33
+ "target_parameters": null,
34
+ "task_type": "CAUSAL_LM",
35
+ "trainable_token_indices": null,
36
+ "use_dora": false,
37
+ "use_qalora": false,
38
+ "use_rslora": false
39
+ }
checkpoint-1383/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1790b7928f19c11eea9b3ae25f2fe1a066a98224c55c01b0bee9e895ec7a9e48
3
+ size 54560368
checkpoint-1383/chat_template.jinja ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if messages[0]["role"] == "system" %}
2
+ {%- set system_message = messages[0]["content"] %}
3
+ {%- set loop_messages = messages[1:] %}
4
+ {%- else %}
5
+ {%- set loop_messages = messages %}
6
+ {%- endif %}
7
+ {%- if not tools is defined %}
8
+ {%- set tools = none %}
9
+ {%- endif %}
10
+ {%- set user_messages = loop_messages | selectattr("role", "equalto", "user") | list %}
11
+
12
+ {#- This block checks for alternating user/assistant messages, skipping tool calling messages #}
13
+ {%- set ns = namespace() %}
14
+ {%- set ns.index = 0 %}
15
+ {%- for message in loop_messages %}
16
+ {%- if not (message.role == "tool" or message.role == "tool_results" or (message.tool_calls is defined and message.tool_calls is not none)) %}
17
+ {%- if (message["role"] == "user") != (ns.index % 2 == 0) %}
18
+ {{- raise_exception("After the optional system message, conversation roles must alternate user/assistant/user/assistant/...") }}
19
+ {%- endif %}
20
+ {%- set ns.index = ns.index + 1 %}
21
+ {%- endif %}
22
+ {%- endfor %}
23
+
24
+ {{- bos_token }}
25
+ {%- for message in loop_messages %}
26
+ {%- if message["role"] == "user" %}
27
+ {%- if tools is not none and (message == user_messages[-1]) %}
28
+ {{- "[AVAILABLE_TOOLS] [" }}
29
+ {%- for tool in tools %}
30
+ {%- set tool = tool.function %}
31
+ {{- '{"type": "function", "function": {' }}
32
+ {%- for key, val in tool.items() if key != "return" %}
33
+ {%- if val is string %}
34
+ {{- '"' + key + '": "' + val + '"' }}
35
+ {%- else %}
36
+ {{- '"' + key + '": ' + val|tojson }}
37
+ {%- endif %}
38
+ {%- if not loop.last %}
39
+ {{- ", " }}
40
+ {%- endif %}
41
+ {%- endfor %}
42
+ {{- "}}" }}
43
+ {%- if not loop.last %}
44
+ {{- ", " }}
45
+ {%- else %}
46
+ {{- "]" }}
47
+ {%- endif %}
48
+ {%- endfor %}
49
+ {{- "[/AVAILABLE_TOOLS]" }}
50
+ {%- endif %}
51
+ {%- if loop.last and system_message is defined %}
52
+ {{- "[INST] " + system_message + "\n\n" + message["content"] + "[/INST]" }}
53
+ {%- else %}
54
+ {{- "[INST] " + message["content"] + "[/INST]" }}
55
+ {%- endif %}
56
+ {%- elif message.tool_calls is defined and message.tool_calls is not none %}
57
+ {{- "[TOOL_CALLS] [" }}
58
+ {%- for tool_call in message.tool_calls %}
59
+ {%- set out = tool_call.function|tojson %}
60
+ {{- out[:-1] }}
61
+ {%- if not tool_call.id is defined or tool_call.id|length != 9 %}
62
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
63
+ {%- endif %}
64
+ {{- ', "id": "' + tool_call.id + '"}' }}
65
+ {%- if not loop.last %}
66
+ {{- ", " }}
67
+ {%- else %}
68
+ {{- "]" + eos_token }}
69
+ {%- endif %}
70
+ {%- endfor %}
71
+ {%- elif message["role"] == "assistant" %}
72
+ {{- " " + message["content"]|trim + eos_token}}
73
+ {%- elif message["role"] == "tool_results" or message["role"] == "tool" %}
74
+ {%- if message.content is defined and message.content.content is defined %}
75
+ {%- set content = message.content.content %}
76
+ {%- else %}
77
+ {%- set content = message.content %}
78
+ {%- endif %}
79
+ {{- '[TOOL_RESULTS] {"content": ' + content|string + ", " }}
80
+ {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}
81
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
82
+ {%- endif %}
83
+ {{- '"call_id": "' + message.tool_call_id + '"}[/TOOL_RESULTS]' }}
84
+ {%- else %}
85
+ {{- raise_exception("Only user and assistant roles are supported, with the exception of an initial optional system message!") }}
86
+ {%- endif %}
87
+ {%- endfor %}
checkpoint-1383/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bf8d723239a9f8765ea2c9af5351cd7324f0b96a218f4a6ebba596375afa9aa
3
+ size 1593
checkpoint-1383/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bc6cb51ab7ae63fbe677bc72e378a12642b92ac6970fa9ba77833061d06c09b
3
+ size 14645
checkpoint-1383/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9915a1d438f134c1517f1be983ddad5176ebd284be94defdc54893cbdfb31dd6
3
+ size 1465
checkpoint-1383/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
checkpoint-1383/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
checkpoint-1383/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1383/trainer_state.json ADDED
@@ -0,0 +1,1276 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1383,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.021719948415122516,
14
+ "grad_norm": 0.0,
15
+ "learning_rate": 0.00019869848156182213,
16
+ "loss": 6.0531,
17
+ "mean_token_accuracy": 0.37881034857127815,
18
+ "num_tokens": 78573.0,
19
+ "step": 10
20
+ },
21
+ {
22
+ "epoch": 0.04343989683024503,
23
+ "grad_norm": 0.0,
24
+ "learning_rate": 0.00019725234996384672,
25
+ "loss": 5.9903,
26
+ "mean_token_accuracy": 0.3887314551044255,
27
+ "num_tokens": 157496.0,
28
+ "step": 20
29
+ },
30
+ {
31
+ "epoch": 0.06515984524536754,
32
+ "grad_norm": 0.0,
33
+ "learning_rate": 0.00019580621836587129,
34
+ "loss": 5.9929,
35
+ "mean_token_accuracy": 0.3850487937917933,
36
+ "num_tokens": 236512.0,
37
+ "step": 30
38
+ },
39
+ {
40
+ "epoch": 0.08687979366049006,
41
+ "grad_norm": 0.0,
42
+ "learning_rate": 0.00019436008676789588,
43
+ "loss": 5.986,
44
+ "mean_token_accuracy": 0.3865805763518438,
45
+ "num_tokens": 315608.0,
46
+ "step": 40
47
+ },
48
+ {
49
+ "epoch": 0.10859974207561257,
50
+ "grad_norm": 0.0,
51
+ "learning_rate": 0.00019291395516992047,
52
+ "loss": 6.0318,
53
+ "mean_token_accuracy": 0.38490854618139564,
54
+ "num_tokens": 394791.0,
55
+ "step": 50
56
+ },
57
+ {
58
+ "epoch": 0.13031969049073508,
59
+ "grad_norm": 0.0,
60
+ "learning_rate": 0.00019146782357194506,
61
+ "loss": 5.9875,
62
+ "mean_token_accuracy": 0.38546840590424836,
63
+ "num_tokens": 473456.0,
64
+ "step": 60
65
+ },
66
+ {
67
+ "epoch": 0.1520396389058576,
68
+ "grad_norm": 0.0,
69
+ "learning_rate": 0.00019002169197396965,
70
+ "loss": 5.9773,
71
+ "mean_token_accuracy": 0.3837947838823311,
72
+ "num_tokens": 552247.0,
73
+ "step": 70
74
+ },
75
+ {
76
+ "epoch": 0.17375958732098012,
77
+ "grad_norm": 0.0,
78
+ "learning_rate": 0.0001885755603759942,
79
+ "loss": 6.0197,
80
+ "mean_token_accuracy": 0.38375892234034836,
81
+ "num_tokens": 631096.0,
82
+ "step": 80
83
+ },
84
+ {
85
+ "epoch": 0.19547953573610263,
86
+ "grad_norm": 0.0,
87
+ "learning_rate": 0.0001871294287780188,
88
+ "loss": 6.0014,
89
+ "mean_token_accuracy": 0.38519658900331705,
90
+ "num_tokens": 710275.0,
91
+ "step": 90
92
+ },
93
+ {
94
+ "epoch": 0.21719948415122514,
95
+ "grad_norm": 0.0,
96
+ "learning_rate": 0.0001856832971800434,
97
+ "loss": 6.0217,
98
+ "mean_token_accuracy": 0.3826989881345071,
99
+ "num_tokens": 789174.0,
100
+ "step": 100
101
+ },
102
+ {
103
+ "epoch": 0.23891943256634765,
104
+ "grad_norm": 0.0,
105
+ "learning_rate": 0.00018423716558206799,
106
+ "loss": 6.0294,
107
+ "mean_token_accuracy": 0.3842650496866554,
108
+ "num_tokens": 867922.0,
109
+ "step": 110
110
+ },
111
+ {
112
+ "epoch": 0.26063938098147016,
113
+ "grad_norm": 0.0,
114
+ "learning_rate": 0.00018279103398409255,
115
+ "loss": 5.9548,
116
+ "mean_token_accuracy": 0.3875111517496407,
117
+ "num_tokens": 947073.0,
118
+ "step": 120
119
+ },
120
+ {
121
+ "epoch": 0.2823593293965927,
122
+ "grad_norm": 0.0,
123
+ "learning_rate": 0.00018134490238611714,
124
+ "loss": 6.0071,
125
+ "mean_token_accuracy": 0.38376112943515184,
126
+ "num_tokens": 1026211.0,
127
+ "step": 130
128
+ },
129
+ {
130
+ "epoch": 0.3040792778117152,
131
+ "grad_norm": 0.0,
132
+ "learning_rate": 0.00017989877078814173,
133
+ "loss": 6.0105,
134
+ "mean_token_accuracy": 0.38463024909142407,
135
+ "num_tokens": 1105147.0,
136
+ "step": 140
137
+ },
138
+ {
139
+ "epoch": 0.3257992262268377,
140
+ "grad_norm": 0.0,
141
+ "learning_rate": 0.0001784526391901663,
142
+ "loss": 6.0449,
143
+ "mean_token_accuracy": 0.38241248747799544,
144
+ "num_tokens": 1183837.0,
145
+ "step": 150
146
+ },
147
+ {
148
+ "epoch": 0.34751917464196025,
149
+ "grad_norm": 0.0,
150
+ "learning_rate": 0.00017700650759219091,
151
+ "loss": 5.9946,
152
+ "mean_token_accuracy": 0.3821917780907825,
153
+ "num_tokens": 1262676.0,
154
+ "step": 160
155
+ },
156
+ {
157
+ "epoch": 0.36923912305708273,
158
+ "grad_norm": 0.0,
159
+ "learning_rate": 0.00017556037599421548,
160
+ "loss": 6.0049,
161
+ "mean_token_accuracy": 0.3842250820598565,
162
+ "num_tokens": 1341759.0,
163
+ "step": 170
164
+ },
165
+ {
166
+ "epoch": 0.39095907147220527,
167
+ "grad_norm": 0.0,
168
+ "learning_rate": 0.00017411424439624007,
169
+ "loss": 5.9736,
170
+ "mean_token_accuracy": 0.3861917880363762,
171
+ "num_tokens": 1420987.0,
172
+ "step": 180
173
+ },
174
+ {
175
+ "epoch": 0.41267901988732775,
176
+ "grad_norm": 0.0,
177
+ "learning_rate": 0.00017266811279826466,
178
+ "loss": 6.0431,
179
+ "mean_token_accuracy": 0.38146835477091373,
180
+ "num_tokens": 1499903.0,
181
+ "step": 190
182
+ },
183
+ {
184
+ "epoch": 0.4343989683024503,
185
+ "grad_norm": 0.0,
186
+ "learning_rate": 0.00017122198120028922,
187
+ "loss": 6.0069,
188
+ "mean_token_accuracy": 0.3861038032686338,
189
+ "num_tokens": 1578880.0,
190
+ "step": 200
191
+ },
192
+ {
193
+ "epoch": 0.4561189167175728,
194
+ "grad_norm": 0.0,
195
+ "learning_rate": 0.00016977584960231381,
196
+ "loss": 6.0091,
197
+ "mean_token_accuracy": 0.3845075036631897,
198
+ "num_tokens": 1657709.0,
199
+ "step": 210
200
+ },
201
+ {
202
+ "epoch": 0.4778388651326953,
203
+ "grad_norm": 0.0,
204
+ "learning_rate": 0.0001683297180043384,
205
+ "loss": 6.0029,
206
+ "mean_token_accuracy": 0.38288453239947556,
207
+ "num_tokens": 1736620.0,
208
+ "step": 220
209
+ },
210
+ {
211
+ "epoch": 0.49955881354781784,
212
+ "grad_norm": 0.0,
213
+ "learning_rate": 0.000166883586406363,
214
+ "loss": 5.9933,
215
+ "mean_token_accuracy": 0.38531269936356694,
216
+ "num_tokens": 1815478.0,
217
+ "step": 230
218
+ },
219
+ {
220
+ "epoch": 0.5212787619629403,
221
+ "grad_norm": 0.0,
222
+ "learning_rate": 0.00016543745480838756,
223
+ "loss": 6.0334,
224
+ "mean_token_accuracy": 0.3833939728909172,
225
+ "num_tokens": 1894256.0,
226
+ "step": 240
227
+ },
228
+ {
229
+ "epoch": 0.5429987103780628,
230
+ "grad_norm": 0.0,
231
+ "learning_rate": 0.00016399132321041215,
232
+ "loss": 6.0484,
233
+ "mean_token_accuracy": 0.3841994108865038,
234
+ "num_tokens": 1972997.0,
235
+ "step": 250
236
+ },
237
+ {
238
+ "epoch": 0.5647186587931854,
239
+ "grad_norm": 0.0,
240
+ "learning_rate": 0.00016254519161243674,
241
+ "loss": 6.0285,
242
+ "mean_token_accuracy": 0.38443261398933826,
243
+ "num_tokens": 2051890.0,
244
+ "step": 260
245
+ },
246
+ {
247
+ "epoch": 0.5864386072083079,
248
+ "grad_norm": 0.0,
249
+ "learning_rate": 0.00016109906001446133,
250
+ "loss": 5.983,
251
+ "mean_token_accuracy": 0.385862308065407,
252
+ "num_tokens": 2130746.0,
253
+ "step": 270
254
+ },
255
+ {
256
+ "epoch": 0.6081585556234304,
257
+ "grad_norm": 0.0,
258
+ "learning_rate": 0.00015965292841648592,
259
+ "loss": 6.0046,
260
+ "mean_token_accuracy": 0.3859730801777914,
261
+ "num_tokens": 2209603.0,
262
+ "step": 280
263
+ },
264
+ {
265
+ "epoch": 0.6298785040385529,
266
+ "grad_norm": 0.0,
267
+ "learning_rate": 0.0001582067968185105,
268
+ "loss": 5.9887,
269
+ "mean_token_accuracy": 0.38499277490191164,
270
+ "num_tokens": 2288743.0,
271
+ "step": 290
272
+ },
273
+ {
274
+ "epoch": 0.6515984524536754,
275
+ "grad_norm": 0.0,
276
+ "learning_rate": 0.00015676066522053508,
277
+ "loss": 5.9788,
278
+ "mean_token_accuracy": 0.3864025830756873,
279
+ "num_tokens": 2367648.0,
280
+ "step": 300
281
+ },
282
+ {
283
+ "epoch": 0.6733184008687979,
284
+ "grad_norm": 0.0,
285
+ "learning_rate": 0.00015531453362255964,
286
+ "loss": 6.0162,
287
+ "mean_token_accuracy": 0.38221894631860776,
288
+ "num_tokens": 2446256.0,
289
+ "step": 310
290
+ },
291
+ {
292
+ "epoch": 0.6950383492839205,
293
+ "grad_norm": 0.0,
294
+ "learning_rate": 0.00015386840202458423,
295
+ "loss": 6.0017,
296
+ "mean_token_accuracy": 0.38538019855041056,
297
+ "num_tokens": 2525065.0,
298
+ "step": 320
299
+ },
300
+ {
301
+ "epoch": 0.716758297699043,
302
+ "grad_norm": 0.0,
303
+ "learning_rate": 0.00015242227042660883,
304
+ "loss": 5.9569,
305
+ "mean_token_accuracy": 0.3852926092222333,
306
+ "num_tokens": 2604183.0,
307
+ "step": 330
308
+ },
309
+ {
310
+ "epoch": 0.7384782461141655,
311
+ "grad_norm": 0.0,
312
+ "learning_rate": 0.00015097613882863342,
313
+ "loss": 6.0072,
314
+ "mean_token_accuracy": 0.38616998645011336,
315
+ "num_tokens": 2683084.0,
316
+ "step": 340
317
+ },
318
+ {
319
+ "epoch": 0.7601981945292879,
320
+ "grad_norm": 0.0,
321
+ "learning_rate": 0.000149530007230658,
322
+ "loss": 6.0428,
323
+ "mean_token_accuracy": 0.3831095602363348,
324
+ "num_tokens": 2761824.0,
325
+ "step": 350
326
+ },
327
+ {
328
+ "epoch": 0.7819181429444105,
329
+ "grad_norm": 0.0,
330
+ "learning_rate": 0.00014808387563268257,
331
+ "loss": 5.9781,
332
+ "mean_token_accuracy": 0.3861894382047467,
333
+ "num_tokens": 2840709.0,
334
+ "step": 360
335
+ },
336
+ {
337
+ "epoch": 0.803638091359533,
338
+ "grad_norm": 0.0,
339
+ "learning_rate": 0.00014663774403470716,
340
+ "loss": 5.9898,
341
+ "mean_token_accuracy": 0.3846066597965546,
342
+ "num_tokens": 2919385.0,
343
+ "step": 370
344
+ },
345
+ {
346
+ "epoch": 0.8253580397746555,
347
+ "grad_norm": 0.0,
348
+ "learning_rate": 0.00014519161243673173,
349
+ "loss": 6.0098,
350
+ "mean_token_accuracy": 0.38357423364650456,
351
+ "num_tokens": 2998308.0,
352
+ "step": 380
353
+ },
354
+ {
355
+ "epoch": 0.8470779881897781,
356
+ "grad_norm": 0.0,
357
+ "learning_rate": 0.00014374548083875634,
358
+ "loss": 6.0349,
359
+ "mean_token_accuracy": 0.3809299209038727,
360
+ "num_tokens": 3076962.0,
361
+ "step": 390
362
+ },
363
+ {
364
+ "epoch": 0.8687979366049006,
365
+ "grad_norm": 0.0,
366
+ "learning_rate": 0.0001422993492407809,
367
+ "loss": 6.0294,
368
+ "mean_token_accuracy": 0.38217092433478683,
369
+ "num_tokens": 3155983.0,
370
+ "step": 400
371
+ },
372
+ {
373
+ "epoch": 0.890517885020023,
374
+ "grad_norm": 0.0,
375
+ "learning_rate": 0.0001408532176428055,
376
+ "loss": 5.9728,
377
+ "mean_token_accuracy": 0.3870704318396747,
378
+ "num_tokens": 3234960.0,
379
+ "step": 410
380
+ },
381
+ {
382
+ "epoch": 0.9122378334351456,
383
+ "grad_norm": 0.0,
384
+ "learning_rate": 0.0001394070860448301,
385
+ "loss": 6.0541,
386
+ "mean_token_accuracy": 0.38209462116938087,
387
+ "num_tokens": 3313797.0,
388
+ "step": 420
389
+ },
390
+ {
391
+ "epoch": 0.9339577818502681,
392
+ "grad_norm": 0.0,
393
+ "learning_rate": 0.00013796095444685465,
394
+ "loss": 5.9871,
395
+ "mean_token_accuracy": 0.38401242352556436,
396
+ "num_tokens": 3392880.0,
397
+ "step": 430
398
+ },
399
+ {
400
+ "epoch": 0.9556777302653906,
401
+ "grad_norm": 0.0,
402
+ "learning_rate": 0.00013651482284887927,
403
+ "loss": 5.9901,
404
+ "mean_token_accuracy": 0.38375150066567587,
405
+ "num_tokens": 3471597.0,
406
+ "step": 440
407
+ },
408
+ {
409
+ "epoch": 0.9773976786805131,
410
+ "grad_norm": 0.0,
411
+ "learning_rate": 0.00013506869125090384,
412
+ "loss": 5.9797,
413
+ "mean_token_accuracy": 0.38728573790285736,
414
+ "num_tokens": 3550697.0,
415
+ "step": 450
416
+ },
417
+ {
418
+ "epoch": 0.9991176270956357,
419
+ "grad_norm": 0.0,
420
+ "learning_rate": 0.00013362255965292843,
421
+ "loss": 5.9971,
422
+ "mean_token_accuracy": 0.38420643559657036,
423
+ "num_tokens": 3629481.0,
424
+ "step": 460
425
+ },
426
+ {
427
+ "epoch": 1.0195479535736103,
428
+ "grad_norm": 0.0,
429
+ "learning_rate": 0.00013217642805495302,
430
+ "loss": 6.0223,
431
+ "mean_token_accuracy": 0.3833757105509308,
432
+ "num_tokens": 3703746.0,
433
+ "step": 470
434
+ },
435
+ {
436
+ "epoch": 1.0412679019887328,
437
+ "grad_norm": 0.0,
438
+ "learning_rate": 0.00013073029645697758,
439
+ "loss": 6.0146,
440
+ "mean_token_accuracy": 0.38417146790307016,
441
+ "num_tokens": 3782783.0,
442
+ "step": 480
443
+ },
444
+ {
445
+ "epoch": 1.0629878504038552,
446
+ "grad_norm": 0.0,
447
+ "learning_rate": 0.00012928416485900217,
448
+ "loss": 5.9713,
449
+ "mean_token_accuracy": 0.3863212176831439,
450
+ "num_tokens": 3862048.0,
451
+ "step": 490
452
+ },
453
+ {
454
+ "epoch": 1.0847077988189777,
455
+ "grad_norm": 0.0,
456
+ "learning_rate": 0.00012783803326102676,
457
+ "loss": 5.9944,
458
+ "mean_token_accuracy": 0.38522618702845646,
459
+ "num_tokens": 3941220.0,
460
+ "step": 500
461
+ },
462
+ {
463
+ "epoch": 1.1064277472341004,
464
+ "grad_norm": 0.0,
465
+ "learning_rate": 0.00012639190166305135,
466
+ "loss": 5.9895,
467
+ "mean_token_accuracy": 0.38548907284857703,
468
+ "num_tokens": 4020089.0,
469
+ "step": 510
470
+ },
471
+ {
472
+ "epoch": 1.128147695649223,
473
+ "grad_norm": 0.0,
474
+ "learning_rate": 0.00012494577006507592,
475
+ "loss": 5.9901,
476
+ "mean_token_accuracy": 0.3841517207212746,
477
+ "num_tokens": 4098981.0,
478
+ "step": 520
479
+ },
480
+ {
481
+ "epoch": 1.1498676440643454,
482
+ "grad_norm": 0.0,
483
+ "learning_rate": 0.0001234996384671005,
484
+ "loss": 6.0125,
485
+ "mean_token_accuracy": 0.38393973018974065,
486
+ "num_tokens": 4177983.0,
487
+ "step": 530
488
+ },
489
+ {
490
+ "epoch": 1.1715875924794679,
491
+ "grad_norm": 0.0,
492
+ "learning_rate": 0.00012205350686912509,
493
+ "loss": 5.9999,
494
+ "mean_token_accuracy": 0.3848019931232557,
495
+ "num_tokens": 4256818.0,
496
+ "step": 540
497
+ },
498
+ {
499
+ "epoch": 1.1933075408945903,
500
+ "grad_norm": 0.0,
501
+ "learning_rate": 0.00012060737527114966,
502
+ "loss": 6.0146,
503
+ "mean_token_accuracy": 0.3834400905063376,
504
+ "num_tokens": 4335754.0,
505
+ "step": 550
506
+ },
507
+ {
508
+ "epoch": 1.2150274893097128,
509
+ "grad_norm": 0.0,
510
+ "learning_rate": 0.00011916124367317427,
511
+ "loss": 5.9832,
512
+ "mean_token_accuracy": 0.38411646473687144,
513
+ "num_tokens": 4414480.0,
514
+ "step": 560
515
+ },
516
+ {
517
+ "epoch": 1.2367474377248353,
518
+ "grad_norm": 0.0,
519
+ "learning_rate": 0.00011771511207519885,
520
+ "loss": 6.0393,
521
+ "mean_token_accuracy": 0.3809544609161094,
522
+ "num_tokens": 4493321.0,
523
+ "step": 570
524
+ },
525
+ {
526
+ "epoch": 1.2584673861399578,
527
+ "grad_norm": 0.0,
528
+ "learning_rate": 0.00011626898047722344,
529
+ "loss": 5.9989,
530
+ "mean_token_accuracy": 0.38565440035890786,
531
+ "num_tokens": 4572297.0,
532
+ "step": 580
533
+ },
534
+ {
535
+ "epoch": 1.2801873345550805,
536
+ "grad_norm": 0.0,
537
+ "learning_rate": 0.00011482284887924801,
538
+ "loss": 5.9983,
539
+ "mean_token_accuracy": 0.3863610655302182,
540
+ "num_tokens": 4651247.0,
541
+ "step": 590
542
+ },
543
+ {
544
+ "epoch": 1.301907282970203,
545
+ "grad_norm": 0.0,
546
+ "learning_rate": 0.00011337671728127259,
547
+ "loss": 6.0116,
548
+ "mean_token_accuracy": 0.3831069741398096,
549
+ "num_tokens": 4730012.0,
550
+ "step": 600
551
+ },
552
+ {
553
+ "epoch": 1.3236272313853255,
554
+ "grad_norm": 0.0,
555
+ "learning_rate": 0.0001119305856832972,
556
+ "loss": 6.0195,
557
+ "mean_token_accuracy": 0.3857766560278833,
558
+ "num_tokens": 4809013.0,
559
+ "step": 610
560
+ },
561
+ {
562
+ "epoch": 1.345347179800448,
563
+ "grad_norm": 0.0,
564
+ "learning_rate": 0.00011048445408532177,
565
+ "loss": 6.0108,
566
+ "mean_token_accuracy": 0.3835394912166521,
567
+ "num_tokens": 4888027.0,
568
+ "step": 620
569
+ },
570
+ {
571
+ "epoch": 1.3670671282155704,
572
+ "grad_norm": 0.0,
573
+ "learning_rate": 0.00010903832248734635,
574
+ "loss": 5.9558,
575
+ "mean_token_accuracy": 0.3883327366434969,
576
+ "num_tokens": 4966859.0,
577
+ "step": 630
578
+ },
579
+ {
580
+ "epoch": 1.3887870766306931,
581
+ "grad_norm": 0.0,
582
+ "learning_rate": 0.00010759219088937094,
583
+ "loss": 6.0088,
584
+ "mean_token_accuracy": 0.385754154715687,
585
+ "num_tokens": 5045905.0,
586
+ "step": 640
587
+ },
588
+ {
589
+ "epoch": 1.4105070250458156,
590
+ "grad_norm": 0.0,
591
+ "learning_rate": 0.00010614605929139552,
592
+ "loss": 6.0236,
593
+ "mean_token_accuracy": 0.3829167880234309,
594
+ "num_tokens": 5124413.0,
595
+ "step": 650
596
+ },
597
+ {
598
+ "epoch": 1.432226973460938,
599
+ "grad_norm": 0.0,
600
+ "learning_rate": 0.0001046999276934201,
601
+ "loss": 6.0154,
602
+ "mean_token_accuracy": 0.3834054367849603,
603
+ "num_tokens": 5203273.0,
604
+ "step": 660
605
+ },
606
+ {
607
+ "epoch": 1.4539469218760606,
608
+ "grad_norm": 0.0,
609
+ "learning_rate": 0.0001032537960954447,
610
+ "loss": 6.005,
611
+ "mean_token_accuracy": 0.3848568681394681,
612
+ "num_tokens": 5281984.0,
613
+ "step": 670
614
+ },
615
+ {
616
+ "epoch": 1.475666870291183,
617
+ "grad_norm": 0.0,
618
+ "learning_rate": 0.00010180766449746928,
619
+ "loss": 5.9805,
620
+ "mean_token_accuracy": 0.3862292483681813,
621
+ "num_tokens": 5360846.0,
622
+ "step": 680
623
+ },
624
+ {
625
+ "epoch": 1.4973868187063055,
626
+ "grad_norm": 0.0,
627
+ "learning_rate": 0.00010036153289949386,
628
+ "loss": 5.9871,
629
+ "mean_token_accuracy": 0.3851381171494722,
630
+ "num_tokens": 5439758.0,
631
+ "step": 690
632
+ },
633
+ {
634
+ "epoch": 1.519106767121428,
635
+ "grad_norm": 0.0,
636
+ "learning_rate": 9.891540130151843e-05,
637
+ "loss": 6.0254,
638
+ "mean_token_accuracy": 0.3831110308994539,
639
+ "num_tokens": 5518768.0,
640
+ "step": 700
641
+ },
642
+ {
643
+ "epoch": 1.5408267155365505,
644
+ "grad_norm": 0.0,
645
+ "learning_rate": 9.746926970354303e-05,
646
+ "loss": 6.0339,
647
+ "mean_token_accuracy": 0.3808896674308926,
648
+ "num_tokens": 5597672.0,
649
+ "step": 710
650
+ },
651
+ {
652
+ "epoch": 1.562546663951673,
653
+ "grad_norm": 0.0,
654
+ "learning_rate": 9.602313810556762e-05,
655
+ "loss": 5.994,
656
+ "mean_token_accuracy": 0.3865658549708314,
657
+ "num_tokens": 5676393.0,
658
+ "step": 720
659
+ },
660
+ {
661
+ "epoch": 1.5842666123667957,
662
+ "grad_norm": 0.0,
663
+ "learning_rate": 9.45770065075922e-05,
664
+ "loss": 6.0026,
665
+ "mean_token_accuracy": 0.38676879862323404,
666
+ "num_tokens": 5755416.0,
667
+ "step": 730
668
+ },
669
+ {
670
+ "epoch": 1.6059865607819181,
671
+ "grad_norm": 0.0,
672
+ "learning_rate": 9.313087490961678e-05,
673
+ "loss": 6.0175,
674
+ "mean_token_accuracy": 0.38389978763880206,
675
+ "num_tokens": 5834322.0,
676
+ "step": 740
677
+ },
678
+ {
679
+ "epoch": 1.6277065091970406,
680
+ "grad_norm": 0.0,
681
+ "learning_rate": 9.168474331164136e-05,
682
+ "loss": 6.0224,
683
+ "mean_token_accuracy": 0.3840769061935134,
684
+ "num_tokens": 5912925.0,
685
+ "step": 750
686
+ },
687
+ {
688
+ "epoch": 1.6494264576121633,
689
+ "grad_norm": 0.0,
690
+ "learning_rate": 9.023861171366594e-05,
691
+ "loss": 5.9683,
692
+ "mean_token_accuracy": 0.38740592543035746,
693
+ "num_tokens": 5991913.0,
694
+ "step": 760
695
+ },
696
+ {
697
+ "epoch": 1.6711464060272858,
698
+ "grad_norm": 0.0,
699
+ "learning_rate": 8.879248011569053e-05,
700
+ "loss": 6.0136,
701
+ "mean_token_accuracy": 0.3827170055825263,
702
+ "num_tokens": 6070546.0,
703
+ "step": 770
704
+ },
705
+ {
706
+ "epoch": 1.6928663544424083,
707
+ "grad_norm": 0.0,
708
+ "learning_rate": 8.734634851771512e-05,
709
+ "loss": 5.9688,
710
+ "mean_token_accuracy": 0.3871995336958207,
711
+ "num_tokens": 6149623.0,
712
+ "step": 780
713
+ },
714
+ {
715
+ "epoch": 1.7145863028575308,
716
+ "grad_norm": 0.0,
717
+ "learning_rate": 8.59002169197397e-05,
718
+ "loss": 6.0639,
719
+ "mean_token_accuracy": 0.3800558194401674,
720
+ "num_tokens": 6228396.0,
721
+ "step": 790
722
+ },
723
+ {
724
+ "epoch": 1.7363062512726533,
725
+ "grad_norm": 0.0,
726
+ "learning_rate": 8.445408532176429e-05,
727
+ "loss": 5.9778,
728
+ "mean_token_accuracy": 0.3854194077430293,
729
+ "num_tokens": 6307375.0,
730
+ "step": 800
731
+ },
732
+ {
733
+ "epoch": 1.7580261996877757,
734
+ "grad_norm": 0.0,
735
+ "learning_rate": 8.300795372378887e-05,
736
+ "loss": 6.0294,
737
+ "mean_token_accuracy": 0.38350161886774004,
738
+ "num_tokens": 6386536.0,
739
+ "step": 810
740
+ },
741
+ {
742
+ "epoch": 1.7797461481028982,
743
+ "grad_norm": 0.0,
744
+ "learning_rate": 8.156182212581345e-05,
745
+ "loss": 5.9736,
746
+ "mean_token_accuracy": 0.38414000715129076,
747
+ "num_tokens": 6465708.0,
748
+ "step": 820
749
+ },
750
+ {
751
+ "epoch": 1.8014660965180207,
752
+ "grad_norm": 0.0,
753
+ "learning_rate": 8.011569052783804e-05,
754
+ "loss": 5.9847,
755
+ "mean_token_accuracy": 0.3877883184002712,
756
+ "num_tokens": 6544782.0,
757
+ "step": 830
758
+ },
759
+ {
760
+ "epoch": 1.8231860449331432,
761
+ "grad_norm": 0.0,
762
+ "learning_rate": 7.866955892986261e-05,
763
+ "loss": 6.043,
764
+ "mean_token_accuracy": 0.38251175949117167,
765
+ "num_tokens": 6623497.0,
766
+ "step": 840
767
+ },
768
+ {
769
+ "epoch": 1.8449059933482657,
770
+ "grad_norm": 0.0,
771
+ "learning_rate": 7.72234273318872e-05,
772
+ "loss": 6.0297,
773
+ "mean_token_accuracy": 0.38357739897910503,
774
+ "num_tokens": 6702159.0,
775
+ "step": 850
776
+ },
777
+ {
778
+ "epoch": 1.8666259417633884,
779
+ "grad_norm": 0.0,
780
+ "learning_rate": 7.57772957339118e-05,
781
+ "loss": 6.0335,
782
+ "mean_token_accuracy": 0.38290644709486515,
783
+ "num_tokens": 6780886.0,
784
+ "step": 860
785
+ },
786
+ {
787
+ "epoch": 1.8883458901785108,
788
+ "grad_norm": 0.0,
789
+ "learning_rate": 7.433116413593637e-05,
790
+ "loss": 6.0273,
791
+ "mean_token_accuracy": 0.38307004772359504,
792
+ "num_tokens": 6859533.0,
793
+ "step": 870
794
+ },
795
+ {
796
+ "epoch": 1.9100658385936333,
797
+ "grad_norm": 0.0,
798
+ "learning_rate": 7.288503253796096e-05,
799
+ "loss": 5.9827,
800
+ "mean_token_accuracy": 0.385389854805544,
801
+ "num_tokens": 6938490.0,
802
+ "step": 880
803
+ },
804
+ {
805
+ "epoch": 1.9317857870087558,
806
+ "grad_norm": 0.0,
807
+ "learning_rate": 7.143890093998554e-05,
808
+ "loss": 6.0212,
809
+ "mean_token_accuracy": 0.382453205762431,
810
+ "num_tokens": 7017297.0,
811
+ "step": 890
812
+ },
813
+ {
814
+ "epoch": 1.9535057354238785,
815
+ "grad_norm": 0.0,
816
+ "learning_rate": 6.999276934201012e-05,
817
+ "loss": 6.0043,
818
+ "mean_token_accuracy": 0.38360455771908164,
819
+ "num_tokens": 7096419.0,
820
+ "step": 900
821
+ },
822
+ {
823
+ "epoch": 1.975225683839001,
824
+ "grad_norm": 0.0,
825
+ "learning_rate": 6.854663774403471e-05,
826
+ "loss": 5.985,
827
+ "mean_token_accuracy": 0.38632277538999915,
828
+ "num_tokens": 7175650.0,
829
+ "step": 910
830
+ },
831
+ {
832
+ "epoch": 1.9969456322541235,
833
+ "grad_norm": 0.0,
834
+ "learning_rate": 6.71005061460593e-05,
835
+ "loss": 6.009,
836
+ "mean_token_accuracy": 0.38311313565354793,
837
+ "num_tokens": 7254378.0,
838
+ "step": 920
839
+ },
840
+ {
841
+ "epoch": 2.017375958732098,
842
+ "grad_norm": 0.0,
843
+ "learning_rate": 6.565437454808388e-05,
844
+ "loss": 5.968,
845
+ "mean_token_accuracy": 0.38770305797259674,
846
+ "num_tokens": 7328545.0,
847
+ "step": 930
848
+ },
849
+ {
850
+ "epoch": 2.0390959071472206,
851
+ "grad_norm": 0.0,
852
+ "learning_rate": 6.420824295010847e-05,
853
+ "loss": 5.9938,
854
+ "mean_token_accuracy": 0.38461664781207217,
855
+ "num_tokens": 7407672.0,
856
+ "step": 940
857
+ },
858
+ {
859
+ "epoch": 2.060815855562343,
860
+ "grad_norm": 0.0,
861
+ "learning_rate": 6.276211135213305e-05,
862
+ "loss": 6.0073,
863
+ "mean_token_accuracy": 0.3848728087265044,
864
+ "num_tokens": 7486852.0,
865
+ "step": 950
866
+ },
867
+ {
868
+ "epoch": 2.0825358039774655,
869
+ "grad_norm": 0.0,
870
+ "learning_rate": 6.131597975415762e-05,
871
+ "loss": 5.9564,
872
+ "mean_token_accuracy": 0.3856648310320452,
873
+ "num_tokens": 7565680.0,
874
+ "step": 960
875
+ },
876
+ {
877
+ "epoch": 2.104255752392588,
878
+ "grad_norm": 0.0,
879
+ "learning_rate": 5.9869848156182215e-05,
880
+ "loss": 6.0092,
881
+ "mean_token_accuracy": 0.38208459480665624,
882
+ "num_tokens": 7644328.0,
883
+ "step": 970
884
+ },
885
+ {
886
+ "epoch": 2.1259757008077105,
887
+ "grad_norm": 0.0,
888
+ "learning_rate": 5.84237165582068e-05,
889
+ "loss": 6.0331,
890
+ "mean_token_accuracy": 0.38403523166198283,
891
+ "num_tokens": 7723297.0,
892
+ "step": 980
893
+ },
894
+ {
895
+ "epoch": 2.147695649222833,
896
+ "grad_norm": 0.0,
897
+ "learning_rate": 5.697758496023138e-05,
898
+ "loss": 6.0003,
899
+ "mean_token_accuracy": 0.38371361922472713,
900
+ "num_tokens": 7802228.0,
901
+ "step": 990
902
+ },
903
+ {
904
+ "epoch": 2.1694155976379554,
905
+ "grad_norm": 0.0,
906
+ "learning_rate": 5.553145336225597e-05,
907
+ "loss": 5.9941,
908
+ "mean_token_accuracy": 0.386648180836346,
909
+ "num_tokens": 7881292.0,
910
+ "step": 1000
911
+ },
912
+ {
913
+ "epoch": 2.191135546053078,
914
+ "grad_norm": 0.0,
915
+ "learning_rate": 5.408532176428055e-05,
916
+ "loss": 5.9629,
917
+ "mean_token_accuracy": 0.3859988093841821,
918
+ "num_tokens": 7960306.0,
919
+ "step": 1010
920
+ },
921
+ {
922
+ "epoch": 2.212855494468201,
923
+ "grad_norm": 0.0,
924
+ "learning_rate": 5.263919016630514e-05,
925
+ "loss": 5.9615,
926
+ "mean_token_accuracy": 0.3859686201903969,
927
+ "num_tokens": 8039403.0,
928
+ "step": 1020
929
+ },
930
+ {
931
+ "epoch": 2.2345754428833233,
932
+ "grad_norm": 0.0,
933
+ "learning_rate": 5.119305856832972e-05,
934
+ "loss": 5.9744,
935
+ "mean_token_accuracy": 0.3849541787989438,
936
+ "num_tokens": 8118810.0,
937
+ "step": 1030
938
+ },
939
+ {
940
+ "epoch": 2.256295391298446,
941
+ "grad_norm": 0.0,
942
+ "learning_rate": 4.9746926970354304e-05,
943
+ "loss": 6.0192,
944
+ "mean_token_accuracy": 0.3843318622326478,
945
+ "num_tokens": 8197744.0,
946
+ "step": 1040
947
+ },
948
+ {
949
+ "epoch": 2.2780153397135683,
950
+ "grad_norm": 0.0,
951
+ "learning_rate": 4.830079537237889e-05,
952
+ "loss": 6.0364,
953
+ "mean_token_accuracy": 0.38417832332197577,
954
+ "num_tokens": 8276673.0,
955
+ "step": 1050
956
+ },
957
+ {
958
+ "epoch": 2.2997352881286908,
959
+ "grad_norm": 0.0,
960
+ "learning_rate": 4.685466377440347e-05,
961
+ "loss": 6.0178,
962
+ "mean_token_accuracy": 0.38303852297831326,
963
+ "num_tokens": 8355486.0,
964
+ "step": 1060
965
+ },
966
+ {
967
+ "epoch": 2.3214552365438132,
968
+ "grad_norm": 0.0,
969
+ "learning_rate": 4.540853217642806e-05,
970
+ "loss": 5.9567,
971
+ "mean_token_accuracy": 0.38857606865931305,
972
+ "num_tokens": 8434638.0,
973
+ "step": 1070
974
+ },
975
+ {
976
+ "epoch": 2.3431751849589357,
977
+ "grad_norm": 0.0,
978
+ "learning_rate": 4.396240057845264e-05,
979
+ "loss": 5.9893,
980
+ "mean_token_accuracy": 0.38565383905079215,
981
+ "num_tokens": 8513720.0,
982
+ "step": 1080
983
+ },
984
+ {
985
+ "epoch": 2.364895133374058,
986
+ "grad_norm": 0.0,
987
+ "learning_rate": 4.2516268980477226e-05,
988
+ "loss": 5.9812,
989
+ "mean_token_accuracy": 0.38872407528106123,
990
+ "num_tokens": 8592514.0,
991
+ "step": 1090
992
+ },
993
+ {
994
+ "epoch": 2.3866150817891807,
995
+ "grad_norm": 0.0,
996
+ "learning_rate": 4.107013738250181e-05,
997
+ "loss": 6.0234,
998
+ "mean_token_accuracy": 0.3825469396775588,
999
+ "num_tokens": 8671231.0,
1000
+ "step": 1100
1001
+ },
1002
+ {
1003
+ "epoch": 2.408335030204303,
1004
+ "grad_norm": 0.0,
1005
+ "learning_rate": 3.9624005784526394e-05,
1006
+ "loss": 6.0149,
1007
+ "mean_token_accuracy": 0.3845805463381112,
1008
+ "num_tokens": 8750288.0,
1009
+ "step": 1110
1010
+ },
1011
+ {
1012
+ "epoch": 2.4300549786194257,
1013
+ "grad_norm": 0.0,
1014
+ "learning_rate": 3.817787418655098e-05,
1015
+ "loss": 6.0056,
1016
+ "mean_token_accuracy": 0.38355012007523326,
1017
+ "num_tokens": 8829159.0,
1018
+ "step": 1120
1019
+ },
1020
+ {
1021
+ "epoch": 2.451774927034548,
1022
+ "grad_norm": 0.0,
1023
+ "learning_rate": 3.673174258857556e-05,
1024
+ "loss": 5.9945,
1025
+ "mean_token_accuracy": 0.38476743231294674,
1026
+ "num_tokens": 8907706.0,
1027
+ "step": 1130
1028
+ },
1029
+ {
1030
+ "epoch": 2.4734948754496706,
1031
+ "grad_norm": 0.0,
1032
+ "learning_rate": 3.5285610990600147e-05,
1033
+ "loss": 5.9712,
1034
+ "mean_token_accuracy": 0.3861268714419566,
1035
+ "num_tokens": 8986526.0,
1036
+ "step": 1140
1037
+ },
1038
+ {
1039
+ "epoch": 2.495214823864793,
1040
+ "grad_norm": 0.0,
1041
+ "learning_rate": 3.383947939262473e-05,
1042
+ "loss": 6.0243,
1043
+ "mean_token_accuracy": 0.3839300020830706,
1044
+ "num_tokens": 9065399.0,
1045
+ "step": 1150
1046
+ },
1047
+ {
1048
+ "epoch": 2.5169347722799156,
1049
+ "grad_norm": 0.0,
1050
+ "learning_rate": 3.2393347794649315e-05,
1051
+ "loss": 6.0197,
1052
+ "mean_token_accuracy": 0.3838517373194918,
1053
+ "num_tokens": 9144182.0,
1054
+ "step": 1160
1055
+ },
1056
+ {
1057
+ "epoch": 2.5386547206950385,
1058
+ "grad_norm": 0.0,
1059
+ "learning_rate": 3.09472161966739e-05,
1060
+ "loss": 6.0336,
1061
+ "mean_token_accuracy": 0.38345675652381034,
1062
+ "num_tokens": 9222743.0,
1063
+ "step": 1170
1064
+ },
1065
+ {
1066
+ "epoch": 2.560374669110161,
1067
+ "grad_norm": 0.0,
1068
+ "learning_rate": 2.950108459869848e-05,
1069
+ "loss": 6.0275,
1070
+ "mean_token_accuracy": 0.38377260488923637,
1071
+ "num_tokens": 9301686.0,
1072
+ "step": 1180
1073
+ },
1074
+ {
1075
+ "epoch": 2.5820946175252835,
1076
+ "grad_norm": 0.0,
1077
+ "learning_rate": 2.8054953000723068e-05,
1078
+ "loss": 6.0401,
1079
+ "mean_token_accuracy": 0.38272831091890114,
1080
+ "num_tokens": 9380398.0,
1081
+ "step": 1190
1082
+ },
1083
+ {
1084
+ "epoch": 2.603814565940406,
1085
+ "grad_norm": 0.0,
1086
+ "learning_rate": 2.6608821402747652e-05,
1087
+ "loss": 6.003,
1088
+ "mean_token_accuracy": 0.38574083771090956,
1089
+ "num_tokens": 9459228.0,
1090
+ "step": 1200
1091
+ },
1092
+ {
1093
+ "epoch": 2.6255345143555284,
1094
+ "grad_norm": 0.0,
1095
+ "learning_rate": 2.516268980477224e-05,
1096
+ "loss": 6.0233,
1097
+ "mean_token_accuracy": 0.3827782694483176,
1098
+ "num_tokens": 9538141.0,
1099
+ "step": 1210
1100
+ },
1101
+ {
1102
+ "epoch": 2.647254462770651,
1103
+ "grad_norm": 0.0,
1104
+ "learning_rate": 2.3716558206796817e-05,
1105
+ "loss": 6.0003,
1106
+ "mean_token_accuracy": 0.3853726448956877,
1107
+ "num_tokens": 9617091.0,
1108
+ "step": 1220
1109
+ },
1110
+ {
1111
+ "epoch": 2.6689744111857734,
1112
+ "grad_norm": 0.0,
1113
+ "learning_rate": 2.2270426608821404e-05,
1114
+ "loss": 6.0271,
1115
+ "mean_token_accuracy": 0.38530398234725,
1116
+ "num_tokens": 9696158.0,
1117
+ "step": 1230
1118
+ },
1119
+ {
1120
+ "epoch": 2.690694359600896,
1121
+ "grad_norm": 0.0,
1122
+ "learning_rate": 2.082429501084599e-05,
1123
+ "loss": 5.9816,
1124
+ "mean_token_accuracy": 0.38289073556661607,
1125
+ "num_tokens": 9775025.0,
1126
+ "step": 1240
1127
+ },
1128
+ {
1129
+ "epoch": 2.7124143080160183,
1130
+ "grad_norm": 0.0,
1131
+ "learning_rate": 1.9378163412870573e-05,
1132
+ "loss": 6.003,
1133
+ "mean_token_accuracy": 0.3852554757846519,
1134
+ "num_tokens": 9853991.0,
1135
+ "step": 1250
1136
+ },
1137
+ {
1138
+ "epoch": 2.734134256431141,
1139
+ "grad_norm": 0.0,
1140
+ "learning_rate": 1.7932031814895157e-05,
1141
+ "loss": 6.0387,
1142
+ "mean_token_accuracy": 0.38218798900488765,
1143
+ "num_tokens": 9932459.0,
1144
+ "step": 1260
1145
+ },
1146
+ {
1147
+ "epoch": 2.7558542048462638,
1148
+ "grad_norm": 0.0,
1149
+ "learning_rate": 1.648590021691974e-05,
1150
+ "loss": 6.0218,
1151
+ "mean_token_accuracy": 0.38361733745550736,
1152
+ "num_tokens": 10011147.0,
1153
+ "step": 1270
1154
+ },
1155
+ {
1156
+ "epoch": 2.7775741532613862,
1157
+ "grad_norm": 0.0,
1158
+ "learning_rate": 1.5039768618944326e-05,
1159
+ "loss": 6.0164,
1160
+ "mean_token_accuracy": 0.3833353664376773,
1161
+ "num_tokens": 10089867.0,
1162
+ "step": 1280
1163
+ },
1164
+ {
1165
+ "epoch": 2.7992941016765087,
1166
+ "grad_norm": 0.0,
1167
+ "learning_rate": 1.3593637020968908e-05,
1168
+ "loss": 6.0203,
1169
+ "mean_token_accuracy": 0.3844827389344573,
1170
+ "num_tokens": 10168558.0,
1171
+ "step": 1290
1172
+ },
1173
+ {
1174
+ "epoch": 2.821014050091631,
1175
+ "grad_norm": 0.0,
1176
+ "learning_rate": 1.2147505422993492e-05,
1177
+ "loss": 6.0223,
1178
+ "mean_token_accuracy": 0.38475555207114664,
1179
+ "num_tokens": 10247626.0,
1180
+ "step": 1300
1181
+ },
1182
+ {
1183
+ "epoch": 2.8427339985067537,
1184
+ "grad_norm": 0.0,
1185
+ "learning_rate": 1.0701373825018076e-05,
1186
+ "loss": 5.997,
1187
+ "mean_token_accuracy": 0.3833994207670912,
1188
+ "num_tokens": 10326786.0,
1189
+ "step": 1310
1190
+ },
1191
+ {
1192
+ "epoch": 2.864453946921876,
1193
+ "grad_norm": 0.0,
1194
+ "learning_rate": 9.25524222704266e-06,
1195
+ "loss": 6.0283,
1196
+ "mean_token_accuracy": 0.3812030048458837,
1197
+ "num_tokens": 10405619.0,
1198
+ "step": 1320
1199
+ },
1200
+ {
1201
+ "epoch": 2.8861738953369986,
1202
+ "grad_norm": 0.0,
1203
+ "learning_rate": 7.809110629067245e-06,
1204
+ "loss": 6.0422,
1205
+ "mean_token_accuracy": 0.381096905877348,
1206
+ "num_tokens": 10484466.0,
1207
+ "step": 1330
1208
+ },
1209
+ {
1210
+ "epoch": 2.907893843752121,
1211
+ "grad_norm": 0.0,
1212
+ "learning_rate": 6.36297903109183e-06,
1213
+ "loss": 5.9859,
1214
+ "mean_token_accuracy": 0.38405414449516684,
1215
+ "num_tokens": 10563350.0,
1216
+ "step": 1340
1217
+ },
1218
+ {
1219
+ "epoch": 2.9296137921672436,
1220
+ "grad_norm": 0.0,
1221
+ "learning_rate": 4.916847433116413e-06,
1222
+ "loss": 6.0283,
1223
+ "mean_token_accuracy": 0.38398846148047594,
1224
+ "num_tokens": 10642421.0,
1225
+ "step": 1350
1226
+ },
1227
+ {
1228
+ "epoch": 2.951333740582366,
1229
+ "grad_norm": 0.0,
1230
+ "learning_rate": 3.470715835140998e-06,
1231
+ "loss": 6.0132,
1232
+ "mean_token_accuracy": 0.3835720970411785,
1233
+ "num_tokens": 10721272.0,
1234
+ "step": 1360
1235
+ },
1236
+ {
1237
+ "epoch": 2.9730536889974886,
1238
+ "grad_norm": 0.0,
1239
+ "learning_rate": 2.024584237165582e-06,
1240
+ "loss": 6.0098,
1241
+ "mean_token_accuracy": 0.3844847684260458,
1242
+ "num_tokens": 10800144.0,
1243
+ "step": 1370
1244
+ },
1245
+ {
1246
+ "epoch": 2.994773637412611,
1247
+ "grad_norm": 0.0,
1248
+ "learning_rate": 5.784526391901663e-07,
1249
+ "loss": 5.962,
1250
+ "mean_token_accuracy": 0.3870904964976944,
1251
+ "num_tokens": 10879173.0,
1252
+ "step": 1380
1253
+ }
1254
+ ],
1255
+ "logging_steps": 10,
1256
+ "max_steps": 1383,
1257
+ "num_input_tokens_seen": 0,
1258
+ "num_train_epochs": 3,
1259
+ "save_steps": 500,
1260
+ "stateful_callbacks": {
1261
+ "TrainerControl": {
1262
+ "args": {
1263
+ "should_epoch_stop": false,
1264
+ "should_evaluate": false,
1265
+ "should_log": false,
1266
+ "should_save": true,
1267
+ "should_training_stop": true
1268
+ },
1269
+ "attributes": {}
1270
+ }
1271
+ },
1272
+ "total_flos": 4.6605276253782835e+17,
1273
+ "train_batch_size": 1,
1274
+ "trial_name": null,
1275
+ "trial_params": null
1276
+ }
checkpoint-1383/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131a0a21414525756ac7e71774e33b16bf1fd48bed5fd2484806044f425159ba
3
+ size 6161
checkpoint-461/README.md ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Mistral-7B-Instruct-v0.3
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:mistralai/Mistral-7B-Instruct-v0.3
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ ---
12
+
13
+ # Model Card for Model ID
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ <!-- Provide a longer summary of what this model is. -->
24
+
25
+
26
+
27
+ - **Developed by:** [More Information Needed]
28
+ - **Funded by [optional]:** [More Information Needed]
29
+ - **Shared by [optional]:** [More Information Needed]
30
+ - **Model type:** [More Information Needed]
31
+ - **Language(s) (NLP):** [More Information Needed]
32
+ - **License:** [More Information Needed]
33
+ - **Finetuned from model [optional]:** [More Information Needed]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ <!-- Provide the basic links for the model. -->
38
+
39
+ - **Repository:** [More Information Needed]
40
+ - **Paper [optional]:** [More Information Needed]
41
+ - **Demo [optional]:** [More Information Needed]
42
+
43
+ ## Uses
44
+
45
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
+
47
+ ### Direct Use
48
+
49
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Downstream Use [optional]
54
+
55
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+
57
+ [More Information Needed]
58
+
59
+ ### Out-of-Scope Use
60
+
61
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
74
+
75
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
76
+
77
+ ## How to Get Started with the Model
78
+
79
+ Use the code below to get started with the model.
80
+
81
+ [More Information Needed]
82
+
83
+ ## Training Details
84
+
85
+ ### Training Data
86
+
87
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
+
89
+ [More Information Needed]
90
+
91
+ ### Training Procedure
92
+
93
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
+
95
+ #### Preprocessing [optional]
96
+
97
+ [More Information Needed]
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
103
+
104
+ #### Speeds, Sizes, Times [optional]
105
+
106
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ## Evaluation
111
+
112
+ <!-- This section describes the evaluation protocols and provides the results. -->
113
+
114
+ ### Testing Data, Factors & Metrics
115
+
116
+ #### Testing Data
117
+
118
+ <!-- This should link to a Dataset Card if possible. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Factors
123
+
124
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
125
+
126
+ [More Information Needed]
127
+
128
+ #### Metrics
129
+
130
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
131
+
132
+ [More Information Needed]
133
+
134
+ ### Results
135
+
136
+ [More Information Needed]
137
+
138
+ #### Summary
139
+
140
+
141
+
142
+ ## Model Examination [optional]
143
+
144
+ <!-- Relevant interpretability work for the model goes here -->
145
+
146
+ [More Information Needed]
147
+
148
+ ## Environmental Impact
149
+
150
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
151
+
152
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
153
+
154
+ - **Hardware Type:** [More Information Needed]
155
+ - **Hours used:** [More Information Needed]
156
+ - **Cloud Provider:** [More Information Needed]
157
+ - **Compute Region:** [More Information Needed]
158
+ - **Carbon Emitted:** [More Information Needed]
159
+
160
+ ## Technical Specifications [optional]
161
+
162
+ ### Model Architecture and Objective
163
+
164
+ [More Information Needed]
165
+
166
+ ### Compute Infrastructure
167
+
168
+ [More Information Needed]
169
+
170
+ #### Hardware
171
+
172
+ [More Information Needed]
173
+
174
+ #### Software
175
+
176
+ [More Information Needed]
177
+
178
+ ## Citation [optional]
179
+
180
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
181
+
182
+ **BibTeX:**
183
+
184
+ [More Information Needed]
185
+
186
+ **APA:**
187
+
188
+ [More Information Needed]
189
+
190
+ ## Glossary [optional]
191
+
192
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
193
+
194
+ [More Information Needed]
195
+
196
+ ## More Information [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Authors [optional]
201
+
202
+ [More Information Needed]
203
+
204
+ ## Model Card Contact
205
+
206
+ [More Information Needed]
207
+ ### Framework versions
208
+
209
+ - PEFT 0.17.0
checkpoint-461/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj"
32
+ ],
33
+ "target_parameters": null,
34
+ "task_type": "CAUSAL_LM",
35
+ "trainable_token_indices": null,
36
+ "use_dora": false,
37
+ "use_qalora": false,
38
+ "use_rslora": false
39
+ }
checkpoint-461/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1790b7928f19c11eea9b3ae25f2fe1a066a98224c55c01b0bee9e895ec7a9e48
3
+ size 54560368
checkpoint-461/chat_template.jinja ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if messages[0]["role"] == "system" %}
2
+ {%- set system_message = messages[0]["content"] %}
3
+ {%- set loop_messages = messages[1:] %}
4
+ {%- else %}
5
+ {%- set loop_messages = messages %}
6
+ {%- endif %}
7
+ {%- if not tools is defined %}
8
+ {%- set tools = none %}
9
+ {%- endif %}
10
+ {%- set user_messages = loop_messages | selectattr("role", "equalto", "user") | list %}
11
+
12
+ {#- This block checks for alternating user/assistant messages, skipping tool calling messages #}
13
+ {%- set ns = namespace() %}
14
+ {%- set ns.index = 0 %}
15
+ {%- for message in loop_messages %}
16
+ {%- if not (message.role == "tool" or message.role == "tool_results" or (message.tool_calls is defined and message.tool_calls is not none)) %}
17
+ {%- if (message["role"] == "user") != (ns.index % 2 == 0) %}
18
+ {{- raise_exception("After the optional system message, conversation roles must alternate user/assistant/user/assistant/...") }}
19
+ {%- endif %}
20
+ {%- set ns.index = ns.index + 1 %}
21
+ {%- endif %}
22
+ {%- endfor %}
23
+
24
+ {{- bos_token }}
25
+ {%- for message in loop_messages %}
26
+ {%- if message["role"] == "user" %}
27
+ {%- if tools is not none and (message == user_messages[-1]) %}
28
+ {{- "[AVAILABLE_TOOLS] [" }}
29
+ {%- for tool in tools %}
30
+ {%- set tool = tool.function %}
31
+ {{- '{"type": "function", "function": {' }}
32
+ {%- for key, val in tool.items() if key != "return" %}
33
+ {%- if val is string %}
34
+ {{- '"' + key + '": "' + val + '"' }}
35
+ {%- else %}
36
+ {{- '"' + key + '": ' + val|tojson }}
37
+ {%- endif %}
38
+ {%- if not loop.last %}
39
+ {{- ", " }}
40
+ {%- endif %}
41
+ {%- endfor %}
42
+ {{- "}}" }}
43
+ {%- if not loop.last %}
44
+ {{- ", " }}
45
+ {%- else %}
46
+ {{- "]" }}
47
+ {%- endif %}
48
+ {%- endfor %}
49
+ {{- "[/AVAILABLE_TOOLS]" }}
50
+ {%- endif %}
51
+ {%- if loop.last and system_message is defined %}
52
+ {{- "[INST] " + system_message + "\n\n" + message["content"] + "[/INST]" }}
53
+ {%- else %}
54
+ {{- "[INST] " + message["content"] + "[/INST]" }}
55
+ {%- endif %}
56
+ {%- elif message.tool_calls is defined and message.tool_calls is not none %}
57
+ {{- "[TOOL_CALLS] [" }}
58
+ {%- for tool_call in message.tool_calls %}
59
+ {%- set out = tool_call.function|tojson %}
60
+ {{- out[:-1] }}
61
+ {%- if not tool_call.id is defined or tool_call.id|length != 9 %}
62
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
63
+ {%- endif %}
64
+ {{- ', "id": "' + tool_call.id + '"}' }}
65
+ {%- if not loop.last %}
66
+ {{- ", " }}
67
+ {%- else %}
68
+ {{- "]" + eos_token }}
69
+ {%- endif %}
70
+ {%- endfor %}
71
+ {%- elif message["role"] == "assistant" %}
72
+ {{- " " + message["content"]|trim + eos_token}}
73
+ {%- elif message["role"] == "tool_results" or message["role"] == "tool" %}
74
+ {%- if message.content is defined and message.content.content is defined %}
75
+ {%- set content = message.content.content %}
76
+ {%- else %}
77
+ {%- set content = message.content %}
78
+ {%- endif %}
79
+ {{- '[TOOL_RESULTS] {"content": ' + content|string + ", " }}
80
+ {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}
81
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
82
+ {%- endif %}
83
+ {{- '"call_id": "' + message.tool_call_id + '"}[/TOOL_RESULTS]' }}
84
+ {%- else %}
85
+ {{- raise_exception("Only user and assistant roles are supported, with the exception of an initial optional system message!") }}
86
+ {%- endif %}
87
+ {%- endfor %}
checkpoint-461/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05d4690c9f75679e3413c951426acd91577e23ee13a26df1957e90b9e5a8893f
3
+ size 1593
checkpoint-461/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fea141dda0bb9d06867529f5f3e0869ed584743a0c64dab3e70fc8a6c34be4ad
3
+ size 14645
checkpoint-461/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ce7281ad70bf1f6664bc4ba855e401060fe9cb5ebab1f99f4310600a12d6c33
3
+ size 1465
checkpoint-461/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
checkpoint-461/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
checkpoint-461/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-461/trainer_state.json ADDED
@@ -0,0 +1,448 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 461,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.021719948415122516,
14
+ "grad_norm": 0.0,
15
+ "learning_rate": 0.00019869848156182213,
16
+ "loss": 6.0531,
17
+ "mean_token_accuracy": 0.37881034857127815,
18
+ "num_tokens": 78573.0,
19
+ "step": 10
20
+ },
21
+ {
22
+ "epoch": 0.04343989683024503,
23
+ "grad_norm": 0.0,
24
+ "learning_rate": 0.00019725234996384672,
25
+ "loss": 5.9903,
26
+ "mean_token_accuracy": 0.3887314551044255,
27
+ "num_tokens": 157496.0,
28
+ "step": 20
29
+ },
30
+ {
31
+ "epoch": 0.06515984524536754,
32
+ "grad_norm": 0.0,
33
+ "learning_rate": 0.00019580621836587129,
34
+ "loss": 5.9929,
35
+ "mean_token_accuracy": 0.3850487937917933,
36
+ "num_tokens": 236512.0,
37
+ "step": 30
38
+ },
39
+ {
40
+ "epoch": 0.08687979366049006,
41
+ "grad_norm": 0.0,
42
+ "learning_rate": 0.00019436008676789588,
43
+ "loss": 5.986,
44
+ "mean_token_accuracy": 0.3865805763518438,
45
+ "num_tokens": 315608.0,
46
+ "step": 40
47
+ },
48
+ {
49
+ "epoch": 0.10859974207561257,
50
+ "grad_norm": 0.0,
51
+ "learning_rate": 0.00019291395516992047,
52
+ "loss": 6.0318,
53
+ "mean_token_accuracy": 0.38490854618139564,
54
+ "num_tokens": 394791.0,
55
+ "step": 50
56
+ },
57
+ {
58
+ "epoch": 0.13031969049073508,
59
+ "grad_norm": 0.0,
60
+ "learning_rate": 0.00019146782357194506,
61
+ "loss": 5.9875,
62
+ "mean_token_accuracy": 0.38546840590424836,
63
+ "num_tokens": 473456.0,
64
+ "step": 60
65
+ },
66
+ {
67
+ "epoch": 0.1520396389058576,
68
+ "grad_norm": 0.0,
69
+ "learning_rate": 0.00019002169197396965,
70
+ "loss": 5.9773,
71
+ "mean_token_accuracy": 0.3837947838823311,
72
+ "num_tokens": 552247.0,
73
+ "step": 70
74
+ },
75
+ {
76
+ "epoch": 0.17375958732098012,
77
+ "grad_norm": 0.0,
78
+ "learning_rate": 0.0001885755603759942,
79
+ "loss": 6.0197,
80
+ "mean_token_accuracy": 0.38375892234034836,
81
+ "num_tokens": 631096.0,
82
+ "step": 80
83
+ },
84
+ {
85
+ "epoch": 0.19547953573610263,
86
+ "grad_norm": 0.0,
87
+ "learning_rate": 0.0001871294287780188,
88
+ "loss": 6.0014,
89
+ "mean_token_accuracy": 0.38519658900331705,
90
+ "num_tokens": 710275.0,
91
+ "step": 90
92
+ },
93
+ {
94
+ "epoch": 0.21719948415122514,
95
+ "grad_norm": 0.0,
96
+ "learning_rate": 0.0001856832971800434,
97
+ "loss": 6.0217,
98
+ "mean_token_accuracy": 0.3826989881345071,
99
+ "num_tokens": 789174.0,
100
+ "step": 100
101
+ },
102
+ {
103
+ "epoch": 0.23891943256634765,
104
+ "grad_norm": 0.0,
105
+ "learning_rate": 0.00018423716558206799,
106
+ "loss": 6.0294,
107
+ "mean_token_accuracy": 0.3842650496866554,
108
+ "num_tokens": 867922.0,
109
+ "step": 110
110
+ },
111
+ {
112
+ "epoch": 0.26063938098147016,
113
+ "grad_norm": 0.0,
114
+ "learning_rate": 0.00018279103398409255,
115
+ "loss": 5.9548,
116
+ "mean_token_accuracy": 0.3875111517496407,
117
+ "num_tokens": 947073.0,
118
+ "step": 120
119
+ },
120
+ {
121
+ "epoch": 0.2823593293965927,
122
+ "grad_norm": 0.0,
123
+ "learning_rate": 0.00018134490238611714,
124
+ "loss": 6.0071,
125
+ "mean_token_accuracy": 0.38376112943515184,
126
+ "num_tokens": 1026211.0,
127
+ "step": 130
128
+ },
129
+ {
130
+ "epoch": 0.3040792778117152,
131
+ "grad_norm": 0.0,
132
+ "learning_rate": 0.00017989877078814173,
133
+ "loss": 6.0105,
134
+ "mean_token_accuracy": 0.38463024909142407,
135
+ "num_tokens": 1105147.0,
136
+ "step": 140
137
+ },
138
+ {
139
+ "epoch": 0.3257992262268377,
140
+ "grad_norm": 0.0,
141
+ "learning_rate": 0.0001784526391901663,
142
+ "loss": 6.0449,
143
+ "mean_token_accuracy": 0.38241248747799544,
144
+ "num_tokens": 1183837.0,
145
+ "step": 150
146
+ },
147
+ {
148
+ "epoch": 0.34751917464196025,
149
+ "grad_norm": 0.0,
150
+ "learning_rate": 0.00017700650759219091,
151
+ "loss": 5.9946,
152
+ "mean_token_accuracy": 0.3821917780907825,
153
+ "num_tokens": 1262676.0,
154
+ "step": 160
155
+ },
156
+ {
157
+ "epoch": 0.36923912305708273,
158
+ "grad_norm": 0.0,
159
+ "learning_rate": 0.00017556037599421548,
160
+ "loss": 6.0049,
161
+ "mean_token_accuracy": 0.3842250820598565,
162
+ "num_tokens": 1341759.0,
163
+ "step": 170
164
+ },
165
+ {
166
+ "epoch": 0.39095907147220527,
167
+ "grad_norm": 0.0,
168
+ "learning_rate": 0.00017411424439624007,
169
+ "loss": 5.9736,
170
+ "mean_token_accuracy": 0.3861917880363762,
171
+ "num_tokens": 1420987.0,
172
+ "step": 180
173
+ },
174
+ {
175
+ "epoch": 0.41267901988732775,
176
+ "grad_norm": 0.0,
177
+ "learning_rate": 0.00017266811279826466,
178
+ "loss": 6.0431,
179
+ "mean_token_accuracy": 0.38146835477091373,
180
+ "num_tokens": 1499903.0,
181
+ "step": 190
182
+ },
183
+ {
184
+ "epoch": 0.4343989683024503,
185
+ "grad_norm": 0.0,
186
+ "learning_rate": 0.00017122198120028922,
187
+ "loss": 6.0069,
188
+ "mean_token_accuracy": 0.3861038032686338,
189
+ "num_tokens": 1578880.0,
190
+ "step": 200
191
+ },
192
+ {
193
+ "epoch": 0.4561189167175728,
194
+ "grad_norm": 0.0,
195
+ "learning_rate": 0.00016977584960231381,
196
+ "loss": 6.0091,
197
+ "mean_token_accuracy": 0.3845075036631897,
198
+ "num_tokens": 1657709.0,
199
+ "step": 210
200
+ },
201
+ {
202
+ "epoch": 0.4778388651326953,
203
+ "grad_norm": 0.0,
204
+ "learning_rate": 0.0001683297180043384,
205
+ "loss": 6.0029,
206
+ "mean_token_accuracy": 0.38288453239947556,
207
+ "num_tokens": 1736620.0,
208
+ "step": 220
209
+ },
210
+ {
211
+ "epoch": 0.49955881354781784,
212
+ "grad_norm": 0.0,
213
+ "learning_rate": 0.000166883586406363,
214
+ "loss": 5.9933,
215
+ "mean_token_accuracy": 0.38531269936356694,
216
+ "num_tokens": 1815478.0,
217
+ "step": 230
218
+ },
219
+ {
220
+ "epoch": 0.5212787619629403,
221
+ "grad_norm": 0.0,
222
+ "learning_rate": 0.00016543745480838756,
223
+ "loss": 6.0334,
224
+ "mean_token_accuracy": 0.3833939728909172,
225
+ "num_tokens": 1894256.0,
226
+ "step": 240
227
+ },
228
+ {
229
+ "epoch": 0.5429987103780628,
230
+ "grad_norm": 0.0,
231
+ "learning_rate": 0.00016399132321041215,
232
+ "loss": 6.0484,
233
+ "mean_token_accuracy": 0.3841994108865038,
234
+ "num_tokens": 1972997.0,
235
+ "step": 250
236
+ },
237
+ {
238
+ "epoch": 0.5647186587931854,
239
+ "grad_norm": 0.0,
240
+ "learning_rate": 0.00016254519161243674,
241
+ "loss": 6.0285,
242
+ "mean_token_accuracy": 0.38443261398933826,
243
+ "num_tokens": 2051890.0,
244
+ "step": 260
245
+ },
246
+ {
247
+ "epoch": 0.5864386072083079,
248
+ "grad_norm": 0.0,
249
+ "learning_rate": 0.00016109906001446133,
250
+ "loss": 5.983,
251
+ "mean_token_accuracy": 0.385862308065407,
252
+ "num_tokens": 2130746.0,
253
+ "step": 270
254
+ },
255
+ {
256
+ "epoch": 0.6081585556234304,
257
+ "grad_norm": 0.0,
258
+ "learning_rate": 0.00015965292841648592,
259
+ "loss": 6.0046,
260
+ "mean_token_accuracy": 0.3859730801777914,
261
+ "num_tokens": 2209603.0,
262
+ "step": 280
263
+ },
264
+ {
265
+ "epoch": 0.6298785040385529,
266
+ "grad_norm": 0.0,
267
+ "learning_rate": 0.0001582067968185105,
268
+ "loss": 5.9887,
269
+ "mean_token_accuracy": 0.38499277490191164,
270
+ "num_tokens": 2288743.0,
271
+ "step": 290
272
+ },
273
+ {
274
+ "epoch": 0.6515984524536754,
275
+ "grad_norm": 0.0,
276
+ "learning_rate": 0.00015676066522053508,
277
+ "loss": 5.9788,
278
+ "mean_token_accuracy": 0.3864025830756873,
279
+ "num_tokens": 2367648.0,
280
+ "step": 300
281
+ },
282
+ {
283
+ "epoch": 0.6733184008687979,
284
+ "grad_norm": 0.0,
285
+ "learning_rate": 0.00015531453362255964,
286
+ "loss": 6.0162,
287
+ "mean_token_accuracy": 0.38221894631860776,
288
+ "num_tokens": 2446256.0,
289
+ "step": 310
290
+ },
291
+ {
292
+ "epoch": 0.6950383492839205,
293
+ "grad_norm": 0.0,
294
+ "learning_rate": 0.00015386840202458423,
295
+ "loss": 6.0017,
296
+ "mean_token_accuracy": 0.38538019855041056,
297
+ "num_tokens": 2525065.0,
298
+ "step": 320
299
+ },
300
+ {
301
+ "epoch": 0.716758297699043,
302
+ "grad_norm": 0.0,
303
+ "learning_rate": 0.00015242227042660883,
304
+ "loss": 5.9569,
305
+ "mean_token_accuracy": 0.3852926092222333,
306
+ "num_tokens": 2604183.0,
307
+ "step": 330
308
+ },
309
+ {
310
+ "epoch": 0.7384782461141655,
311
+ "grad_norm": 0.0,
312
+ "learning_rate": 0.00015097613882863342,
313
+ "loss": 6.0072,
314
+ "mean_token_accuracy": 0.38616998645011336,
315
+ "num_tokens": 2683084.0,
316
+ "step": 340
317
+ },
318
+ {
319
+ "epoch": 0.7601981945292879,
320
+ "grad_norm": 0.0,
321
+ "learning_rate": 0.000149530007230658,
322
+ "loss": 6.0428,
323
+ "mean_token_accuracy": 0.3831095602363348,
324
+ "num_tokens": 2761824.0,
325
+ "step": 350
326
+ },
327
+ {
328
+ "epoch": 0.7819181429444105,
329
+ "grad_norm": 0.0,
330
+ "learning_rate": 0.00014808387563268257,
331
+ "loss": 5.9781,
332
+ "mean_token_accuracy": 0.3861894382047467,
333
+ "num_tokens": 2840709.0,
334
+ "step": 360
335
+ },
336
+ {
337
+ "epoch": 0.803638091359533,
338
+ "grad_norm": 0.0,
339
+ "learning_rate": 0.00014663774403470716,
340
+ "loss": 5.9898,
341
+ "mean_token_accuracy": 0.3846066597965546,
342
+ "num_tokens": 2919385.0,
343
+ "step": 370
344
+ },
345
+ {
346
+ "epoch": 0.8253580397746555,
347
+ "grad_norm": 0.0,
348
+ "learning_rate": 0.00014519161243673173,
349
+ "loss": 6.0098,
350
+ "mean_token_accuracy": 0.38357423364650456,
351
+ "num_tokens": 2998308.0,
352
+ "step": 380
353
+ },
354
+ {
355
+ "epoch": 0.8470779881897781,
356
+ "grad_norm": 0.0,
357
+ "learning_rate": 0.00014374548083875634,
358
+ "loss": 6.0349,
359
+ "mean_token_accuracy": 0.3809299209038727,
360
+ "num_tokens": 3076962.0,
361
+ "step": 390
362
+ },
363
+ {
364
+ "epoch": 0.8687979366049006,
365
+ "grad_norm": 0.0,
366
+ "learning_rate": 0.0001422993492407809,
367
+ "loss": 6.0294,
368
+ "mean_token_accuracy": 0.38217092433478683,
369
+ "num_tokens": 3155983.0,
370
+ "step": 400
371
+ },
372
+ {
373
+ "epoch": 0.890517885020023,
374
+ "grad_norm": 0.0,
375
+ "learning_rate": 0.0001408532176428055,
376
+ "loss": 5.9728,
377
+ "mean_token_accuracy": 0.3870704318396747,
378
+ "num_tokens": 3234960.0,
379
+ "step": 410
380
+ },
381
+ {
382
+ "epoch": 0.9122378334351456,
383
+ "grad_norm": 0.0,
384
+ "learning_rate": 0.0001394070860448301,
385
+ "loss": 6.0541,
386
+ "mean_token_accuracy": 0.38209462116938087,
387
+ "num_tokens": 3313797.0,
388
+ "step": 420
389
+ },
390
+ {
391
+ "epoch": 0.9339577818502681,
392
+ "grad_norm": 0.0,
393
+ "learning_rate": 0.00013796095444685465,
394
+ "loss": 5.9871,
395
+ "mean_token_accuracy": 0.38401242352556436,
396
+ "num_tokens": 3392880.0,
397
+ "step": 430
398
+ },
399
+ {
400
+ "epoch": 0.9556777302653906,
401
+ "grad_norm": 0.0,
402
+ "learning_rate": 0.00013651482284887927,
403
+ "loss": 5.9901,
404
+ "mean_token_accuracy": 0.38375150066567587,
405
+ "num_tokens": 3471597.0,
406
+ "step": 440
407
+ },
408
+ {
409
+ "epoch": 0.9773976786805131,
410
+ "grad_norm": 0.0,
411
+ "learning_rate": 0.00013506869125090384,
412
+ "loss": 5.9797,
413
+ "mean_token_accuracy": 0.38728573790285736,
414
+ "num_tokens": 3550697.0,
415
+ "step": 450
416
+ },
417
+ {
418
+ "epoch": 0.9991176270956357,
419
+ "grad_norm": 0.0,
420
+ "learning_rate": 0.00013362255965292843,
421
+ "loss": 5.9971,
422
+ "mean_token_accuracy": 0.38420643559657036,
423
+ "num_tokens": 3629481.0,
424
+ "step": 460
425
+ }
426
+ ],
427
+ "logging_steps": 10,
428
+ "max_steps": 1383,
429
+ "num_input_tokens_seen": 0,
430
+ "num_train_epochs": 3,
431
+ "save_steps": 500,
432
+ "stateful_callbacks": {
433
+ "TrainerControl": {
434
+ "args": {
435
+ "should_epoch_stop": false,
436
+ "should_evaluate": false,
437
+ "should_log": false,
438
+ "should_save": true,
439
+ "should_training_stop": false
440
+ },
441
+ "attributes": {}
442
+ }
443
+ },
444
+ "total_flos": 1.553509208459428e+17,
445
+ "train_batch_size": 1,
446
+ "trial_name": null,
447
+ "trial_params": null
448
+ }
checkpoint-461/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131a0a21414525756ac7e71774e33b16bf1fd48bed5fd2484806044f425159ba
3
+ size 6161
checkpoint-922/README.md ADDED
@@ -0,0 +1,209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Mistral-7B-Instruct-v0.3
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:mistralai/Mistral-7B-Instruct-v0.3
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ ---
12
+
13
+ # Model Card for Model ID
14
+
15
+ <!-- Provide a quick summary of what the model is/does. -->
16
+
17
+
18
+
19
+ ## Model Details
20
+
21
+ ### Model Description
22
+
23
+ <!-- Provide a longer summary of what this model is. -->
24
+
25
+
26
+
27
+ - **Developed by:** [More Information Needed]
28
+ - **Funded by [optional]:** [More Information Needed]
29
+ - **Shared by [optional]:** [More Information Needed]
30
+ - **Model type:** [More Information Needed]
31
+ - **Language(s) (NLP):** [More Information Needed]
32
+ - **License:** [More Information Needed]
33
+ - **Finetuned from model [optional]:** [More Information Needed]
34
+
35
+ ### Model Sources [optional]
36
+
37
+ <!-- Provide the basic links for the model. -->
38
+
39
+ - **Repository:** [More Information Needed]
40
+ - **Paper [optional]:** [More Information Needed]
41
+ - **Demo [optional]:** [More Information Needed]
42
+
43
+ ## Uses
44
+
45
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
+
47
+ ### Direct Use
48
+
49
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
50
+
51
+ [More Information Needed]
52
+
53
+ ### Downstream Use [optional]
54
+
55
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
56
+
57
+ [More Information Needed]
58
+
59
+ ### Out-of-Scope Use
60
+
61
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
62
+
63
+ [More Information Needed]
64
+
65
+ ## Bias, Risks, and Limitations
66
+
67
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
68
+
69
+ [More Information Needed]
70
+
71
+ ### Recommendations
72
+
73
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
74
+
75
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
76
+
77
+ ## How to Get Started with the Model
78
+
79
+ Use the code below to get started with the model.
80
+
81
+ [More Information Needed]
82
+
83
+ ## Training Details
84
+
85
+ ### Training Data
86
+
87
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
88
+
89
+ [More Information Needed]
90
+
91
+ ### Training Procedure
92
+
93
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
94
+
95
+ #### Preprocessing [optional]
96
+
97
+ [More Information Needed]
98
+
99
+
100
+ #### Training Hyperparameters
101
+
102
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
103
+
104
+ #### Speeds, Sizes, Times [optional]
105
+
106
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
107
+
108
+ [More Information Needed]
109
+
110
+ ## Evaluation
111
+
112
+ <!-- This section describes the evaluation protocols and provides the results. -->
113
+
114
+ ### Testing Data, Factors & Metrics
115
+
116
+ #### Testing Data
117
+
118
+ <!-- This should link to a Dataset Card if possible. -->
119
+
120
+ [More Information Needed]
121
+
122
+ #### Factors
123
+
124
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
125
+
126
+ [More Information Needed]
127
+
128
+ #### Metrics
129
+
130
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
131
+
132
+ [More Information Needed]
133
+
134
+ ### Results
135
+
136
+ [More Information Needed]
137
+
138
+ #### Summary
139
+
140
+
141
+
142
+ ## Model Examination [optional]
143
+
144
+ <!-- Relevant interpretability work for the model goes here -->
145
+
146
+ [More Information Needed]
147
+
148
+ ## Environmental Impact
149
+
150
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
151
+
152
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
153
+
154
+ - **Hardware Type:** [More Information Needed]
155
+ - **Hours used:** [More Information Needed]
156
+ - **Cloud Provider:** [More Information Needed]
157
+ - **Compute Region:** [More Information Needed]
158
+ - **Carbon Emitted:** [More Information Needed]
159
+
160
+ ## Technical Specifications [optional]
161
+
162
+ ### Model Architecture and Objective
163
+
164
+ [More Information Needed]
165
+
166
+ ### Compute Infrastructure
167
+
168
+ [More Information Needed]
169
+
170
+ #### Hardware
171
+
172
+ [More Information Needed]
173
+
174
+ #### Software
175
+
176
+ [More Information Needed]
177
+
178
+ ## Citation [optional]
179
+
180
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
181
+
182
+ **BibTeX:**
183
+
184
+ [More Information Needed]
185
+
186
+ **APA:**
187
+
188
+ [More Information Needed]
189
+
190
+ ## Glossary [optional]
191
+
192
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
193
+
194
+ [More Information Needed]
195
+
196
+ ## More Information [optional]
197
+
198
+ [More Information Needed]
199
+
200
+ ## Model Card Authors [optional]
201
+
202
+ [More Information Needed]
203
+
204
+ ## Model Card Contact
205
+
206
+ [More Information Needed]
207
+ ### Framework versions
208
+
209
+ - PEFT 0.17.0
checkpoint-922/adapter_config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.3",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 32,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 16,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "o_proj",
30
+ "q_proj",
31
+ "k_proj"
32
+ ],
33
+ "target_parameters": null,
34
+ "task_type": "CAUSAL_LM",
35
+ "trainable_token_indices": null,
36
+ "use_dora": false,
37
+ "use_qalora": false,
38
+ "use_rslora": false
39
+ }
checkpoint-922/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1790b7928f19c11eea9b3ae25f2fe1a066a98224c55c01b0bee9e895ec7a9e48
3
+ size 54560368
checkpoint-922/chat_template.jinja ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if messages[0]["role"] == "system" %}
2
+ {%- set system_message = messages[0]["content"] %}
3
+ {%- set loop_messages = messages[1:] %}
4
+ {%- else %}
5
+ {%- set loop_messages = messages %}
6
+ {%- endif %}
7
+ {%- if not tools is defined %}
8
+ {%- set tools = none %}
9
+ {%- endif %}
10
+ {%- set user_messages = loop_messages | selectattr("role", "equalto", "user") | list %}
11
+
12
+ {#- This block checks for alternating user/assistant messages, skipping tool calling messages #}
13
+ {%- set ns = namespace() %}
14
+ {%- set ns.index = 0 %}
15
+ {%- for message in loop_messages %}
16
+ {%- if not (message.role == "tool" or message.role == "tool_results" or (message.tool_calls is defined and message.tool_calls is not none)) %}
17
+ {%- if (message["role"] == "user") != (ns.index % 2 == 0) %}
18
+ {{- raise_exception("After the optional system message, conversation roles must alternate user/assistant/user/assistant/...") }}
19
+ {%- endif %}
20
+ {%- set ns.index = ns.index + 1 %}
21
+ {%- endif %}
22
+ {%- endfor %}
23
+
24
+ {{- bos_token }}
25
+ {%- for message in loop_messages %}
26
+ {%- if message["role"] == "user" %}
27
+ {%- if tools is not none and (message == user_messages[-1]) %}
28
+ {{- "[AVAILABLE_TOOLS] [" }}
29
+ {%- for tool in tools %}
30
+ {%- set tool = tool.function %}
31
+ {{- '{"type": "function", "function": {' }}
32
+ {%- for key, val in tool.items() if key != "return" %}
33
+ {%- if val is string %}
34
+ {{- '"' + key + '": "' + val + '"' }}
35
+ {%- else %}
36
+ {{- '"' + key + '": ' + val|tojson }}
37
+ {%- endif %}
38
+ {%- if not loop.last %}
39
+ {{- ", " }}
40
+ {%- endif %}
41
+ {%- endfor %}
42
+ {{- "}}" }}
43
+ {%- if not loop.last %}
44
+ {{- ", " }}
45
+ {%- else %}
46
+ {{- "]" }}
47
+ {%- endif %}
48
+ {%- endfor %}
49
+ {{- "[/AVAILABLE_TOOLS]" }}
50
+ {%- endif %}
51
+ {%- if loop.last and system_message is defined %}
52
+ {{- "[INST] " + system_message + "\n\n" + message["content"] + "[/INST]" }}
53
+ {%- else %}
54
+ {{- "[INST] " + message["content"] + "[/INST]" }}
55
+ {%- endif %}
56
+ {%- elif message.tool_calls is defined and message.tool_calls is not none %}
57
+ {{- "[TOOL_CALLS] [" }}
58
+ {%- for tool_call in message.tool_calls %}
59
+ {%- set out = tool_call.function|tojson %}
60
+ {{- out[:-1] }}
61
+ {%- if not tool_call.id is defined or tool_call.id|length != 9 %}
62
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
63
+ {%- endif %}
64
+ {{- ', "id": "' + tool_call.id + '"}' }}
65
+ {%- if not loop.last %}
66
+ {{- ", " }}
67
+ {%- else %}
68
+ {{- "]" + eos_token }}
69
+ {%- endif %}
70
+ {%- endfor %}
71
+ {%- elif message["role"] == "assistant" %}
72
+ {{- " " + message["content"]|trim + eos_token}}
73
+ {%- elif message["role"] == "tool_results" or message["role"] == "tool" %}
74
+ {%- if message.content is defined and message.content.content is defined %}
75
+ {%- set content = message.content.content %}
76
+ {%- else %}
77
+ {%- set content = message.content %}
78
+ {%- endif %}
79
+ {{- '[TOOL_RESULTS] {"content": ' + content|string + ", " }}
80
+ {%- if not message.tool_call_id is defined or message.tool_call_id|length != 9 %}
81
+ {{- raise_exception("Tool call IDs should be alphanumeric strings with length 9!") }}
82
+ {%- endif %}
83
+ {{- '"call_id": "' + message.tool_call_id + '"}[/TOOL_RESULTS]' }}
84
+ {%- else %}
85
+ {{- raise_exception("Only user and assistant roles are supported, with the exception of an initial optional system message!") }}
86
+ {%- endif %}
87
+ {%- endfor %}
checkpoint-922/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d25237402eecb22fcce8a7b527f350a3dea69b08a73fb37359db364a0e2692a8
3
+ size 1593
checkpoint-922/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7a8c296296f6fbd072906c578880fb09f4a8720fc9aef9a7ee1b60d9ab6bb9f
3
+ size 14645
checkpoint-922/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c56a513ca46c68f1d30f1c250981246cbe3ec666f9a8922ea6f553f4e3931ea
3
+ size 1465
checkpoint-922/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
checkpoint-922/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
checkpoint-922/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-922/trainer_state.json ADDED
@@ -0,0 +1,862 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 2.0,
6
+ "eval_steps": 500,
7
+ "global_step": 922,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.021719948415122516,
14
+ "grad_norm": 0.0,
15
+ "learning_rate": 0.00019869848156182213,
16
+ "loss": 6.0531,
17
+ "mean_token_accuracy": 0.37881034857127815,
18
+ "num_tokens": 78573.0,
19
+ "step": 10
20
+ },
21
+ {
22
+ "epoch": 0.04343989683024503,
23
+ "grad_norm": 0.0,
24
+ "learning_rate": 0.00019725234996384672,
25
+ "loss": 5.9903,
26
+ "mean_token_accuracy": 0.3887314551044255,
27
+ "num_tokens": 157496.0,
28
+ "step": 20
29
+ },
30
+ {
31
+ "epoch": 0.06515984524536754,
32
+ "grad_norm": 0.0,
33
+ "learning_rate": 0.00019580621836587129,
34
+ "loss": 5.9929,
35
+ "mean_token_accuracy": 0.3850487937917933,
36
+ "num_tokens": 236512.0,
37
+ "step": 30
38
+ },
39
+ {
40
+ "epoch": 0.08687979366049006,
41
+ "grad_norm": 0.0,
42
+ "learning_rate": 0.00019436008676789588,
43
+ "loss": 5.986,
44
+ "mean_token_accuracy": 0.3865805763518438,
45
+ "num_tokens": 315608.0,
46
+ "step": 40
47
+ },
48
+ {
49
+ "epoch": 0.10859974207561257,
50
+ "grad_norm": 0.0,
51
+ "learning_rate": 0.00019291395516992047,
52
+ "loss": 6.0318,
53
+ "mean_token_accuracy": 0.38490854618139564,
54
+ "num_tokens": 394791.0,
55
+ "step": 50
56
+ },
57
+ {
58
+ "epoch": 0.13031969049073508,
59
+ "grad_norm": 0.0,
60
+ "learning_rate": 0.00019146782357194506,
61
+ "loss": 5.9875,
62
+ "mean_token_accuracy": 0.38546840590424836,
63
+ "num_tokens": 473456.0,
64
+ "step": 60
65
+ },
66
+ {
67
+ "epoch": 0.1520396389058576,
68
+ "grad_norm": 0.0,
69
+ "learning_rate": 0.00019002169197396965,
70
+ "loss": 5.9773,
71
+ "mean_token_accuracy": 0.3837947838823311,
72
+ "num_tokens": 552247.0,
73
+ "step": 70
74
+ },
75
+ {
76
+ "epoch": 0.17375958732098012,
77
+ "grad_norm": 0.0,
78
+ "learning_rate": 0.0001885755603759942,
79
+ "loss": 6.0197,
80
+ "mean_token_accuracy": 0.38375892234034836,
81
+ "num_tokens": 631096.0,
82
+ "step": 80
83
+ },
84
+ {
85
+ "epoch": 0.19547953573610263,
86
+ "grad_norm": 0.0,
87
+ "learning_rate": 0.0001871294287780188,
88
+ "loss": 6.0014,
89
+ "mean_token_accuracy": 0.38519658900331705,
90
+ "num_tokens": 710275.0,
91
+ "step": 90
92
+ },
93
+ {
94
+ "epoch": 0.21719948415122514,
95
+ "grad_norm": 0.0,
96
+ "learning_rate": 0.0001856832971800434,
97
+ "loss": 6.0217,
98
+ "mean_token_accuracy": 0.3826989881345071,
99
+ "num_tokens": 789174.0,
100
+ "step": 100
101
+ },
102
+ {
103
+ "epoch": 0.23891943256634765,
104
+ "grad_norm": 0.0,
105
+ "learning_rate": 0.00018423716558206799,
106
+ "loss": 6.0294,
107
+ "mean_token_accuracy": 0.3842650496866554,
108
+ "num_tokens": 867922.0,
109
+ "step": 110
110
+ },
111
+ {
112
+ "epoch": 0.26063938098147016,
113
+ "grad_norm": 0.0,
114
+ "learning_rate": 0.00018279103398409255,
115
+ "loss": 5.9548,
116
+ "mean_token_accuracy": 0.3875111517496407,
117
+ "num_tokens": 947073.0,
118
+ "step": 120
119
+ },
120
+ {
121
+ "epoch": 0.2823593293965927,
122
+ "grad_norm": 0.0,
123
+ "learning_rate": 0.00018134490238611714,
124
+ "loss": 6.0071,
125
+ "mean_token_accuracy": 0.38376112943515184,
126
+ "num_tokens": 1026211.0,
127
+ "step": 130
128
+ },
129
+ {
130
+ "epoch": 0.3040792778117152,
131
+ "grad_norm": 0.0,
132
+ "learning_rate": 0.00017989877078814173,
133
+ "loss": 6.0105,
134
+ "mean_token_accuracy": 0.38463024909142407,
135
+ "num_tokens": 1105147.0,
136
+ "step": 140
137
+ },
138
+ {
139
+ "epoch": 0.3257992262268377,
140
+ "grad_norm": 0.0,
141
+ "learning_rate": 0.0001784526391901663,
142
+ "loss": 6.0449,
143
+ "mean_token_accuracy": 0.38241248747799544,
144
+ "num_tokens": 1183837.0,
145
+ "step": 150
146
+ },
147
+ {
148
+ "epoch": 0.34751917464196025,
149
+ "grad_norm": 0.0,
150
+ "learning_rate": 0.00017700650759219091,
151
+ "loss": 5.9946,
152
+ "mean_token_accuracy": 0.3821917780907825,
153
+ "num_tokens": 1262676.0,
154
+ "step": 160
155
+ },
156
+ {
157
+ "epoch": 0.36923912305708273,
158
+ "grad_norm": 0.0,
159
+ "learning_rate": 0.00017556037599421548,
160
+ "loss": 6.0049,
161
+ "mean_token_accuracy": 0.3842250820598565,
162
+ "num_tokens": 1341759.0,
163
+ "step": 170
164
+ },
165
+ {
166
+ "epoch": 0.39095907147220527,
167
+ "grad_norm": 0.0,
168
+ "learning_rate": 0.00017411424439624007,
169
+ "loss": 5.9736,
170
+ "mean_token_accuracy": 0.3861917880363762,
171
+ "num_tokens": 1420987.0,
172
+ "step": 180
173
+ },
174
+ {
175
+ "epoch": 0.41267901988732775,
176
+ "grad_norm": 0.0,
177
+ "learning_rate": 0.00017266811279826466,
178
+ "loss": 6.0431,
179
+ "mean_token_accuracy": 0.38146835477091373,
180
+ "num_tokens": 1499903.0,
181
+ "step": 190
182
+ },
183
+ {
184
+ "epoch": 0.4343989683024503,
185
+ "grad_norm": 0.0,
186
+ "learning_rate": 0.00017122198120028922,
187
+ "loss": 6.0069,
188
+ "mean_token_accuracy": 0.3861038032686338,
189
+ "num_tokens": 1578880.0,
190
+ "step": 200
191
+ },
192
+ {
193
+ "epoch": 0.4561189167175728,
194
+ "grad_norm": 0.0,
195
+ "learning_rate": 0.00016977584960231381,
196
+ "loss": 6.0091,
197
+ "mean_token_accuracy": 0.3845075036631897,
198
+ "num_tokens": 1657709.0,
199
+ "step": 210
200
+ },
201
+ {
202
+ "epoch": 0.4778388651326953,
203
+ "grad_norm": 0.0,
204
+ "learning_rate": 0.0001683297180043384,
205
+ "loss": 6.0029,
206
+ "mean_token_accuracy": 0.38288453239947556,
207
+ "num_tokens": 1736620.0,
208
+ "step": 220
209
+ },
210
+ {
211
+ "epoch": 0.49955881354781784,
212
+ "grad_norm": 0.0,
213
+ "learning_rate": 0.000166883586406363,
214
+ "loss": 5.9933,
215
+ "mean_token_accuracy": 0.38531269936356694,
216
+ "num_tokens": 1815478.0,
217
+ "step": 230
218
+ },
219
+ {
220
+ "epoch": 0.5212787619629403,
221
+ "grad_norm": 0.0,
222
+ "learning_rate": 0.00016543745480838756,
223
+ "loss": 6.0334,
224
+ "mean_token_accuracy": 0.3833939728909172,
225
+ "num_tokens": 1894256.0,
226
+ "step": 240
227
+ },
228
+ {
229
+ "epoch": 0.5429987103780628,
230
+ "grad_norm": 0.0,
231
+ "learning_rate": 0.00016399132321041215,
232
+ "loss": 6.0484,
233
+ "mean_token_accuracy": 0.3841994108865038,
234
+ "num_tokens": 1972997.0,
235
+ "step": 250
236
+ },
237
+ {
238
+ "epoch": 0.5647186587931854,
239
+ "grad_norm": 0.0,
240
+ "learning_rate": 0.00016254519161243674,
241
+ "loss": 6.0285,
242
+ "mean_token_accuracy": 0.38443261398933826,
243
+ "num_tokens": 2051890.0,
244
+ "step": 260
245
+ },
246
+ {
247
+ "epoch": 0.5864386072083079,
248
+ "grad_norm": 0.0,
249
+ "learning_rate": 0.00016109906001446133,
250
+ "loss": 5.983,
251
+ "mean_token_accuracy": 0.385862308065407,
252
+ "num_tokens": 2130746.0,
253
+ "step": 270
254
+ },
255
+ {
256
+ "epoch": 0.6081585556234304,
257
+ "grad_norm": 0.0,
258
+ "learning_rate": 0.00015965292841648592,
259
+ "loss": 6.0046,
260
+ "mean_token_accuracy": 0.3859730801777914,
261
+ "num_tokens": 2209603.0,
262
+ "step": 280
263
+ },
264
+ {
265
+ "epoch": 0.6298785040385529,
266
+ "grad_norm": 0.0,
267
+ "learning_rate": 0.0001582067968185105,
268
+ "loss": 5.9887,
269
+ "mean_token_accuracy": 0.38499277490191164,
270
+ "num_tokens": 2288743.0,
271
+ "step": 290
272
+ },
273
+ {
274
+ "epoch": 0.6515984524536754,
275
+ "grad_norm": 0.0,
276
+ "learning_rate": 0.00015676066522053508,
277
+ "loss": 5.9788,
278
+ "mean_token_accuracy": 0.3864025830756873,
279
+ "num_tokens": 2367648.0,
280
+ "step": 300
281
+ },
282
+ {
283
+ "epoch": 0.6733184008687979,
284
+ "grad_norm": 0.0,
285
+ "learning_rate": 0.00015531453362255964,
286
+ "loss": 6.0162,
287
+ "mean_token_accuracy": 0.38221894631860776,
288
+ "num_tokens": 2446256.0,
289
+ "step": 310
290
+ },
291
+ {
292
+ "epoch": 0.6950383492839205,
293
+ "grad_norm": 0.0,
294
+ "learning_rate": 0.00015386840202458423,
295
+ "loss": 6.0017,
296
+ "mean_token_accuracy": 0.38538019855041056,
297
+ "num_tokens": 2525065.0,
298
+ "step": 320
299
+ },
300
+ {
301
+ "epoch": 0.716758297699043,
302
+ "grad_norm": 0.0,
303
+ "learning_rate": 0.00015242227042660883,
304
+ "loss": 5.9569,
305
+ "mean_token_accuracy": 0.3852926092222333,
306
+ "num_tokens": 2604183.0,
307
+ "step": 330
308
+ },
309
+ {
310
+ "epoch": 0.7384782461141655,
311
+ "grad_norm": 0.0,
312
+ "learning_rate": 0.00015097613882863342,
313
+ "loss": 6.0072,
314
+ "mean_token_accuracy": 0.38616998645011336,
315
+ "num_tokens": 2683084.0,
316
+ "step": 340
317
+ },
318
+ {
319
+ "epoch": 0.7601981945292879,
320
+ "grad_norm": 0.0,
321
+ "learning_rate": 0.000149530007230658,
322
+ "loss": 6.0428,
323
+ "mean_token_accuracy": 0.3831095602363348,
324
+ "num_tokens": 2761824.0,
325
+ "step": 350
326
+ },
327
+ {
328
+ "epoch": 0.7819181429444105,
329
+ "grad_norm": 0.0,
330
+ "learning_rate": 0.00014808387563268257,
331
+ "loss": 5.9781,
332
+ "mean_token_accuracy": 0.3861894382047467,
333
+ "num_tokens": 2840709.0,
334
+ "step": 360
335
+ },
336
+ {
337
+ "epoch": 0.803638091359533,
338
+ "grad_norm": 0.0,
339
+ "learning_rate": 0.00014663774403470716,
340
+ "loss": 5.9898,
341
+ "mean_token_accuracy": 0.3846066597965546,
342
+ "num_tokens": 2919385.0,
343
+ "step": 370
344
+ },
345
+ {
346
+ "epoch": 0.8253580397746555,
347
+ "grad_norm": 0.0,
348
+ "learning_rate": 0.00014519161243673173,
349
+ "loss": 6.0098,
350
+ "mean_token_accuracy": 0.38357423364650456,
351
+ "num_tokens": 2998308.0,
352
+ "step": 380
353
+ },
354
+ {
355
+ "epoch": 0.8470779881897781,
356
+ "grad_norm": 0.0,
357
+ "learning_rate": 0.00014374548083875634,
358
+ "loss": 6.0349,
359
+ "mean_token_accuracy": 0.3809299209038727,
360
+ "num_tokens": 3076962.0,
361
+ "step": 390
362
+ },
363
+ {
364
+ "epoch": 0.8687979366049006,
365
+ "grad_norm": 0.0,
366
+ "learning_rate": 0.0001422993492407809,
367
+ "loss": 6.0294,
368
+ "mean_token_accuracy": 0.38217092433478683,
369
+ "num_tokens": 3155983.0,
370
+ "step": 400
371
+ },
372
+ {
373
+ "epoch": 0.890517885020023,
374
+ "grad_norm": 0.0,
375
+ "learning_rate": 0.0001408532176428055,
376
+ "loss": 5.9728,
377
+ "mean_token_accuracy": 0.3870704318396747,
378
+ "num_tokens": 3234960.0,
379
+ "step": 410
380
+ },
381
+ {
382
+ "epoch": 0.9122378334351456,
383
+ "grad_norm": 0.0,
384
+ "learning_rate": 0.0001394070860448301,
385
+ "loss": 6.0541,
386
+ "mean_token_accuracy": 0.38209462116938087,
387
+ "num_tokens": 3313797.0,
388
+ "step": 420
389
+ },
390
+ {
391
+ "epoch": 0.9339577818502681,
392
+ "grad_norm": 0.0,
393
+ "learning_rate": 0.00013796095444685465,
394
+ "loss": 5.9871,
395
+ "mean_token_accuracy": 0.38401242352556436,
396
+ "num_tokens": 3392880.0,
397
+ "step": 430
398
+ },
399
+ {
400
+ "epoch": 0.9556777302653906,
401
+ "grad_norm": 0.0,
402
+ "learning_rate": 0.00013651482284887927,
403
+ "loss": 5.9901,
404
+ "mean_token_accuracy": 0.38375150066567587,
405
+ "num_tokens": 3471597.0,
406
+ "step": 440
407
+ },
408
+ {
409
+ "epoch": 0.9773976786805131,
410
+ "grad_norm": 0.0,
411
+ "learning_rate": 0.00013506869125090384,
412
+ "loss": 5.9797,
413
+ "mean_token_accuracy": 0.38728573790285736,
414
+ "num_tokens": 3550697.0,
415
+ "step": 450
416
+ },
417
+ {
418
+ "epoch": 0.9991176270956357,
419
+ "grad_norm": 0.0,
420
+ "learning_rate": 0.00013362255965292843,
421
+ "loss": 5.9971,
422
+ "mean_token_accuracy": 0.38420643559657036,
423
+ "num_tokens": 3629481.0,
424
+ "step": 460
425
+ },
426
+ {
427
+ "epoch": 1.0195479535736103,
428
+ "grad_norm": 0.0,
429
+ "learning_rate": 0.00013217642805495302,
430
+ "loss": 6.0223,
431
+ "mean_token_accuracy": 0.3833757105509308,
432
+ "num_tokens": 3703746.0,
433
+ "step": 470
434
+ },
435
+ {
436
+ "epoch": 1.0412679019887328,
437
+ "grad_norm": 0.0,
438
+ "learning_rate": 0.00013073029645697758,
439
+ "loss": 6.0146,
440
+ "mean_token_accuracy": 0.38417146790307016,
441
+ "num_tokens": 3782783.0,
442
+ "step": 480
443
+ },
444
+ {
445
+ "epoch": 1.0629878504038552,
446
+ "grad_norm": 0.0,
447
+ "learning_rate": 0.00012928416485900217,
448
+ "loss": 5.9713,
449
+ "mean_token_accuracy": 0.3863212176831439,
450
+ "num_tokens": 3862048.0,
451
+ "step": 490
452
+ },
453
+ {
454
+ "epoch": 1.0847077988189777,
455
+ "grad_norm": 0.0,
456
+ "learning_rate": 0.00012783803326102676,
457
+ "loss": 5.9944,
458
+ "mean_token_accuracy": 0.38522618702845646,
459
+ "num_tokens": 3941220.0,
460
+ "step": 500
461
+ },
462
+ {
463
+ "epoch": 1.1064277472341004,
464
+ "grad_norm": 0.0,
465
+ "learning_rate": 0.00012639190166305135,
466
+ "loss": 5.9895,
467
+ "mean_token_accuracy": 0.38548907284857703,
468
+ "num_tokens": 4020089.0,
469
+ "step": 510
470
+ },
471
+ {
472
+ "epoch": 1.128147695649223,
473
+ "grad_norm": 0.0,
474
+ "learning_rate": 0.00012494577006507592,
475
+ "loss": 5.9901,
476
+ "mean_token_accuracy": 0.3841517207212746,
477
+ "num_tokens": 4098981.0,
478
+ "step": 520
479
+ },
480
+ {
481
+ "epoch": 1.1498676440643454,
482
+ "grad_norm": 0.0,
483
+ "learning_rate": 0.0001234996384671005,
484
+ "loss": 6.0125,
485
+ "mean_token_accuracy": 0.38393973018974065,
486
+ "num_tokens": 4177983.0,
487
+ "step": 530
488
+ },
489
+ {
490
+ "epoch": 1.1715875924794679,
491
+ "grad_norm": 0.0,
492
+ "learning_rate": 0.00012205350686912509,
493
+ "loss": 5.9999,
494
+ "mean_token_accuracy": 0.3848019931232557,
495
+ "num_tokens": 4256818.0,
496
+ "step": 540
497
+ },
498
+ {
499
+ "epoch": 1.1933075408945903,
500
+ "grad_norm": 0.0,
501
+ "learning_rate": 0.00012060737527114966,
502
+ "loss": 6.0146,
503
+ "mean_token_accuracy": 0.3834400905063376,
504
+ "num_tokens": 4335754.0,
505
+ "step": 550
506
+ },
507
+ {
508
+ "epoch": 1.2150274893097128,
509
+ "grad_norm": 0.0,
510
+ "learning_rate": 0.00011916124367317427,
511
+ "loss": 5.9832,
512
+ "mean_token_accuracy": 0.38411646473687144,
513
+ "num_tokens": 4414480.0,
514
+ "step": 560
515
+ },
516
+ {
517
+ "epoch": 1.2367474377248353,
518
+ "grad_norm": 0.0,
519
+ "learning_rate": 0.00011771511207519885,
520
+ "loss": 6.0393,
521
+ "mean_token_accuracy": 0.3809544609161094,
522
+ "num_tokens": 4493321.0,
523
+ "step": 570
524
+ },
525
+ {
526
+ "epoch": 1.2584673861399578,
527
+ "grad_norm": 0.0,
528
+ "learning_rate": 0.00011626898047722344,
529
+ "loss": 5.9989,
530
+ "mean_token_accuracy": 0.38565440035890786,
531
+ "num_tokens": 4572297.0,
532
+ "step": 580
533
+ },
534
+ {
535
+ "epoch": 1.2801873345550805,
536
+ "grad_norm": 0.0,
537
+ "learning_rate": 0.00011482284887924801,
538
+ "loss": 5.9983,
539
+ "mean_token_accuracy": 0.3863610655302182,
540
+ "num_tokens": 4651247.0,
541
+ "step": 590
542
+ },
543
+ {
544
+ "epoch": 1.301907282970203,
545
+ "grad_norm": 0.0,
546
+ "learning_rate": 0.00011337671728127259,
547
+ "loss": 6.0116,
548
+ "mean_token_accuracy": 0.3831069741398096,
549
+ "num_tokens": 4730012.0,
550
+ "step": 600
551
+ },
552
+ {
553
+ "epoch": 1.3236272313853255,
554
+ "grad_norm": 0.0,
555
+ "learning_rate": 0.0001119305856832972,
556
+ "loss": 6.0195,
557
+ "mean_token_accuracy": 0.3857766560278833,
558
+ "num_tokens": 4809013.0,
559
+ "step": 610
560
+ },
561
+ {
562
+ "epoch": 1.345347179800448,
563
+ "grad_norm": 0.0,
564
+ "learning_rate": 0.00011048445408532177,
565
+ "loss": 6.0108,
566
+ "mean_token_accuracy": 0.3835394912166521,
567
+ "num_tokens": 4888027.0,
568
+ "step": 620
569
+ },
570
+ {
571
+ "epoch": 1.3670671282155704,
572
+ "grad_norm": 0.0,
573
+ "learning_rate": 0.00010903832248734635,
574
+ "loss": 5.9558,
575
+ "mean_token_accuracy": 0.3883327366434969,
576
+ "num_tokens": 4966859.0,
577
+ "step": 630
578
+ },
579
+ {
580
+ "epoch": 1.3887870766306931,
581
+ "grad_norm": 0.0,
582
+ "learning_rate": 0.00010759219088937094,
583
+ "loss": 6.0088,
584
+ "mean_token_accuracy": 0.385754154715687,
585
+ "num_tokens": 5045905.0,
586
+ "step": 640
587
+ },
588
+ {
589
+ "epoch": 1.4105070250458156,
590
+ "grad_norm": 0.0,
591
+ "learning_rate": 0.00010614605929139552,
592
+ "loss": 6.0236,
593
+ "mean_token_accuracy": 0.3829167880234309,
594
+ "num_tokens": 5124413.0,
595
+ "step": 650
596
+ },
597
+ {
598
+ "epoch": 1.432226973460938,
599
+ "grad_norm": 0.0,
600
+ "learning_rate": 0.0001046999276934201,
601
+ "loss": 6.0154,
602
+ "mean_token_accuracy": 0.3834054367849603,
603
+ "num_tokens": 5203273.0,
604
+ "step": 660
605
+ },
606
+ {
607
+ "epoch": 1.4539469218760606,
608
+ "grad_norm": 0.0,
609
+ "learning_rate": 0.0001032537960954447,
610
+ "loss": 6.005,
611
+ "mean_token_accuracy": 0.3848568681394681,
612
+ "num_tokens": 5281984.0,
613
+ "step": 670
614
+ },
615
+ {
616
+ "epoch": 1.475666870291183,
617
+ "grad_norm": 0.0,
618
+ "learning_rate": 0.00010180766449746928,
619
+ "loss": 5.9805,
620
+ "mean_token_accuracy": 0.3862292483681813,
621
+ "num_tokens": 5360846.0,
622
+ "step": 680
623
+ },
624
+ {
625
+ "epoch": 1.4973868187063055,
626
+ "grad_norm": 0.0,
627
+ "learning_rate": 0.00010036153289949386,
628
+ "loss": 5.9871,
629
+ "mean_token_accuracy": 0.3851381171494722,
630
+ "num_tokens": 5439758.0,
631
+ "step": 690
632
+ },
633
+ {
634
+ "epoch": 1.519106767121428,
635
+ "grad_norm": 0.0,
636
+ "learning_rate": 9.891540130151843e-05,
637
+ "loss": 6.0254,
638
+ "mean_token_accuracy": 0.3831110308994539,
639
+ "num_tokens": 5518768.0,
640
+ "step": 700
641
+ },
642
+ {
643
+ "epoch": 1.5408267155365505,
644
+ "grad_norm": 0.0,
645
+ "learning_rate": 9.746926970354303e-05,
646
+ "loss": 6.0339,
647
+ "mean_token_accuracy": 0.3808896674308926,
648
+ "num_tokens": 5597672.0,
649
+ "step": 710
650
+ },
651
+ {
652
+ "epoch": 1.562546663951673,
653
+ "grad_norm": 0.0,
654
+ "learning_rate": 9.602313810556762e-05,
655
+ "loss": 5.994,
656
+ "mean_token_accuracy": 0.3865658549708314,
657
+ "num_tokens": 5676393.0,
658
+ "step": 720
659
+ },
660
+ {
661
+ "epoch": 1.5842666123667957,
662
+ "grad_norm": 0.0,
663
+ "learning_rate": 9.45770065075922e-05,
664
+ "loss": 6.0026,
665
+ "mean_token_accuracy": 0.38676879862323404,
666
+ "num_tokens": 5755416.0,
667
+ "step": 730
668
+ },
669
+ {
670
+ "epoch": 1.6059865607819181,
671
+ "grad_norm": 0.0,
672
+ "learning_rate": 9.313087490961678e-05,
673
+ "loss": 6.0175,
674
+ "mean_token_accuracy": 0.38389978763880206,
675
+ "num_tokens": 5834322.0,
676
+ "step": 740
677
+ },
678
+ {
679
+ "epoch": 1.6277065091970406,
680
+ "grad_norm": 0.0,
681
+ "learning_rate": 9.168474331164136e-05,
682
+ "loss": 6.0224,
683
+ "mean_token_accuracy": 0.3840769061935134,
684
+ "num_tokens": 5912925.0,
685
+ "step": 750
686
+ },
687
+ {
688
+ "epoch": 1.6494264576121633,
689
+ "grad_norm": 0.0,
690
+ "learning_rate": 9.023861171366594e-05,
691
+ "loss": 5.9683,
692
+ "mean_token_accuracy": 0.38740592543035746,
693
+ "num_tokens": 5991913.0,
694
+ "step": 760
695
+ },
696
+ {
697
+ "epoch": 1.6711464060272858,
698
+ "grad_norm": 0.0,
699
+ "learning_rate": 8.879248011569053e-05,
700
+ "loss": 6.0136,
701
+ "mean_token_accuracy": 0.3827170055825263,
702
+ "num_tokens": 6070546.0,
703
+ "step": 770
704
+ },
705
+ {
706
+ "epoch": 1.6928663544424083,
707
+ "grad_norm": 0.0,
708
+ "learning_rate": 8.734634851771512e-05,
709
+ "loss": 5.9688,
710
+ "mean_token_accuracy": 0.3871995336958207,
711
+ "num_tokens": 6149623.0,
712
+ "step": 780
713
+ },
714
+ {
715
+ "epoch": 1.7145863028575308,
716
+ "grad_norm": 0.0,
717
+ "learning_rate": 8.59002169197397e-05,
718
+ "loss": 6.0639,
719
+ "mean_token_accuracy": 0.3800558194401674,
720
+ "num_tokens": 6228396.0,
721
+ "step": 790
722
+ },
723
+ {
724
+ "epoch": 1.7363062512726533,
725
+ "grad_norm": 0.0,
726
+ "learning_rate": 8.445408532176429e-05,
727
+ "loss": 5.9778,
728
+ "mean_token_accuracy": 0.3854194077430293,
729
+ "num_tokens": 6307375.0,
730
+ "step": 800
731
+ },
732
+ {
733
+ "epoch": 1.7580261996877757,
734
+ "grad_norm": 0.0,
735
+ "learning_rate": 8.300795372378887e-05,
736
+ "loss": 6.0294,
737
+ "mean_token_accuracy": 0.38350161886774004,
738
+ "num_tokens": 6386536.0,
739
+ "step": 810
740
+ },
741
+ {
742
+ "epoch": 1.7797461481028982,
743
+ "grad_norm": 0.0,
744
+ "learning_rate": 8.156182212581345e-05,
745
+ "loss": 5.9736,
746
+ "mean_token_accuracy": 0.38414000715129076,
747
+ "num_tokens": 6465708.0,
748
+ "step": 820
749
+ },
750
+ {
751
+ "epoch": 1.8014660965180207,
752
+ "grad_norm": 0.0,
753
+ "learning_rate": 8.011569052783804e-05,
754
+ "loss": 5.9847,
755
+ "mean_token_accuracy": 0.3877883184002712,
756
+ "num_tokens": 6544782.0,
757
+ "step": 830
758
+ },
759
+ {
760
+ "epoch": 1.8231860449331432,
761
+ "grad_norm": 0.0,
762
+ "learning_rate": 7.866955892986261e-05,
763
+ "loss": 6.043,
764
+ "mean_token_accuracy": 0.38251175949117167,
765
+ "num_tokens": 6623497.0,
766
+ "step": 840
767
+ },
768
+ {
769
+ "epoch": 1.8449059933482657,
770
+ "grad_norm": 0.0,
771
+ "learning_rate": 7.72234273318872e-05,
772
+ "loss": 6.0297,
773
+ "mean_token_accuracy": 0.38357739897910503,
774
+ "num_tokens": 6702159.0,
775
+ "step": 850
776
+ },
777
+ {
778
+ "epoch": 1.8666259417633884,
779
+ "grad_norm": 0.0,
780
+ "learning_rate": 7.57772957339118e-05,
781
+ "loss": 6.0335,
782
+ "mean_token_accuracy": 0.38290644709486515,
783
+ "num_tokens": 6780886.0,
784
+ "step": 860
785
+ },
786
+ {
787
+ "epoch": 1.8883458901785108,
788
+ "grad_norm": 0.0,
789
+ "learning_rate": 7.433116413593637e-05,
790
+ "loss": 6.0273,
791
+ "mean_token_accuracy": 0.38307004772359504,
792
+ "num_tokens": 6859533.0,
793
+ "step": 870
794
+ },
795
+ {
796
+ "epoch": 1.9100658385936333,
797
+ "grad_norm": 0.0,
798
+ "learning_rate": 7.288503253796096e-05,
799
+ "loss": 5.9827,
800
+ "mean_token_accuracy": 0.385389854805544,
801
+ "num_tokens": 6938490.0,
802
+ "step": 880
803
+ },
804
+ {
805
+ "epoch": 1.9317857870087558,
806
+ "grad_norm": 0.0,
807
+ "learning_rate": 7.143890093998554e-05,
808
+ "loss": 6.0212,
809
+ "mean_token_accuracy": 0.382453205762431,
810
+ "num_tokens": 7017297.0,
811
+ "step": 890
812
+ },
813
+ {
814
+ "epoch": 1.9535057354238785,
815
+ "grad_norm": 0.0,
816
+ "learning_rate": 6.999276934201012e-05,
817
+ "loss": 6.0043,
818
+ "mean_token_accuracy": 0.38360455771908164,
819
+ "num_tokens": 7096419.0,
820
+ "step": 900
821
+ },
822
+ {
823
+ "epoch": 1.975225683839001,
824
+ "grad_norm": 0.0,
825
+ "learning_rate": 6.854663774403471e-05,
826
+ "loss": 5.985,
827
+ "mean_token_accuracy": 0.38632277538999915,
828
+ "num_tokens": 7175650.0,
829
+ "step": 910
830
+ },
831
+ {
832
+ "epoch": 1.9969456322541235,
833
+ "grad_norm": 0.0,
834
+ "learning_rate": 6.71005061460593e-05,
835
+ "loss": 6.009,
836
+ "mean_token_accuracy": 0.38311313565354793,
837
+ "num_tokens": 7254378.0,
838
+ "step": 920
839
+ }
840
+ ],
841
+ "logging_steps": 10,
842
+ "max_steps": 1383,
843
+ "num_input_tokens_seen": 0,
844
+ "num_train_epochs": 3,
845
+ "save_steps": 500,
846
+ "stateful_callbacks": {
847
+ "TrainerControl": {
848
+ "args": {
849
+ "should_epoch_stop": false,
850
+ "should_evaluate": false,
851
+ "should_log": false,
852
+ "should_save": true,
853
+ "should_training_stop": false
854
+ },
855
+ "attributes": {}
856
+ }
857
+ },
858
+ "total_flos": 3.107018416918856e+17,
859
+ "train_batch_size": 1,
860
+ "trial_name": null,
861
+ "trial_params": null
862
+ }
checkpoint-922/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131a0a21414525756ac7e71774e33b16bf1fd48bed5fd2484806044f425159ba
3
+ size 6161
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:131a0a21414525756ac7e71774e33b16bf1fd48bed5fd2484806044f425159ba
3
+ size 6161