hazarsozer commited on
Commit
2d6f639
·
verified ·
1 Parent(s): 7c85bb3

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +63 -0
  2. adapter_config.json +50 -0
  3. adapter_model.safetensors +3 -0
  4. checkpoint-1000/README.md +210 -0
  5. checkpoint-1000/adapter_config.json +50 -0
  6. checkpoint-1000/adapter_model.safetensors +3 -0
  7. checkpoint-1000/optimizer.pt +3 -0
  8. checkpoint-1000/rng_state.pth +3 -0
  9. checkpoint-1000/scheduler.pt +3 -0
  10. checkpoint-1000/special_tokens_map.json +30 -0
  11. checkpoint-1000/tokenizer.json +0 -0
  12. checkpoint-1000/tokenizer.model +3 -0
  13. checkpoint-1000/tokenizer_config.json +0 -0
  14. checkpoint-1000/trainer_state.json +734 -0
  15. checkpoint-1000/training_args.bin +3 -0
  16. checkpoint-1500/README.md +210 -0
  17. checkpoint-1500/adapter_config.json +50 -0
  18. checkpoint-1500/adapter_model.safetensors +3 -0
  19. checkpoint-1500/optimizer.pt +3 -0
  20. checkpoint-1500/rng_state.pth +3 -0
  21. checkpoint-1500/scheduler.pt +3 -0
  22. checkpoint-1500/special_tokens_map.json +30 -0
  23. checkpoint-1500/tokenizer.json +0 -0
  24. checkpoint-1500/tokenizer.model +3 -0
  25. checkpoint-1500/tokenizer_config.json +0 -0
  26. checkpoint-1500/trainer_state.json +1084 -0
  27. checkpoint-1500/training_args.bin +3 -0
  28. checkpoint-1545/README.md +210 -0
  29. checkpoint-1545/adapter_config.json +50 -0
  30. checkpoint-1545/adapter_model.safetensors +3 -0
  31. checkpoint-1545/optimizer.pt +3 -0
  32. checkpoint-1545/rng_state.pth +3 -0
  33. checkpoint-1545/scheduler.pt +3 -0
  34. checkpoint-1545/special_tokens_map.json +30 -0
  35. checkpoint-1545/tokenizer.json +0 -0
  36. checkpoint-1545/tokenizer.model +3 -0
  37. checkpoint-1545/tokenizer_config.json +0 -0
  38. checkpoint-1545/trainer_state.json +1112 -0
  39. checkpoint-1545/training_args.bin +3 -0
  40. checkpoint-500/README.md +210 -0
  41. checkpoint-500/adapter_config.json +50 -0
  42. checkpoint-500/adapter_model.safetensors +3 -0
  43. checkpoint-500/optimizer.pt +3 -0
  44. checkpoint-500/rng_state.pth +3 -0
  45. checkpoint-500/scheduler.pt +3 -0
  46. checkpoint-500/special_tokens_map.json +30 -0
  47. checkpoint-500/tokenizer.json +0 -0
  48. checkpoint-500/tokenizer.model +3 -0
  49. checkpoint-500/tokenizer_config.json +0 -0
  50. checkpoint-500/trainer_state.json +384 -0
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/mistral-7b-v0.3-bnb-4bit
3
+ library_name: peft
4
+ model_name: mistral_qlora
5
+ tags:
6
+ - base_model:adapter:unsloth/mistral-7b-v0.3-bnb-4bit
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ - unsloth
12
+ licence: license
13
+ pipeline_tag: text-generation
14
+ ---
15
+
16
+ # Model Card for mistral_qlora
17
+
18
+ This model is a fine-tuned version of [unsloth/mistral-7b-v0.3-bnb-4bit](https://huggingface.co/unsloth/mistral-7b-v0.3-bnb-4bit).
19
+ It has been trained using [TRL](https://github.com/huggingface/trl).
20
+
21
+ ## Quick start
22
+
23
+ ```python
24
+ from transformers import pipeline
25
+
26
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
27
+ generator = pipeline("text-generation", model="None", device="cuda")
28
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
29
+ print(output["generated_text"])
30
+ ```
31
+
32
+ ## Training procedure
33
+
34
+
35
+
36
+
37
+ This model was trained with SFT.
38
+
39
+ ### Framework versions
40
+
41
+ - PEFT 0.18.0
42
+ - TRL: 0.24.0
43
+ - Transformers: 4.57.3
44
+ - Pytorch: 2.6.0+cu124
45
+ - Datasets: 4.3.0
46
+ - Tokenizers: 0.22.1
47
+
48
+ ## Citations
49
+
50
+
51
+
52
+ Cite TRL as:
53
+
54
+ ```bibtex
55
+ @misc{vonwerra2022trl,
56
+ title = {{TRL: Transformer Reinforcement Learning}},
57
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
58
+ year = 2020,
59
+ journal = {GitHub repository},
60
+ publisher = {GitHub},
61
+ howpublished = {\url{https://github.com/huggingface/trl}}
62
+ }
63
+ ```
adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "MistralForCausalLM",
7
+ "parent_library": "transformers.models.mistral.modeling_mistral",
8
+ "unsloth_fixed": true
9
+ },
10
+ "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
11
+ "bias": "none",
12
+ "corda_config": null,
13
+ "ensure_weight_tying": false,
14
+ "eva_config": null,
15
+ "exclude_modules": null,
16
+ "fan_in_fan_out": false,
17
+ "inference_mode": true,
18
+ "init_lora_weights": true,
19
+ "layer_replication": null,
20
+ "layers_pattern": null,
21
+ "layers_to_transform": null,
22
+ "loftq_config": {},
23
+ "lora_alpha": 32,
24
+ "lora_bias": false,
25
+ "lora_dropout": 0,
26
+ "megatron_config": null,
27
+ "megatron_core": "megatron.core",
28
+ "modules_to_save": null,
29
+ "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
+ "qalora_group_size": 16,
32
+ "r": 16,
33
+ "rank_pattern": {},
34
+ "revision": null,
35
+ "target_modules": [
36
+ "k_proj",
37
+ "gate_proj",
38
+ "o_proj",
39
+ "v_proj",
40
+ "q_proj",
41
+ "down_proj",
42
+ "up_proj"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": "CAUSAL_LM",
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a03703be3951badd986bde85e958c3be749b7cc1f8e0ddf6c2a7dac0e8060a8b
3
+ size 167832240
checkpoint-1000/README.md ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/mistral-7b-v0.3-bnb-4bit
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:unsloth/mistral-7b-v0.3-bnb-4bit
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ - unsloth
12
+ ---
13
+
14
+ # Model Card for Model ID
15
+
16
+ <!-- Provide a quick summary of what the model is/does. -->
17
+
18
+
19
+
20
+ ## Model Details
21
+
22
+ ### Model Description
23
+
24
+ <!-- Provide a longer summary of what this model is. -->
25
+
26
+
27
+
28
+ - **Developed by:** [More Information Needed]
29
+ - **Funded by [optional]:** [More Information Needed]
30
+ - **Shared by [optional]:** [More Information Needed]
31
+ - **Model type:** [More Information Needed]
32
+ - **Language(s) (NLP):** [More Information Needed]
33
+ - **License:** [More Information Needed]
34
+ - **Finetuned from model [optional]:** [More Information Needed]
35
+
36
+ ### Model Sources [optional]
37
+
38
+ <!-- Provide the basic links for the model. -->
39
+
40
+ - **Repository:** [More Information Needed]
41
+ - **Paper [optional]:** [More Information Needed]
42
+ - **Demo [optional]:** [More Information Needed]
43
+
44
+ ## Uses
45
+
46
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
47
+
48
+ ### Direct Use
49
+
50
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
51
+
52
+ [More Information Needed]
53
+
54
+ ### Downstream Use [optional]
55
+
56
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
57
+
58
+ [More Information Needed]
59
+
60
+ ### Out-of-Scope Use
61
+
62
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
63
+
64
+ [More Information Needed]
65
+
66
+ ## Bias, Risks, and Limitations
67
+
68
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
69
+
70
+ [More Information Needed]
71
+
72
+ ### Recommendations
73
+
74
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
75
+
76
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
77
+
78
+ ## How to Get Started with the Model
79
+
80
+ Use the code below to get started with the model.
81
+
82
+ [More Information Needed]
83
+
84
+ ## Training Details
85
+
86
+ ### Training Data
87
+
88
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
89
+
90
+ [More Information Needed]
91
+
92
+ ### Training Procedure
93
+
94
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
95
+
96
+ #### Preprocessing [optional]
97
+
98
+ [More Information Needed]
99
+
100
+
101
+ #### Training Hyperparameters
102
+
103
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
104
+
105
+ #### Speeds, Sizes, Times [optional]
106
+
107
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
108
+
109
+ [More Information Needed]
110
+
111
+ ## Evaluation
112
+
113
+ <!-- This section describes the evaluation protocols and provides the results. -->
114
+
115
+ ### Testing Data, Factors & Metrics
116
+
117
+ #### Testing Data
118
+
119
+ <!-- This should link to a Dataset Card if possible. -->
120
+
121
+ [More Information Needed]
122
+
123
+ #### Factors
124
+
125
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
126
+
127
+ [More Information Needed]
128
+
129
+ #### Metrics
130
+
131
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
132
+
133
+ [More Information Needed]
134
+
135
+ ### Results
136
+
137
+ [More Information Needed]
138
+
139
+ #### Summary
140
+
141
+
142
+
143
+ ## Model Examination [optional]
144
+
145
+ <!-- Relevant interpretability work for the model goes here -->
146
+
147
+ [More Information Needed]
148
+
149
+ ## Environmental Impact
150
+
151
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
152
+
153
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
154
+
155
+ - **Hardware Type:** [More Information Needed]
156
+ - **Hours used:** [More Information Needed]
157
+ - **Cloud Provider:** [More Information Needed]
158
+ - **Compute Region:** [More Information Needed]
159
+ - **Carbon Emitted:** [More Information Needed]
160
+
161
+ ## Technical Specifications [optional]
162
+
163
+ ### Model Architecture and Objective
164
+
165
+ [More Information Needed]
166
+
167
+ ### Compute Infrastructure
168
+
169
+ [More Information Needed]
170
+
171
+ #### Hardware
172
+
173
+ [More Information Needed]
174
+
175
+ #### Software
176
+
177
+ [More Information Needed]
178
+
179
+ ## Citation [optional]
180
+
181
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
182
+
183
+ **BibTeX:**
184
+
185
+ [More Information Needed]
186
+
187
+ **APA:**
188
+
189
+ [More Information Needed]
190
+
191
+ ## Glossary [optional]
192
+
193
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
194
+
195
+ [More Information Needed]
196
+
197
+ ## More Information [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Authors [optional]
202
+
203
+ [More Information Needed]
204
+
205
+ ## Model Card Contact
206
+
207
+ [More Information Needed]
208
+ ### Framework versions
209
+
210
+ - PEFT 0.18.0
checkpoint-1000/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "MistralForCausalLM",
7
+ "parent_library": "transformers.models.mistral.modeling_mistral",
8
+ "unsloth_fixed": true
9
+ },
10
+ "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
11
+ "bias": "none",
12
+ "corda_config": null,
13
+ "ensure_weight_tying": false,
14
+ "eva_config": null,
15
+ "exclude_modules": null,
16
+ "fan_in_fan_out": false,
17
+ "inference_mode": true,
18
+ "init_lora_weights": true,
19
+ "layer_replication": null,
20
+ "layers_pattern": null,
21
+ "layers_to_transform": null,
22
+ "loftq_config": {},
23
+ "lora_alpha": 32,
24
+ "lora_bias": false,
25
+ "lora_dropout": 0,
26
+ "megatron_config": null,
27
+ "megatron_core": "megatron.core",
28
+ "modules_to_save": null,
29
+ "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
+ "qalora_group_size": 16,
32
+ "r": 16,
33
+ "rank_pattern": {},
34
+ "revision": null,
35
+ "target_modules": [
36
+ "k_proj",
37
+ "gate_proj",
38
+ "o_proj",
39
+ "v_proj",
40
+ "q_proj",
41
+ "down_proj",
42
+ "up_proj"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": "CAUSAL_LM",
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
checkpoint-1000/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:148d050a179421d107ed3e714b55602294460c2b390f162cff86f39a6617d977
3
+ size 167832240
checkpoint-1000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31a3b3211330479dd260ccee7e779eaaabf42e11fba8ccc0aac598b17b11809e
3
+ size 85723732
checkpoint-1000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:608fccb6c056ce88cdfd5355e6be2046f4d107a24a87c6b0d2c3b200ce6bb4ea
3
+ size 14244
checkpoint-1000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84f4ebc4f434988747cb05f7d349c675df153a128a9118a06de2f22fe7b363cd
3
+ size 1064
checkpoint-1000/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[control_768]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-1000/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1000/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
checkpoint-1000/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1000/trainer_state.json ADDED
@@ -0,0 +1,734 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.6472491909385113,
6
+ "eval_steps": 500,
7
+ "global_step": 1000,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.006472491909385114,
14
+ "grad_norm": 2.0135042667388916,
15
+ "learning_rate": 1.8e-05,
16
+ "loss": 1.5311,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.012944983818770227,
21
+ "grad_norm": 1.0551478862762451,
22
+ "learning_rate": 3.8e-05,
23
+ "loss": 1.2229,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.019417475728155338,
28
+ "grad_norm": 1.101570725440979,
29
+ "learning_rate": 5.8e-05,
30
+ "loss": 0.9845,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.025889967637540454,
35
+ "grad_norm": 0.9952124357223511,
36
+ "learning_rate": 7.800000000000001e-05,
37
+ "loss": 0.8599,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.032362459546925564,
42
+ "grad_norm": 0.8940704464912415,
43
+ "learning_rate": 9.8e-05,
44
+ "loss": 0.8401,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.038834951456310676,
49
+ "grad_norm": 0.8487522602081299,
50
+ "learning_rate": 9.93979933110368e-05,
51
+ "loss": 0.776,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.045307443365695796,
56
+ "grad_norm": 0.7864758968353271,
57
+ "learning_rate": 9.872909698996656e-05,
58
+ "loss": 0.8153,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.05177993527508091,
63
+ "grad_norm": 0.7295967936515808,
64
+ "learning_rate": 9.806020066889633e-05,
65
+ "loss": 0.8251,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.05825242718446602,
70
+ "grad_norm": 0.7075174450874329,
71
+ "learning_rate": 9.739130434782609e-05,
72
+ "loss": 0.749,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.06472491909385113,
77
+ "grad_norm": 0.6422586441040039,
78
+ "learning_rate": 9.672240802675586e-05,
79
+ "loss": 0.7428,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.07119741100323625,
84
+ "grad_norm": 0.6745761632919312,
85
+ "learning_rate": 9.605351170568563e-05,
86
+ "loss": 0.7806,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.07766990291262135,
91
+ "grad_norm": 0.6769530177116394,
92
+ "learning_rate": 9.53846153846154e-05,
93
+ "loss": 0.796,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.08414239482200647,
98
+ "grad_norm": 0.5150454044342041,
99
+ "learning_rate": 9.471571906354515e-05,
100
+ "loss": 0.7486,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.09061488673139159,
105
+ "grad_norm": 0.6074182391166687,
106
+ "learning_rate": 9.404682274247491e-05,
107
+ "loss": 0.7783,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0970873786407767,
112
+ "grad_norm": 0.6745702624320984,
113
+ "learning_rate": 9.337792642140469e-05,
114
+ "loss": 0.7503,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.10355987055016182,
119
+ "grad_norm": 0.6088436841964722,
120
+ "learning_rate": 9.270903010033445e-05,
121
+ "loss": 0.7771,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.11003236245954692,
126
+ "grad_norm": 0.6351744532585144,
127
+ "learning_rate": 9.204013377926422e-05,
128
+ "loss": 0.7554,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.11650485436893204,
133
+ "grad_norm": 0.5281522274017334,
134
+ "learning_rate": 9.137123745819398e-05,
135
+ "loss": 0.7635,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.12297734627831715,
140
+ "grad_norm": 0.6833154559135437,
141
+ "learning_rate": 9.070234113712374e-05,
142
+ "loss": 0.753,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.12944983818770225,
147
+ "grad_norm": 0.5272049307823181,
148
+ "learning_rate": 9.003344481605351e-05,
149
+ "loss": 0.7384,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.13592233009708737,
154
+ "grad_norm": 0.6461851596832275,
155
+ "learning_rate": 8.936454849498329e-05,
156
+ "loss": 0.7548,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.1423948220064725,
161
+ "grad_norm": 0.5979950428009033,
162
+ "learning_rate": 8.869565217391305e-05,
163
+ "loss": 0.7534,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.1488673139158576,
168
+ "grad_norm": 0.6396514773368835,
169
+ "learning_rate": 8.802675585284281e-05,
170
+ "loss": 0.7436,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.1553398058252427,
175
+ "grad_norm": 0.6104384064674377,
176
+ "learning_rate": 8.735785953177258e-05,
177
+ "loss": 0.7345,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.16181229773462782,
182
+ "grad_norm": 0.5535375475883484,
183
+ "learning_rate": 8.668896321070234e-05,
184
+ "loss": 0.7293,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.16828478964401294,
189
+ "grad_norm": 0.5984886884689331,
190
+ "learning_rate": 8.602006688963212e-05,
191
+ "loss": 0.7173,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.17475728155339806,
196
+ "grad_norm": 0.517716646194458,
197
+ "learning_rate": 8.535117056856188e-05,
198
+ "loss": 0.7694,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.18122977346278318,
203
+ "grad_norm": 0.5802426934242249,
204
+ "learning_rate": 8.468227424749165e-05,
205
+ "loss": 0.7505,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.18770226537216828,
210
+ "grad_norm": 0.5289272665977478,
211
+ "learning_rate": 8.401337792642141e-05,
212
+ "loss": 0.7678,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.1941747572815534,
217
+ "grad_norm": 0.5293733477592468,
218
+ "learning_rate": 8.334448160535117e-05,
219
+ "loss": 0.7242,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.20064724919093851,
224
+ "grad_norm": 0.566146969795227,
225
+ "learning_rate": 8.267558528428094e-05,
226
+ "loss": 0.7248,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.20711974110032363,
231
+ "grad_norm": 0.5496853590011597,
232
+ "learning_rate": 8.200668896321071e-05,
233
+ "loss": 0.7588,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.21359223300970873,
238
+ "grad_norm": 0.5420507192611694,
239
+ "learning_rate": 8.133779264214048e-05,
240
+ "loss": 0.7673,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.22006472491909385,
245
+ "grad_norm": 0.582582950592041,
246
+ "learning_rate": 8.066889632107024e-05,
247
+ "loss": 0.7341,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.22653721682847897,
252
+ "grad_norm": 0.5332103371620178,
253
+ "learning_rate": 8e-05,
254
+ "loss": 0.7518,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.23300970873786409,
259
+ "grad_norm": 0.5132858157157898,
260
+ "learning_rate": 7.933110367892977e-05,
261
+ "loss": 0.752,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.23948220064724918,
266
+ "grad_norm": 0.5356752276420593,
267
+ "learning_rate": 7.866220735785953e-05,
268
+ "loss": 0.7432,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.2459546925566343,
273
+ "grad_norm": 0.5592623353004456,
274
+ "learning_rate": 7.79933110367893e-05,
275
+ "loss": 0.7396,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.2524271844660194,
280
+ "grad_norm": 0.5103404521942139,
281
+ "learning_rate": 7.732441471571906e-05,
282
+ "loss": 0.7253,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.2588996763754045,
287
+ "grad_norm": 0.5380160212516785,
288
+ "learning_rate": 7.665551839464883e-05,
289
+ "loss": 0.7464,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.26537216828478966,
294
+ "grad_norm": 0.5373779535293579,
295
+ "learning_rate": 7.598662207357859e-05,
296
+ "loss": 0.7377,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.27184466019417475,
301
+ "grad_norm": 0.523170530796051,
302
+ "learning_rate": 7.531772575250837e-05,
303
+ "loss": 0.7413,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.2783171521035599,
308
+ "grad_norm": 0.5368985533714294,
309
+ "learning_rate": 7.464882943143813e-05,
310
+ "loss": 0.7227,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.284789644012945,
315
+ "grad_norm": 0.587778627872467,
316
+ "learning_rate": 7.39799331103679e-05,
317
+ "loss": 0.7289,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.2912621359223301,
322
+ "grad_norm": 0.5471896529197693,
323
+ "learning_rate": 7.331103678929766e-05,
324
+ "loss": 0.7301,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.2977346278317152,
329
+ "grad_norm": 0.5446394085884094,
330
+ "learning_rate": 7.264214046822742e-05,
331
+ "loss": 0.7325,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.3042071197411003,
336
+ "grad_norm": 0.5761464238166809,
337
+ "learning_rate": 7.197324414715719e-05,
338
+ "loss": 0.7583,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.3106796116504854,
343
+ "grad_norm": 0.5082632899284363,
344
+ "learning_rate": 7.130434782608696e-05,
345
+ "loss": 0.7382,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.31715210355987056,
350
+ "grad_norm": 0.5441333651542664,
351
+ "learning_rate": 7.063545150501673e-05,
352
+ "loss": 0.7411,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.32362459546925565,
357
+ "grad_norm": 0.5039440393447876,
358
+ "learning_rate": 6.996655518394649e-05,
359
+ "loss": 0.7332,
360
+ "step": 500
361
+ },
362
+ {
363
+ "epoch": 0.3300970873786408,
364
+ "grad_norm": 0.5474230051040649,
365
+ "learning_rate": 6.929765886287626e-05,
366
+ "loss": 0.7216,
367
+ "step": 510
368
+ },
369
+ {
370
+ "epoch": 0.3365695792880259,
371
+ "grad_norm": 0.5532674789428711,
372
+ "learning_rate": 6.862876254180602e-05,
373
+ "loss": 0.7389,
374
+ "step": 520
375
+ },
376
+ {
377
+ "epoch": 0.343042071197411,
378
+ "grad_norm": 0.5631316900253296,
379
+ "learning_rate": 6.79598662207358e-05,
380
+ "loss": 0.7065,
381
+ "step": 530
382
+ },
383
+ {
384
+ "epoch": 0.34951456310679613,
385
+ "grad_norm": 0.45850899815559387,
386
+ "learning_rate": 6.729096989966556e-05,
387
+ "loss": 0.7472,
388
+ "step": 540
389
+ },
390
+ {
391
+ "epoch": 0.3559870550161812,
392
+ "grad_norm": 0.48253533244132996,
393
+ "learning_rate": 6.662207357859532e-05,
394
+ "loss": 0.7312,
395
+ "step": 550
396
+ },
397
+ {
398
+ "epoch": 0.36245954692556637,
399
+ "grad_norm": 0.6004955172538757,
400
+ "learning_rate": 6.595317725752509e-05,
401
+ "loss": 0.7681,
402
+ "step": 560
403
+ },
404
+ {
405
+ "epoch": 0.36893203883495146,
406
+ "grad_norm": 0.517497181892395,
407
+ "learning_rate": 6.528428093645485e-05,
408
+ "loss": 0.7581,
409
+ "step": 570
410
+ },
411
+ {
412
+ "epoch": 0.37540453074433655,
413
+ "grad_norm": 0.5473213195800781,
414
+ "learning_rate": 6.461538461538462e-05,
415
+ "loss": 0.7459,
416
+ "step": 580
417
+ },
418
+ {
419
+ "epoch": 0.3818770226537217,
420
+ "grad_norm": 0.5596595406532288,
421
+ "learning_rate": 6.39464882943144e-05,
422
+ "loss": 0.7322,
423
+ "step": 590
424
+ },
425
+ {
426
+ "epoch": 0.3883495145631068,
427
+ "grad_norm": 0.5507789254188538,
428
+ "learning_rate": 6.327759197324416e-05,
429
+ "loss": 0.7324,
430
+ "step": 600
431
+ },
432
+ {
433
+ "epoch": 0.3948220064724919,
434
+ "grad_norm": 0.5934524536132812,
435
+ "learning_rate": 6.260869565217392e-05,
436
+ "loss": 0.7072,
437
+ "step": 610
438
+ },
439
+ {
440
+ "epoch": 0.40129449838187703,
441
+ "grad_norm": 0.5208216905593872,
442
+ "learning_rate": 6.193979933110368e-05,
443
+ "loss": 0.755,
444
+ "step": 620
445
+ },
446
+ {
447
+ "epoch": 0.4077669902912621,
448
+ "grad_norm": 0.506360650062561,
449
+ "learning_rate": 6.127090301003345e-05,
450
+ "loss": 0.7244,
451
+ "step": 630
452
+ },
453
+ {
454
+ "epoch": 0.41423948220064727,
455
+ "grad_norm": 0.5485314130783081,
456
+ "learning_rate": 6.0602006688963206e-05,
457
+ "loss": 0.7284,
458
+ "step": 640
459
+ },
460
+ {
461
+ "epoch": 0.42071197411003236,
462
+ "grad_norm": 0.5979824066162109,
463
+ "learning_rate": 5.9933110367892977e-05,
464
+ "loss": 0.7065,
465
+ "step": 650
466
+ },
467
+ {
468
+ "epoch": 0.42718446601941745,
469
+ "grad_norm": 0.5200499892234802,
470
+ "learning_rate": 5.926421404682274e-05,
471
+ "loss": 0.7124,
472
+ "step": 660
473
+ },
474
+ {
475
+ "epoch": 0.4336569579288026,
476
+ "grad_norm": 0.5502509474754333,
477
+ "learning_rate": 5.859531772575251e-05,
478
+ "loss": 0.7493,
479
+ "step": 670
480
+ },
481
+ {
482
+ "epoch": 0.4401294498381877,
483
+ "grad_norm": 0.4699322581291199,
484
+ "learning_rate": 5.7926421404682275e-05,
485
+ "loss": 0.7588,
486
+ "step": 680
487
+ },
488
+ {
489
+ "epoch": 0.44660194174757284,
490
+ "grad_norm": 0.4809620678424835,
491
+ "learning_rate": 5.725752508361204e-05,
492
+ "loss": 0.7463,
493
+ "step": 690
494
+ },
495
+ {
496
+ "epoch": 0.45307443365695793,
497
+ "grad_norm": 0.5801950693130493,
498
+ "learning_rate": 5.658862876254181e-05,
499
+ "loss": 0.7206,
500
+ "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.459546925566343,
504
+ "grad_norm": 0.5444105267524719,
505
+ "learning_rate": 5.591973244147157e-05,
506
+ "loss": 0.7197,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.46601941747572817,
511
+ "grad_norm": 0.4769653379917145,
512
+ "learning_rate": 5.525083612040134e-05,
513
+ "loss": 0.7359,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.47249190938511326,
518
+ "grad_norm": 0.5103979706764221,
519
+ "learning_rate": 5.458193979933111e-05,
520
+ "loss": 0.769,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.47896440129449835,
525
+ "grad_norm": 0.543662965297699,
526
+ "learning_rate": 5.391304347826087e-05,
527
+ "loss": 0.7618,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.4854368932038835,
532
+ "grad_norm": 0.5364013314247131,
533
+ "learning_rate": 5.3244147157190635e-05,
534
+ "loss": 0.72,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.4919093851132686,
539
+ "grad_norm": 0.6091406345367432,
540
+ "learning_rate": 5.2575250836120406e-05,
541
+ "loss": 0.6965,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.49838187702265374,
546
+ "grad_norm": 0.5977626442909241,
547
+ "learning_rate": 5.190635451505017e-05,
548
+ "loss": 0.7172,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.5048543689320388,
553
+ "grad_norm": 0.45984843373298645,
554
+ "learning_rate": 5.1237458193979934e-05,
555
+ "loss": 0.7391,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.511326860841424,
560
+ "grad_norm": 0.49561065435409546,
561
+ "learning_rate": 5.0568561872909704e-05,
562
+ "loss": 0.7526,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.517799352750809,
567
+ "grad_norm": 0.4905160367488861,
568
+ "learning_rate": 4.989966555183947e-05,
569
+ "loss": 0.7278,
570
+ "step": 800
571
+ },
572
+ {
573
+ "epoch": 0.5242718446601942,
574
+ "grad_norm": 0.5338721871376038,
575
+ "learning_rate": 4.923076923076924e-05,
576
+ "loss": 0.7297,
577
+ "step": 810
578
+ },
579
+ {
580
+ "epoch": 0.5307443365695793,
581
+ "grad_norm": 0.5137319564819336,
582
+ "learning_rate": 4.8561872909698996e-05,
583
+ "loss": 0.7863,
584
+ "step": 820
585
+ },
586
+ {
587
+ "epoch": 0.5372168284789643,
588
+ "grad_norm": 0.4649944603443146,
589
+ "learning_rate": 4.789297658862876e-05,
590
+ "loss": 0.7525,
591
+ "step": 830
592
+ },
593
+ {
594
+ "epoch": 0.5436893203883495,
595
+ "grad_norm": 0.5467334985733032,
596
+ "learning_rate": 4.722408026755853e-05,
597
+ "loss": 0.7144,
598
+ "step": 840
599
+ },
600
+ {
601
+ "epoch": 0.5501618122977346,
602
+ "grad_norm": 0.47389763593673706,
603
+ "learning_rate": 4.6555183946488294e-05,
604
+ "loss": 0.7217,
605
+ "step": 850
606
+ },
607
+ {
608
+ "epoch": 0.5566343042071198,
609
+ "grad_norm": 0.45604580640792847,
610
+ "learning_rate": 4.5886287625418065e-05,
611
+ "loss": 0.7243,
612
+ "step": 860
613
+ },
614
+ {
615
+ "epoch": 0.5631067961165048,
616
+ "grad_norm": 0.548736572265625,
617
+ "learning_rate": 4.521739130434783e-05,
618
+ "loss": 0.7078,
619
+ "step": 870
620
+ },
621
+ {
622
+ "epoch": 0.56957928802589,
623
+ "grad_norm": 0.5387297868728638,
624
+ "learning_rate": 4.454849498327759e-05,
625
+ "loss": 0.6778,
626
+ "step": 880
627
+ },
628
+ {
629
+ "epoch": 0.5760517799352751,
630
+ "grad_norm": 0.5060684680938721,
631
+ "learning_rate": 4.3879598662207363e-05,
632
+ "loss": 0.7088,
633
+ "step": 890
634
+ },
635
+ {
636
+ "epoch": 0.5825242718446602,
637
+ "grad_norm": 0.5241347551345825,
638
+ "learning_rate": 4.321070234113713e-05,
639
+ "loss": 0.7232,
640
+ "step": 900
641
+ },
642
+ {
643
+ "epoch": 0.5889967637540453,
644
+ "grad_norm": 0.5504106283187866,
645
+ "learning_rate": 4.254180602006689e-05,
646
+ "loss": 0.7118,
647
+ "step": 910
648
+ },
649
+ {
650
+ "epoch": 0.5954692556634305,
651
+ "grad_norm": 0.47307100892066956,
652
+ "learning_rate": 4.187290969899666e-05,
653
+ "loss": 0.7377,
654
+ "step": 920
655
+ },
656
+ {
657
+ "epoch": 0.6019417475728155,
658
+ "grad_norm": 0.5132865905761719,
659
+ "learning_rate": 4.1204013377926426e-05,
660
+ "loss": 0.7237,
661
+ "step": 930
662
+ },
663
+ {
664
+ "epoch": 0.6084142394822006,
665
+ "grad_norm": 0.5692107677459717,
666
+ "learning_rate": 4.053511705685619e-05,
667
+ "loss": 0.6933,
668
+ "step": 940
669
+ },
670
+ {
671
+ "epoch": 0.6148867313915858,
672
+ "grad_norm": 0.5088536143302917,
673
+ "learning_rate": 3.986622073578595e-05,
674
+ "loss": 0.7094,
675
+ "step": 950
676
+ },
677
+ {
678
+ "epoch": 0.6213592233009708,
679
+ "grad_norm": 0.47623276710510254,
680
+ "learning_rate": 3.919732441471572e-05,
681
+ "loss": 0.7219,
682
+ "step": 960
683
+ },
684
+ {
685
+ "epoch": 0.627831715210356,
686
+ "grad_norm": 0.48927319049835205,
687
+ "learning_rate": 3.852842809364549e-05,
688
+ "loss": 0.7324,
689
+ "step": 970
690
+ },
691
+ {
692
+ "epoch": 0.6343042071197411,
693
+ "grad_norm": 0.4764053523540497,
694
+ "learning_rate": 3.785953177257525e-05,
695
+ "loss": 0.7286,
696
+ "step": 980
697
+ },
698
+ {
699
+ "epoch": 0.6407766990291263,
700
+ "grad_norm": 0.4962717592716217,
701
+ "learning_rate": 3.7190635451505016e-05,
702
+ "loss": 0.7471,
703
+ "step": 990
704
+ },
705
+ {
706
+ "epoch": 0.6472491909385113,
707
+ "grad_norm": 0.5014144778251648,
708
+ "learning_rate": 3.6521739130434786e-05,
709
+ "loss": 0.6966,
710
+ "step": 1000
711
+ }
712
+ ],
713
+ "logging_steps": 10,
714
+ "max_steps": 1545,
715
+ "num_input_tokens_seen": 0,
716
+ "num_train_epochs": 1,
717
+ "save_steps": 500,
718
+ "stateful_callbacks": {
719
+ "TrainerControl": {
720
+ "args": {
721
+ "should_epoch_stop": false,
722
+ "should_evaluate": false,
723
+ "should_log": false,
724
+ "should_save": true,
725
+ "should_training_stop": false
726
+ },
727
+ "attributes": {}
728
+ }
729
+ },
730
+ "total_flos": 3.626219816245002e+17,
731
+ "train_batch_size": 16,
732
+ "trial_name": null,
733
+ "trial_params": null
734
+ }
checkpoint-1000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3ca7cc1e4f7451b1bc8044e4e525255bb7182251d1959fd39f0cc8c4abb5629
3
+ size 5752
checkpoint-1500/README.md ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/mistral-7b-v0.3-bnb-4bit
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:unsloth/mistral-7b-v0.3-bnb-4bit
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ - unsloth
12
+ ---
13
+
14
+ # Model Card for Model ID
15
+
16
+ <!-- Provide a quick summary of what the model is/does. -->
17
+
18
+
19
+
20
+ ## Model Details
21
+
22
+ ### Model Description
23
+
24
+ <!-- Provide a longer summary of what this model is. -->
25
+
26
+
27
+
28
+ - **Developed by:** [More Information Needed]
29
+ - **Funded by [optional]:** [More Information Needed]
30
+ - **Shared by [optional]:** [More Information Needed]
31
+ - **Model type:** [More Information Needed]
32
+ - **Language(s) (NLP):** [More Information Needed]
33
+ - **License:** [More Information Needed]
34
+ - **Finetuned from model [optional]:** [More Information Needed]
35
+
36
+ ### Model Sources [optional]
37
+
38
+ <!-- Provide the basic links for the model. -->
39
+
40
+ - **Repository:** [More Information Needed]
41
+ - **Paper [optional]:** [More Information Needed]
42
+ - **Demo [optional]:** [More Information Needed]
43
+
44
+ ## Uses
45
+
46
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
47
+
48
+ ### Direct Use
49
+
50
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
51
+
52
+ [More Information Needed]
53
+
54
+ ### Downstream Use [optional]
55
+
56
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
57
+
58
+ [More Information Needed]
59
+
60
+ ### Out-of-Scope Use
61
+
62
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
63
+
64
+ [More Information Needed]
65
+
66
+ ## Bias, Risks, and Limitations
67
+
68
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
69
+
70
+ [More Information Needed]
71
+
72
+ ### Recommendations
73
+
74
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
75
+
76
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
77
+
78
+ ## How to Get Started with the Model
79
+
80
+ Use the code below to get started with the model.
81
+
82
+ [More Information Needed]
83
+
84
+ ## Training Details
85
+
86
+ ### Training Data
87
+
88
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
89
+
90
+ [More Information Needed]
91
+
92
+ ### Training Procedure
93
+
94
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
95
+
96
+ #### Preprocessing [optional]
97
+
98
+ [More Information Needed]
99
+
100
+
101
+ #### Training Hyperparameters
102
+
103
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
104
+
105
+ #### Speeds, Sizes, Times [optional]
106
+
107
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
108
+
109
+ [More Information Needed]
110
+
111
+ ## Evaluation
112
+
113
+ <!-- This section describes the evaluation protocols and provides the results. -->
114
+
115
+ ### Testing Data, Factors & Metrics
116
+
117
+ #### Testing Data
118
+
119
+ <!-- This should link to a Dataset Card if possible. -->
120
+
121
+ [More Information Needed]
122
+
123
+ #### Factors
124
+
125
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
126
+
127
+ [More Information Needed]
128
+
129
+ #### Metrics
130
+
131
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
132
+
133
+ [More Information Needed]
134
+
135
+ ### Results
136
+
137
+ [More Information Needed]
138
+
139
+ #### Summary
140
+
141
+
142
+
143
+ ## Model Examination [optional]
144
+
145
+ <!-- Relevant interpretability work for the model goes here -->
146
+
147
+ [More Information Needed]
148
+
149
+ ## Environmental Impact
150
+
151
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
152
+
153
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
154
+
155
+ - **Hardware Type:** [More Information Needed]
156
+ - **Hours used:** [More Information Needed]
157
+ - **Cloud Provider:** [More Information Needed]
158
+ - **Compute Region:** [More Information Needed]
159
+ - **Carbon Emitted:** [More Information Needed]
160
+
161
+ ## Technical Specifications [optional]
162
+
163
+ ### Model Architecture and Objective
164
+
165
+ [More Information Needed]
166
+
167
+ ### Compute Infrastructure
168
+
169
+ [More Information Needed]
170
+
171
+ #### Hardware
172
+
173
+ [More Information Needed]
174
+
175
+ #### Software
176
+
177
+ [More Information Needed]
178
+
179
+ ## Citation [optional]
180
+
181
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
182
+
183
+ **BibTeX:**
184
+
185
+ [More Information Needed]
186
+
187
+ **APA:**
188
+
189
+ [More Information Needed]
190
+
191
+ ## Glossary [optional]
192
+
193
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
194
+
195
+ [More Information Needed]
196
+
197
+ ## More Information [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Authors [optional]
202
+
203
+ [More Information Needed]
204
+
205
+ ## Model Card Contact
206
+
207
+ [More Information Needed]
208
+ ### Framework versions
209
+
210
+ - PEFT 0.18.0
checkpoint-1500/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "MistralForCausalLM",
7
+ "parent_library": "transformers.models.mistral.modeling_mistral",
8
+ "unsloth_fixed": true
9
+ },
10
+ "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
11
+ "bias": "none",
12
+ "corda_config": null,
13
+ "ensure_weight_tying": false,
14
+ "eva_config": null,
15
+ "exclude_modules": null,
16
+ "fan_in_fan_out": false,
17
+ "inference_mode": true,
18
+ "init_lora_weights": true,
19
+ "layer_replication": null,
20
+ "layers_pattern": null,
21
+ "layers_to_transform": null,
22
+ "loftq_config": {},
23
+ "lora_alpha": 32,
24
+ "lora_bias": false,
25
+ "lora_dropout": 0,
26
+ "megatron_config": null,
27
+ "megatron_core": "megatron.core",
28
+ "modules_to_save": null,
29
+ "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
+ "qalora_group_size": 16,
32
+ "r": 16,
33
+ "rank_pattern": {},
34
+ "revision": null,
35
+ "target_modules": [
36
+ "k_proj",
37
+ "gate_proj",
38
+ "o_proj",
39
+ "v_proj",
40
+ "q_proj",
41
+ "down_proj",
42
+ "up_proj"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": "CAUSAL_LM",
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
checkpoint-1500/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd9f0ddf073b81967f96c0089920f96ff8c303f417183813c216dc7377adefb3
3
+ size 167832240
checkpoint-1500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a78e398a4641f882c363bf08601ca3a25cabffb0afa15023a9517779a055247
3
+ size 85723732
checkpoint-1500/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:608fccb6c056ce88cdfd5355e6be2046f4d107a24a87c6b0d2c3b200ce6bb4ea
3
+ size 14244
checkpoint-1500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37be4a2012710e99dd33f94c5f188bac520a322a9e33ecf222bfa872639c5aac
3
+ size 1064
checkpoint-1500/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[control_768]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-1500/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1500/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
checkpoint-1500/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1500/trainer_state.json ADDED
@@ -0,0 +1,1084 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.970873786407767,
6
+ "eval_steps": 500,
7
+ "global_step": 1500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.006472491909385114,
14
+ "grad_norm": 2.0135042667388916,
15
+ "learning_rate": 1.8e-05,
16
+ "loss": 1.5311,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.012944983818770227,
21
+ "grad_norm": 1.0551478862762451,
22
+ "learning_rate": 3.8e-05,
23
+ "loss": 1.2229,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.019417475728155338,
28
+ "grad_norm": 1.101570725440979,
29
+ "learning_rate": 5.8e-05,
30
+ "loss": 0.9845,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.025889967637540454,
35
+ "grad_norm": 0.9952124357223511,
36
+ "learning_rate": 7.800000000000001e-05,
37
+ "loss": 0.8599,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.032362459546925564,
42
+ "grad_norm": 0.8940704464912415,
43
+ "learning_rate": 9.8e-05,
44
+ "loss": 0.8401,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.038834951456310676,
49
+ "grad_norm": 0.8487522602081299,
50
+ "learning_rate": 9.93979933110368e-05,
51
+ "loss": 0.776,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.045307443365695796,
56
+ "grad_norm": 0.7864758968353271,
57
+ "learning_rate": 9.872909698996656e-05,
58
+ "loss": 0.8153,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.05177993527508091,
63
+ "grad_norm": 0.7295967936515808,
64
+ "learning_rate": 9.806020066889633e-05,
65
+ "loss": 0.8251,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.05825242718446602,
70
+ "grad_norm": 0.7075174450874329,
71
+ "learning_rate": 9.739130434782609e-05,
72
+ "loss": 0.749,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.06472491909385113,
77
+ "grad_norm": 0.6422586441040039,
78
+ "learning_rate": 9.672240802675586e-05,
79
+ "loss": 0.7428,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.07119741100323625,
84
+ "grad_norm": 0.6745761632919312,
85
+ "learning_rate": 9.605351170568563e-05,
86
+ "loss": 0.7806,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.07766990291262135,
91
+ "grad_norm": 0.6769530177116394,
92
+ "learning_rate": 9.53846153846154e-05,
93
+ "loss": 0.796,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.08414239482200647,
98
+ "grad_norm": 0.5150454044342041,
99
+ "learning_rate": 9.471571906354515e-05,
100
+ "loss": 0.7486,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.09061488673139159,
105
+ "grad_norm": 0.6074182391166687,
106
+ "learning_rate": 9.404682274247491e-05,
107
+ "loss": 0.7783,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0970873786407767,
112
+ "grad_norm": 0.6745702624320984,
113
+ "learning_rate": 9.337792642140469e-05,
114
+ "loss": 0.7503,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.10355987055016182,
119
+ "grad_norm": 0.6088436841964722,
120
+ "learning_rate": 9.270903010033445e-05,
121
+ "loss": 0.7771,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.11003236245954692,
126
+ "grad_norm": 0.6351744532585144,
127
+ "learning_rate": 9.204013377926422e-05,
128
+ "loss": 0.7554,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.11650485436893204,
133
+ "grad_norm": 0.5281522274017334,
134
+ "learning_rate": 9.137123745819398e-05,
135
+ "loss": 0.7635,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.12297734627831715,
140
+ "grad_norm": 0.6833154559135437,
141
+ "learning_rate": 9.070234113712374e-05,
142
+ "loss": 0.753,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.12944983818770225,
147
+ "grad_norm": 0.5272049307823181,
148
+ "learning_rate": 9.003344481605351e-05,
149
+ "loss": 0.7384,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.13592233009708737,
154
+ "grad_norm": 0.6461851596832275,
155
+ "learning_rate": 8.936454849498329e-05,
156
+ "loss": 0.7548,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.1423948220064725,
161
+ "grad_norm": 0.5979950428009033,
162
+ "learning_rate": 8.869565217391305e-05,
163
+ "loss": 0.7534,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.1488673139158576,
168
+ "grad_norm": 0.6396514773368835,
169
+ "learning_rate": 8.802675585284281e-05,
170
+ "loss": 0.7436,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.1553398058252427,
175
+ "grad_norm": 0.6104384064674377,
176
+ "learning_rate": 8.735785953177258e-05,
177
+ "loss": 0.7345,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.16181229773462782,
182
+ "grad_norm": 0.5535375475883484,
183
+ "learning_rate": 8.668896321070234e-05,
184
+ "loss": 0.7293,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.16828478964401294,
189
+ "grad_norm": 0.5984886884689331,
190
+ "learning_rate": 8.602006688963212e-05,
191
+ "loss": 0.7173,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.17475728155339806,
196
+ "grad_norm": 0.517716646194458,
197
+ "learning_rate": 8.535117056856188e-05,
198
+ "loss": 0.7694,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.18122977346278318,
203
+ "grad_norm": 0.5802426934242249,
204
+ "learning_rate": 8.468227424749165e-05,
205
+ "loss": 0.7505,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.18770226537216828,
210
+ "grad_norm": 0.5289272665977478,
211
+ "learning_rate": 8.401337792642141e-05,
212
+ "loss": 0.7678,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.1941747572815534,
217
+ "grad_norm": 0.5293733477592468,
218
+ "learning_rate": 8.334448160535117e-05,
219
+ "loss": 0.7242,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.20064724919093851,
224
+ "grad_norm": 0.566146969795227,
225
+ "learning_rate": 8.267558528428094e-05,
226
+ "loss": 0.7248,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.20711974110032363,
231
+ "grad_norm": 0.5496853590011597,
232
+ "learning_rate": 8.200668896321071e-05,
233
+ "loss": 0.7588,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.21359223300970873,
238
+ "grad_norm": 0.5420507192611694,
239
+ "learning_rate": 8.133779264214048e-05,
240
+ "loss": 0.7673,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.22006472491909385,
245
+ "grad_norm": 0.582582950592041,
246
+ "learning_rate": 8.066889632107024e-05,
247
+ "loss": 0.7341,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.22653721682847897,
252
+ "grad_norm": 0.5332103371620178,
253
+ "learning_rate": 8e-05,
254
+ "loss": 0.7518,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.23300970873786409,
259
+ "grad_norm": 0.5132858157157898,
260
+ "learning_rate": 7.933110367892977e-05,
261
+ "loss": 0.752,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.23948220064724918,
266
+ "grad_norm": 0.5356752276420593,
267
+ "learning_rate": 7.866220735785953e-05,
268
+ "loss": 0.7432,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.2459546925566343,
273
+ "grad_norm": 0.5592623353004456,
274
+ "learning_rate": 7.79933110367893e-05,
275
+ "loss": 0.7396,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.2524271844660194,
280
+ "grad_norm": 0.5103404521942139,
281
+ "learning_rate": 7.732441471571906e-05,
282
+ "loss": 0.7253,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.2588996763754045,
287
+ "grad_norm": 0.5380160212516785,
288
+ "learning_rate": 7.665551839464883e-05,
289
+ "loss": 0.7464,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.26537216828478966,
294
+ "grad_norm": 0.5373779535293579,
295
+ "learning_rate": 7.598662207357859e-05,
296
+ "loss": 0.7377,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.27184466019417475,
301
+ "grad_norm": 0.523170530796051,
302
+ "learning_rate": 7.531772575250837e-05,
303
+ "loss": 0.7413,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.2783171521035599,
308
+ "grad_norm": 0.5368985533714294,
309
+ "learning_rate": 7.464882943143813e-05,
310
+ "loss": 0.7227,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.284789644012945,
315
+ "grad_norm": 0.587778627872467,
316
+ "learning_rate": 7.39799331103679e-05,
317
+ "loss": 0.7289,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.2912621359223301,
322
+ "grad_norm": 0.5471896529197693,
323
+ "learning_rate": 7.331103678929766e-05,
324
+ "loss": 0.7301,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.2977346278317152,
329
+ "grad_norm": 0.5446394085884094,
330
+ "learning_rate": 7.264214046822742e-05,
331
+ "loss": 0.7325,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.3042071197411003,
336
+ "grad_norm": 0.5761464238166809,
337
+ "learning_rate": 7.197324414715719e-05,
338
+ "loss": 0.7583,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.3106796116504854,
343
+ "grad_norm": 0.5082632899284363,
344
+ "learning_rate": 7.130434782608696e-05,
345
+ "loss": 0.7382,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.31715210355987056,
350
+ "grad_norm": 0.5441333651542664,
351
+ "learning_rate": 7.063545150501673e-05,
352
+ "loss": 0.7411,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.32362459546925565,
357
+ "grad_norm": 0.5039440393447876,
358
+ "learning_rate": 6.996655518394649e-05,
359
+ "loss": 0.7332,
360
+ "step": 500
361
+ },
362
+ {
363
+ "epoch": 0.3300970873786408,
364
+ "grad_norm": 0.5474230051040649,
365
+ "learning_rate": 6.929765886287626e-05,
366
+ "loss": 0.7216,
367
+ "step": 510
368
+ },
369
+ {
370
+ "epoch": 0.3365695792880259,
371
+ "grad_norm": 0.5532674789428711,
372
+ "learning_rate": 6.862876254180602e-05,
373
+ "loss": 0.7389,
374
+ "step": 520
375
+ },
376
+ {
377
+ "epoch": 0.343042071197411,
378
+ "grad_norm": 0.5631316900253296,
379
+ "learning_rate": 6.79598662207358e-05,
380
+ "loss": 0.7065,
381
+ "step": 530
382
+ },
383
+ {
384
+ "epoch": 0.34951456310679613,
385
+ "grad_norm": 0.45850899815559387,
386
+ "learning_rate": 6.729096989966556e-05,
387
+ "loss": 0.7472,
388
+ "step": 540
389
+ },
390
+ {
391
+ "epoch": 0.3559870550161812,
392
+ "grad_norm": 0.48253533244132996,
393
+ "learning_rate": 6.662207357859532e-05,
394
+ "loss": 0.7312,
395
+ "step": 550
396
+ },
397
+ {
398
+ "epoch": 0.36245954692556637,
399
+ "grad_norm": 0.6004955172538757,
400
+ "learning_rate": 6.595317725752509e-05,
401
+ "loss": 0.7681,
402
+ "step": 560
403
+ },
404
+ {
405
+ "epoch": 0.36893203883495146,
406
+ "grad_norm": 0.517497181892395,
407
+ "learning_rate": 6.528428093645485e-05,
408
+ "loss": 0.7581,
409
+ "step": 570
410
+ },
411
+ {
412
+ "epoch": 0.37540453074433655,
413
+ "grad_norm": 0.5473213195800781,
414
+ "learning_rate": 6.461538461538462e-05,
415
+ "loss": 0.7459,
416
+ "step": 580
417
+ },
418
+ {
419
+ "epoch": 0.3818770226537217,
420
+ "grad_norm": 0.5596595406532288,
421
+ "learning_rate": 6.39464882943144e-05,
422
+ "loss": 0.7322,
423
+ "step": 590
424
+ },
425
+ {
426
+ "epoch": 0.3883495145631068,
427
+ "grad_norm": 0.5507789254188538,
428
+ "learning_rate": 6.327759197324416e-05,
429
+ "loss": 0.7324,
430
+ "step": 600
431
+ },
432
+ {
433
+ "epoch": 0.3948220064724919,
434
+ "grad_norm": 0.5934524536132812,
435
+ "learning_rate": 6.260869565217392e-05,
436
+ "loss": 0.7072,
437
+ "step": 610
438
+ },
439
+ {
440
+ "epoch": 0.40129449838187703,
441
+ "grad_norm": 0.5208216905593872,
442
+ "learning_rate": 6.193979933110368e-05,
443
+ "loss": 0.755,
444
+ "step": 620
445
+ },
446
+ {
447
+ "epoch": 0.4077669902912621,
448
+ "grad_norm": 0.506360650062561,
449
+ "learning_rate": 6.127090301003345e-05,
450
+ "loss": 0.7244,
451
+ "step": 630
452
+ },
453
+ {
454
+ "epoch": 0.41423948220064727,
455
+ "grad_norm": 0.5485314130783081,
456
+ "learning_rate": 6.0602006688963206e-05,
457
+ "loss": 0.7284,
458
+ "step": 640
459
+ },
460
+ {
461
+ "epoch": 0.42071197411003236,
462
+ "grad_norm": 0.5979824066162109,
463
+ "learning_rate": 5.9933110367892977e-05,
464
+ "loss": 0.7065,
465
+ "step": 650
466
+ },
467
+ {
468
+ "epoch": 0.42718446601941745,
469
+ "grad_norm": 0.5200499892234802,
470
+ "learning_rate": 5.926421404682274e-05,
471
+ "loss": 0.7124,
472
+ "step": 660
473
+ },
474
+ {
475
+ "epoch": 0.4336569579288026,
476
+ "grad_norm": 0.5502509474754333,
477
+ "learning_rate": 5.859531772575251e-05,
478
+ "loss": 0.7493,
479
+ "step": 670
480
+ },
481
+ {
482
+ "epoch": 0.4401294498381877,
483
+ "grad_norm": 0.4699322581291199,
484
+ "learning_rate": 5.7926421404682275e-05,
485
+ "loss": 0.7588,
486
+ "step": 680
487
+ },
488
+ {
489
+ "epoch": 0.44660194174757284,
490
+ "grad_norm": 0.4809620678424835,
491
+ "learning_rate": 5.725752508361204e-05,
492
+ "loss": 0.7463,
493
+ "step": 690
494
+ },
495
+ {
496
+ "epoch": 0.45307443365695793,
497
+ "grad_norm": 0.5801950693130493,
498
+ "learning_rate": 5.658862876254181e-05,
499
+ "loss": 0.7206,
500
+ "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.459546925566343,
504
+ "grad_norm": 0.5444105267524719,
505
+ "learning_rate": 5.591973244147157e-05,
506
+ "loss": 0.7197,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.46601941747572817,
511
+ "grad_norm": 0.4769653379917145,
512
+ "learning_rate": 5.525083612040134e-05,
513
+ "loss": 0.7359,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.47249190938511326,
518
+ "grad_norm": 0.5103979706764221,
519
+ "learning_rate": 5.458193979933111e-05,
520
+ "loss": 0.769,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.47896440129449835,
525
+ "grad_norm": 0.543662965297699,
526
+ "learning_rate": 5.391304347826087e-05,
527
+ "loss": 0.7618,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.4854368932038835,
532
+ "grad_norm": 0.5364013314247131,
533
+ "learning_rate": 5.3244147157190635e-05,
534
+ "loss": 0.72,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.4919093851132686,
539
+ "grad_norm": 0.6091406345367432,
540
+ "learning_rate": 5.2575250836120406e-05,
541
+ "loss": 0.6965,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.49838187702265374,
546
+ "grad_norm": 0.5977626442909241,
547
+ "learning_rate": 5.190635451505017e-05,
548
+ "loss": 0.7172,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.5048543689320388,
553
+ "grad_norm": 0.45984843373298645,
554
+ "learning_rate": 5.1237458193979934e-05,
555
+ "loss": 0.7391,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.511326860841424,
560
+ "grad_norm": 0.49561065435409546,
561
+ "learning_rate": 5.0568561872909704e-05,
562
+ "loss": 0.7526,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.517799352750809,
567
+ "grad_norm": 0.4905160367488861,
568
+ "learning_rate": 4.989966555183947e-05,
569
+ "loss": 0.7278,
570
+ "step": 800
571
+ },
572
+ {
573
+ "epoch": 0.5242718446601942,
574
+ "grad_norm": 0.5338721871376038,
575
+ "learning_rate": 4.923076923076924e-05,
576
+ "loss": 0.7297,
577
+ "step": 810
578
+ },
579
+ {
580
+ "epoch": 0.5307443365695793,
581
+ "grad_norm": 0.5137319564819336,
582
+ "learning_rate": 4.8561872909698996e-05,
583
+ "loss": 0.7863,
584
+ "step": 820
585
+ },
586
+ {
587
+ "epoch": 0.5372168284789643,
588
+ "grad_norm": 0.4649944603443146,
589
+ "learning_rate": 4.789297658862876e-05,
590
+ "loss": 0.7525,
591
+ "step": 830
592
+ },
593
+ {
594
+ "epoch": 0.5436893203883495,
595
+ "grad_norm": 0.5467334985733032,
596
+ "learning_rate": 4.722408026755853e-05,
597
+ "loss": 0.7144,
598
+ "step": 840
599
+ },
600
+ {
601
+ "epoch": 0.5501618122977346,
602
+ "grad_norm": 0.47389763593673706,
603
+ "learning_rate": 4.6555183946488294e-05,
604
+ "loss": 0.7217,
605
+ "step": 850
606
+ },
607
+ {
608
+ "epoch": 0.5566343042071198,
609
+ "grad_norm": 0.45604580640792847,
610
+ "learning_rate": 4.5886287625418065e-05,
611
+ "loss": 0.7243,
612
+ "step": 860
613
+ },
614
+ {
615
+ "epoch": 0.5631067961165048,
616
+ "grad_norm": 0.548736572265625,
617
+ "learning_rate": 4.521739130434783e-05,
618
+ "loss": 0.7078,
619
+ "step": 870
620
+ },
621
+ {
622
+ "epoch": 0.56957928802589,
623
+ "grad_norm": 0.5387297868728638,
624
+ "learning_rate": 4.454849498327759e-05,
625
+ "loss": 0.6778,
626
+ "step": 880
627
+ },
628
+ {
629
+ "epoch": 0.5760517799352751,
630
+ "grad_norm": 0.5060684680938721,
631
+ "learning_rate": 4.3879598662207363e-05,
632
+ "loss": 0.7088,
633
+ "step": 890
634
+ },
635
+ {
636
+ "epoch": 0.5825242718446602,
637
+ "grad_norm": 0.5241347551345825,
638
+ "learning_rate": 4.321070234113713e-05,
639
+ "loss": 0.7232,
640
+ "step": 900
641
+ },
642
+ {
643
+ "epoch": 0.5889967637540453,
644
+ "grad_norm": 0.5504106283187866,
645
+ "learning_rate": 4.254180602006689e-05,
646
+ "loss": 0.7118,
647
+ "step": 910
648
+ },
649
+ {
650
+ "epoch": 0.5954692556634305,
651
+ "grad_norm": 0.47307100892066956,
652
+ "learning_rate": 4.187290969899666e-05,
653
+ "loss": 0.7377,
654
+ "step": 920
655
+ },
656
+ {
657
+ "epoch": 0.6019417475728155,
658
+ "grad_norm": 0.5132865905761719,
659
+ "learning_rate": 4.1204013377926426e-05,
660
+ "loss": 0.7237,
661
+ "step": 930
662
+ },
663
+ {
664
+ "epoch": 0.6084142394822006,
665
+ "grad_norm": 0.5692107677459717,
666
+ "learning_rate": 4.053511705685619e-05,
667
+ "loss": 0.6933,
668
+ "step": 940
669
+ },
670
+ {
671
+ "epoch": 0.6148867313915858,
672
+ "grad_norm": 0.5088536143302917,
673
+ "learning_rate": 3.986622073578595e-05,
674
+ "loss": 0.7094,
675
+ "step": 950
676
+ },
677
+ {
678
+ "epoch": 0.6213592233009708,
679
+ "grad_norm": 0.47623276710510254,
680
+ "learning_rate": 3.919732441471572e-05,
681
+ "loss": 0.7219,
682
+ "step": 960
683
+ },
684
+ {
685
+ "epoch": 0.627831715210356,
686
+ "grad_norm": 0.48927319049835205,
687
+ "learning_rate": 3.852842809364549e-05,
688
+ "loss": 0.7324,
689
+ "step": 970
690
+ },
691
+ {
692
+ "epoch": 0.6343042071197411,
693
+ "grad_norm": 0.4764053523540497,
694
+ "learning_rate": 3.785953177257525e-05,
695
+ "loss": 0.7286,
696
+ "step": 980
697
+ },
698
+ {
699
+ "epoch": 0.6407766990291263,
700
+ "grad_norm": 0.4962717592716217,
701
+ "learning_rate": 3.7190635451505016e-05,
702
+ "loss": 0.7471,
703
+ "step": 990
704
+ },
705
+ {
706
+ "epoch": 0.6472491909385113,
707
+ "grad_norm": 0.5014144778251648,
708
+ "learning_rate": 3.6521739130434786e-05,
709
+ "loss": 0.6966,
710
+ "step": 1000
711
+ },
712
+ {
713
+ "epoch": 0.6537216828478964,
714
+ "grad_norm": 0.44654181599617004,
715
+ "learning_rate": 3.585284280936455e-05,
716
+ "loss": 0.7057,
717
+ "step": 1010
718
+ },
719
+ {
720
+ "epoch": 0.6601941747572816,
721
+ "grad_norm": 0.49160680174827576,
722
+ "learning_rate": 3.5183946488294314e-05,
723
+ "loss": 0.7016,
724
+ "step": 1020
725
+ },
726
+ {
727
+ "epoch": 0.6666666666666666,
728
+ "grad_norm": 0.5010440349578857,
729
+ "learning_rate": 3.4515050167224085e-05,
730
+ "loss": 0.7132,
731
+ "step": 1030
732
+ },
733
+ {
734
+ "epoch": 0.6731391585760518,
735
+ "grad_norm": 0.4979737102985382,
736
+ "learning_rate": 3.384615384615385e-05,
737
+ "loss": 0.7243,
738
+ "step": 1040
739
+ },
740
+ {
741
+ "epoch": 0.6796116504854369,
742
+ "grad_norm": 0.45698127150535583,
743
+ "learning_rate": 3.317725752508362e-05,
744
+ "loss": 0.7535,
745
+ "step": 1050
746
+ },
747
+ {
748
+ "epoch": 0.686084142394822,
749
+ "grad_norm": 0.47817227244377136,
750
+ "learning_rate": 3.250836120401338e-05,
751
+ "loss": 0.7749,
752
+ "step": 1060
753
+ },
754
+ {
755
+ "epoch": 0.6925566343042071,
756
+ "grad_norm": 0.4754548966884613,
757
+ "learning_rate": 3.183946488294314e-05,
758
+ "loss": 0.7229,
759
+ "step": 1070
760
+ },
761
+ {
762
+ "epoch": 0.6990291262135923,
763
+ "grad_norm": 0.5484136343002319,
764
+ "learning_rate": 3.117056856187291e-05,
765
+ "loss": 0.7191,
766
+ "step": 1080
767
+ },
768
+ {
769
+ "epoch": 0.7055016181229773,
770
+ "grad_norm": 0.5162729620933533,
771
+ "learning_rate": 3.0501672240802674e-05,
772
+ "loss": 0.7293,
773
+ "step": 1090
774
+ },
775
+ {
776
+ "epoch": 0.7119741100323624,
777
+ "grad_norm": 0.5177437663078308,
778
+ "learning_rate": 2.9832775919732442e-05,
779
+ "loss": 0.7254,
780
+ "step": 1100
781
+ },
782
+ {
783
+ "epoch": 0.7184466019417476,
784
+ "grad_norm": 0.5128371119499207,
785
+ "learning_rate": 2.916387959866221e-05,
786
+ "loss": 0.728,
787
+ "step": 1110
788
+ },
789
+ {
790
+ "epoch": 0.7249190938511327,
791
+ "grad_norm": 0.5296499729156494,
792
+ "learning_rate": 2.8494983277591973e-05,
793
+ "loss": 0.6976,
794
+ "step": 1120
795
+ },
796
+ {
797
+ "epoch": 0.7313915857605178,
798
+ "grad_norm": 0.5406925678253174,
799
+ "learning_rate": 2.782608695652174e-05,
800
+ "loss": 0.7313,
801
+ "step": 1130
802
+ },
803
+ {
804
+ "epoch": 0.7378640776699029,
805
+ "grad_norm": 0.5372524261474609,
806
+ "learning_rate": 2.7157190635451507e-05,
807
+ "loss": 0.7126,
808
+ "step": 1140
809
+ },
810
+ {
811
+ "epoch": 0.7443365695792881,
812
+ "grad_norm": 0.5130836367607117,
813
+ "learning_rate": 2.6488294314381275e-05,
814
+ "loss": 0.7178,
815
+ "step": 1150
816
+ },
817
+ {
818
+ "epoch": 0.7508090614886731,
819
+ "grad_norm": 0.636240541934967,
820
+ "learning_rate": 2.581939799331104e-05,
821
+ "loss": 0.6861,
822
+ "step": 1160
823
+ },
824
+ {
825
+ "epoch": 0.7572815533980582,
826
+ "grad_norm": 0.4897923767566681,
827
+ "learning_rate": 2.5150501672240806e-05,
828
+ "loss": 0.6889,
829
+ "step": 1170
830
+ },
831
+ {
832
+ "epoch": 0.7637540453074434,
833
+ "grad_norm": 0.4422953128814697,
834
+ "learning_rate": 2.448160535117057e-05,
835
+ "loss": 0.7314,
836
+ "step": 1180
837
+ },
838
+ {
839
+ "epoch": 0.7702265372168284,
840
+ "grad_norm": 0.4565708041191101,
841
+ "learning_rate": 2.3812709030100337e-05,
842
+ "loss": 0.7111,
843
+ "step": 1190
844
+ },
845
+ {
846
+ "epoch": 0.7766990291262136,
847
+ "grad_norm": 0.5054651498794556,
848
+ "learning_rate": 2.31438127090301e-05,
849
+ "loss": 0.742,
850
+ "step": 1200
851
+ },
852
+ {
853
+ "epoch": 0.7831715210355987,
854
+ "grad_norm": 0.5192699432373047,
855
+ "learning_rate": 2.2474916387959868e-05,
856
+ "loss": 0.7211,
857
+ "step": 1210
858
+ },
859
+ {
860
+ "epoch": 0.7896440129449838,
861
+ "grad_norm": 0.4859030842781067,
862
+ "learning_rate": 2.1806020066889632e-05,
863
+ "loss": 0.6962,
864
+ "step": 1220
865
+ },
866
+ {
867
+ "epoch": 0.7961165048543689,
868
+ "grad_norm": 0.4740995466709137,
869
+ "learning_rate": 2.11371237458194e-05,
870
+ "loss": 0.6868,
871
+ "step": 1230
872
+ },
873
+ {
874
+ "epoch": 0.8025889967637541,
875
+ "grad_norm": 0.4981386959552765,
876
+ "learning_rate": 2.0468227424749163e-05,
877
+ "loss": 0.7222,
878
+ "step": 1240
879
+ },
880
+ {
881
+ "epoch": 0.8090614886731392,
882
+ "grad_norm": 0.5139773488044739,
883
+ "learning_rate": 1.979933110367893e-05,
884
+ "loss": 0.7247,
885
+ "step": 1250
886
+ },
887
+ {
888
+ "epoch": 0.8155339805825242,
889
+ "grad_norm": 0.5492443442344666,
890
+ "learning_rate": 1.9130434782608697e-05,
891
+ "loss": 0.7195,
892
+ "step": 1260
893
+ },
894
+ {
895
+ "epoch": 0.8220064724919094,
896
+ "grad_norm": 0.46029913425445557,
897
+ "learning_rate": 1.8461538461538465e-05,
898
+ "loss": 0.6993,
899
+ "step": 1270
900
+ },
901
+ {
902
+ "epoch": 0.8284789644012945,
903
+ "grad_norm": 0.45861348509788513,
904
+ "learning_rate": 1.779264214046823e-05,
905
+ "loss": 0.7214,
906
+ "step": 1280
907
+ },
908
+ {
909
+ "epoch": 0.8349514563106796,
910
+ "grad_norm": 0.48776987195014954,
911
+ "learning_rate": 1.7123745819397992e-05,
912
+ "loss": 0.7554,
913
+ "step": 1290
914
+ },
915
+ {
916
+ "epoch": 0.8414239482200647,
917
+ "grad_norm": 0.44210392236709595,
918
+ "learning_rate": 1.645484949832776e-05,
919
+ "loss": 0.7155,
920
+ "step": 1300
921
+ },
922
+ {
923
+ "epoch": 0.8478964401294499,
924
+ "grad_norm": 0.45907357335090637,
925
+ "learning_rate": 1.5785953177257527e-05,
926
+ "loss": 0.7345,
927
+ "step": 1310
928
+ },
929
+ {
930
+ "epoch": 0.8543689320388349,
931
+ "grad_norm": 0.47504889965057373,
932
+ "learning_rate": 1.5117056856187292e-05,
933
+ "loss": 0.6969,
934
+ "step": 1320
935
+ },
936
+ {
937
+ "epoch": 0.86084142394822,
938
+ "grad_norm": 0.553654670715332,
939
+ "learning_rate": 1.4448160535117058e-05,
940
+ "loss": 0.6983,
941
+ "step": 1330
942
+ },
943
+ {
944
+ "epoch": 0.8673139158576052,
945
+ "grad_norm": 0.5136590003967285,
946
+ "learning_rate": 1.3779264214046825e-05,
947
+ "loss": 0.7423,
948
+ "step": 1340
949
+ },
950
+ {
951
+ "epoch": 0.8737864077669902,
952
+ "grad_norm": 0.44719818234443665,
953
+ "learning_rate": 1.3110367892976589e-05,
954
+ "loss": 0.6991,
955
+ "step": 1350
956
+ },
957
+ {
958
+ "epoch": 0.8802588996763754,
959
+ "grad_norm": 0.5215861797332764,
960
+ "learning_rate": 1.2441471571906355e-05,
961
+ "loss": 0.7171,
962
+ "step": 1360
963
+ },
964
+ {
965
+ "epoch": 0.8867313915857605,
966
+ "grad_norm": 0.48722243309020996,
967
+ "learning_rate": 1.177257525083612e-05,
968
+ "loss": 0.7072,
969
+ "step": 1370
970
+ },
971
+ {
972
+ "epoch": 0.8932038834951457,
973
+ "grad_norm": 0.46748003363609314,
974
+ "learning_rate": 1.1103678929765887e-05,
975
+ "loss": 0.7326,
976
+ "step": 1380
977
+ },
978
+ {
979
+ "epoch": 0.8996763754045307,
980
+ "grad_norm": 0.47341325879096985,
981
+ "learning_rate": 1.0434782608695651e-05,
982
+ "loss": 0.7437,
983
+ "step": 1390
984
+ },
985
+ {
986
+ "epoch": 0.9061488673139159,
987
+ "grad_norm": 0.7238942980766296,
988
+ "learning_rate": 9.765886287625419e-06,
989
+ "loss": 0.6979,
990
+ "step": 1400
991
+ },
992
+ {
993
+ "epoch": 0.912621359223301,
994
+ "grad_norm": 0.4732665717601776,
995
+ "learning_rate": 9.096989966555184e-06,
996
+ "loss": 0.7304,
997
+ "step": 1410
998
+ },
999
+ {
1000
+ "epoch": 0.919093851132686,
1001
+ "grad_norm": 0.5634979009628296,
1002
+ "learning_rate": 8.42809364548495e-06,
1003
+ "loss": 0.7154,
1004
+ "step": 1420
1005
+ },
1006
+ {
1007
+ "epoch": 0.9255663430420712,
1008
+ "grad_norm": 0.5499462485313416,
1009
+ "learning_rate": 7.759197324414715e-06,
1010
+ "loss": 0.7049,
1011
+ "step": 1430
1012
+ },
1013
+ {
1014
+ "epoch": 0.9320388349514563,
1015
+ "grad_norm": 0.4869629144668579,
1016
+ "learning_rate": 7.090301003344482e-06,
1017
+ "loss": 0.7045,
1018
+ "step": 1440
1019
+ },
1020
+ {
1021
+ "epoch": 0.9385113268608414,
1022
+ "grad_norm": 0.4725572168827057,
1023
+ "learning_rate": 6.421404682274247e-06,
1024
+ "loss": 0.7077,
1025
+ "step": 1450
1026
+ },
1027
+ {
1028
+ "epoch": 0.9449838187702265,
1029
+ "grad_norm": 0.46987947821617126,
1030
+ "learning_rate": 5.7525083612040135e-06,
1031
+ "loss": 0.7217,
1032
+ "step": 1460
1033
+ },
1034
+ {
1035
+ "epoch": 0.9514563106796117,
1036
+ "grad_norm": 0.5289896130561829,
1037
+ "learning_rate": 5.08361204013378e-06,
1038
+ "loss": 0.6953,
1039
+ "step": 1470
1040
+ },
1041
+ {
1042
+ "epoch": 0.9579288025889967,
1043
+ "grad_norm": 0.545849621295929,
1044
+ "learning_rate": 4.4147157190635455e-06,
1045
+ "loss": 0.7042,
1046
+ "step": 1480
1047
+ },
1048
+ {
1049
+ "epoch": 0.9644012944983819,
1050
+ "grad_norm": 0.556922197341919,
1051
+ "learning_rate": 3.745819397993311e-06,
1052
+ "loss": 0.7215,
1053
+ "step": 1490
1054
+ },
1055
+ {
1056
+ "epoch": 0.970873786407767,
1057
+ "grad_norm": 0.4752863645553589,
1058
+ "learning_rate": 3.0769230769230774e-06,
1059
+ "loss": 0.705,
1060
+ "step": 1500
1061
+ }
1062
+ ],
1063
+ "logging_steps": 10,
1064
+ "max_steps": 1545,
1065
+ "num_input_tokens_seen": 0,
1066
+ "num_train_epochs": 1,
1067
+ "save_steps": 500,
1068
+ "stateful_callbacks": {
1069
+ "TrainerControl": {
1070
+ "args": {
1071
+ "should_epoch_stop": false,
1072
+ "should_evaluate": false,
1073
+ "should_log": false,
1074
+ "should_save": true,
1075
+ "should_training_stop": false
1076
+ },
1077
+ "attributes": {}
1078
+ }
1079
+ },
1080
+ "total_flos": 5.4469720641542554e+17,
1081
+ "train_batch_size": 16,
1082
+ "trial_name": null,
1083
+ "trial_params": null
1084
+ }
checkpoint-1500/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3ca7cc1e4f7451b1bc8044e4e525255bb7182251d1959fd39f0cc8c4abb5629
3
+ size 5752
checkpoint-1545/README.md ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/mistral-7b-v0.3-bnb-4bit
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:unsloth/mistral-7b-v0.3-bnb-4bit
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ - unsloth
12
+ ---
13
+
14
+ # Model Card for Model ID
15
+
16
+ <!-- Provide a quick summary of what the model is/does. -->
17
+
18
+
19
+
20
+ ## Model Details
21
+
22
+ ### Model Description
23
+
24
+ <!-- Provide a longer summary of what this model is. -->
25
+
26
+
27
+
28
+ - **Developed by:** [More Information Needed]
29
+ - **Funded by [optional]:** [More Information Needed]
30
+ - **Shared by [optional]:** [More Information Needed]
31
+ - **Model type:** [More Information Needed]
32
+ - **Language(s) (NLP):** [More Information Needed]
33
+ - **License:** [More Information Needed]
34
+ - **Finetuned from model [optional]:** [More Information Needed]
35
+
36
+ ### Model Sources [optional]
37
+
38
+ <!-- Provide the basic links for the model. -->
39
+
40
+ - **Repository:** [More Information Needed]
41
+ - **Paper [optional]:** [More Information Needed]
42
+ - **Demo [optional]:** [More Information Needed]
43
+
44
+ ## Uses
45
+
46
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
47
+
48
+ ### Direct Use
49
+
50
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
51
+
52
+ [More Information Needed]
53
+
54
+ ### Downstream Use [optional]
55
+
56
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
57
+
58
+ [More Information Needed]
59
+
60
+ ### Out-of-Scope Use
61
+
62
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
63
+
64
+ [More Information Needed]
65
+
66
+ ## Bias, Risks, and Limitations
67
+
68
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
69
+
70
+ [More Information Needed]
71
+
72
+ ### Recommendations
73
+
74
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
75
+
76
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
77
+
78
+ ## How to Get Started with the Model
79
+
80
+ Use the code below to get started with the model.
81
+
82
+ [More Information Needed]
83
+
84
+ ## Training Details
85
+
86
+ ### Training Data
87
+
88
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
89
+
90
+ [More Information Needed]
91
+
92
+ ### Training Procedure
93
+
94
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
95
+
96
+ #### Preprocessing [optional]
97
+
98
+ [More Information Needed]
99
+
100
+
101
+ #### Training Hyperparameters
102
+
103
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
104
+
105
+ #### Speeds, Sizes, Times [optional]
106
+
107
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
108
+
109
+ [More Information Needed]
110
+
111
+ ## Evaluation
112
+
113
+ <!-- This section describes the evaluation protocols and provides the results. -->
114
+
115
+ ### Testing Data, Factors & Metrics
116
+
117
+ #### Testing Data
118
+
119
+ <!-- This should link to a Dataset Card if possible. -->
120
+
121
+ [More Information Needed]
122
+
123
+ #### Factors
124
+
125
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
126
+
127
+ [More Information Needed]
128
+
129
+ #### Metrics
130
+
131
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
132
+
133
+ [More Information Needed]
134
+
135
+ ### Results
136
+
137
+ [More Information Needed]
138
+
139
+ #### Summary
140
+
141
+
142
+
143
+ ## Model Examination [optional]
144
+
145
+ <!-- Relevant interpretability work for the model goes here -->
146
+
147
+ [More Information Needed]
148
+
149
+ ## Environmental Impact
150
+
151
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
152
+
153
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
154
+
155
+ - **Hardware Type:** [More Information Needed]
156
+ - **Hours used:** [More Information Needed]
157
+ - **Cloud Provider:** [More Information Needed]
158
+ - **Compute Region:** [More Information Needed]
159
+ - **Carbon Emitted:** [More Information Needed]
160
+
161
+ ## Technical Specifications [optional]
162
+
163
+ ### Model Architecture and Objective
164
+
165
+ [More Information Needed]
166
+
167
+ ### Compute Infrastructure
168
+
169
+ [More Information Needed]
170
+
171
+ #### Hardware
172
+
173
+ [More Information Needed]
174
+
175
+ #### Software
176
+
177
+ [More Information Needed]
178
+
179
+ ## Citation [optional]
180
+
181
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
182
+
183
+ **BibTeX:**
184
+
185
+ [More Information Needed]
186
+
187
+ **APA:**
188
+
189
+ [More Information Needed]
190
+
191
+ ## Glossary [optional]
192
+
193
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
194
+
195
+ [More Information Needed]
196
+
197
+ ## More Information [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Authors [optional]
202
+
203
+ [More Information Needed]
204
+
205
+ ## Model Card Contact
206
+
207
+ [More Information Needed]
208
+ ### Framework versions
209
+
210
+ - PEFT 0.18.0
checkpoint-1545/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "MistralForCausalLM",
7
+ "parent_library": "transformers.models.mistral.modeling_mistral",
8
+ "unsloth_fixed": true
9
+ },
10
+ "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
11
+ "bias": "none",
12
+ "corda_config": null,
13
+ "ensure_weight_tying": false,
14
+ "eva_config": null,
15
+ "exclude_modules": null,
16
+ "fan_in_fan_out": false,
17
+ "inference_mode": true,
18
+ "init_lora_weights": true,
19
+ "layer_replication": null,
20
+ "layers_pattern": null,
21
+ "layers_to_transform": null,
22
+ "loftq_config": {},
23
+ "lora_alpha": 32,
24
+ "lora_bias": false,
25
+ "lora_dropout": 0,
26
+ "megatron_config": null,
27
+ "megatron_core": "megatron.core",
28
+ "modules_to_save": null,
29
+ "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
+ "qalora_group_size": 16,
32
+ "r": 16,
33
+ "rank_pattern": {},
34
+ "revision": null,
35
+ "target_modules": [
36
+ "k_proj",
37
+ "gate_proj",
38
+ "o_proj",
39
+ "v_proj",
40
+ "q_proj",
41
+ "down_proj",
42
+ "up_proj"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": "CAUSAL_LM",
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
checkpoint-1545/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a03703be3951badd986bde85e958c3be749b7cc1f8e0ddf6c2a7dac0e8060a8b
3
+ size 167832240
checkpoint-1545/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fc47e024b3dac58745d1b5ae019c59fabcabecceb4d3d9c8a0085116350926f
3
+ size 85723732
checkpoint-1545/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:608fccb6c056ce88cdfd5355e6be2046f4d107a24a87c6b0d2c3b200ce6bb4ea
3
+ size 14244
checkpoint-1545/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3938e5834fb6d7b92270fc83a570825167a94c73c9876e613f5b81e1c8c2381e
3
+ size 1064
checkpoint-1545/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[control_768]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-1545/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1545/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
checkpoint-1545/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1545/trainer_state.json ADDED
@@ -0,0 +1,1112 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 1.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1545,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.006472491909385114,
14
+ "grad_norm": 2.0135042667388916,
15
+ "learning_rate": 1.8e-05,
16
+ "loss": 1.5311,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.012944983818770227,
21
+ "grad_norm": 1.0551478862762451,
22
+ "learning_rate": 3.8e-05,
23
+ "loss": 1.2229,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.019417475728155338,
28
+ "grad_norm": 1.101570725440979,
29
+ "learning_rate": 5.8e-05,
30
+ "loss": 0.9845,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.025889967637540454,
35
+ "grad_norm": 0.9952124357223511,
36
+ "learning_rate": 7.800000000000001e-05,
37
+ "loss": 0.8599,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.032362459546925564,
42
+ "grad_norm": 0.8940704464912415,
43
+ "learning_rate": 9.8e-05,
44
+ "loss": 0.8401,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.038834951456310676,
49
+ "grad_norm": 0.8487522602081299,
50
+ "learning_rate": 9.93979933110368e-05,
51
+ "loss": 0.776,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.045307443365695796,
56
+ "grad_norm": 0.7864758968353271,
57
+ "learning_rate": 9.872909698996656e-05,
58
+ "loss": 0.8153,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.05177993527508091,
63
+ "grad_norm": 0.7295967936515808,
64
+ "learning_rate": 9.806020066889633e-05,
65
+ "loss": 0.8251,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.05825242718446602,
70
+ "grad_norm": 0.7075174450874329,
71
+ "learning_rate": 9.739130434782609e-05,
72
+ "loss": 0.749,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.06472491909385113,
77
+ "grad_norm": 0.6422586441040039,
78
+ "learning_rate": 9.672240802675586e-05,
79
+ "loss": 0.7428,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.07119741100323625,
84
+ "grad_norm": 0.6745761632919312,
85
+ "learning_rate": 9.605351170568563e-05,
86
+ "loss": 0.7806,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.07766990291262135,
91
+ "grad_norm": 0.6769530177116394,
92
+ "learning_rate": 9.53846153846154e-05,
93
+ "loss": 0.796,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.08414239482200647,
98
+ "grad_norm": 0.5150454044342041,
99
+ "learning_rate": 9.471571906354515e-05,
100
+ "loss": 0.7486,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.09061488673139159,
105
+ "grad_norm": 0.6074182391166687,
106
+ "learning_rate": 9.404682274247491e-05,
107
+ "loss": 0.7783,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0970873786407767,
112
+ "grad_norm": 0.6745702624320984,
113
+ "learning_rate": 9.337792642140469e-05,
114
+ "loss": 0.7503,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.10355987055016182,
119
+ "grad_norm": 0.6088436841964722,
120
+ "learning_rate": 9.270903010033445e-05,
121
+ "loss": 0.7771,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.11003236245954692,
126
+ "grad_norm": 0.6351744532585144,
127
+ "learning_rate": 9.204013377926422e-05,
128
+ "loss": 0.7554,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.11650485436893204,
133
+ "grad_norm": 0.5281522274017334,
134
+ "learning_rate": 9.137123745819398e-05,
135
+ "loss": 0.7635,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.12297734627831715,
140
+ "grad_norm": 0.6833154559135437,
141
+ "learning_rate": 9.070234113712374e-05,
142
+ "loss": 0.753,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.12944983818770225,
147
+ "grad_norm": 0.5272049307823181,
148
+ "learning_rate": 9.003344481605351e-05,
149
+ "loss": 0.7384,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.13592233009708737,
154
+ "grad_norm": 0.6461851596832275,
155
+ "learning_rate": 8.936454849498329e-05,
156
+ "loss": 0.7548,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.1423948220064725,
161
+ "grad_norm": 0.5979950428009033,
162
+ "learning_rate": 8.869565217391305e-05,
163
+ "loss": 0.7534,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.1488673139158576,
168
+ "grad_norm": 0.6396514773368835,
169
+ "learning_rate": 8.802675585284281e-05,
170
+ "loss": 0.7436,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.1553398058252427,
175
+ "grad_norm": 0.6104384064674377,
176
+ "learning_rate": 8.735785953177258e-05,
177
+ "loss": 0.7345,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.16181229773462782,
182
+ "grad_norm": 0.5535375475883484,
183
+ "learning_rate": 8.668896321070234e-05,
184
+ "loss": 0.7293,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.16828478964401294,
189
+ "grad_norm": 0.5984886884689331,
190
+ "learning_rate": 8.602006688963212e-05,
191
+ "loss": 0.7173,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.17475728155339806,
196
+ "grad_norm": 0.517716646194458,
197
+ "learning_rate": 8.535117056856188e-05,
198
+ "loss": 0.7694,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.18122977346278318,
203
+ "grad_norm": 0.5802426934242249,
204
+ "learning_rate": 8.468227424749165e-05,
205
+ "loss": 0.7505,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.18770226537216828,
210
+ "grad_norm": 0.5289272665977478,
211
+ "learning_rate": 8.401337792642141e-05,
212
+ "loss": 0.7678,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.1941747572815534,
217
+ "grad_norm": 0.5293733477592468,
218
+ "learning_rate": 8.334448160535117e-05,
219
+ "loss": 0.7242,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.20064724919093851,
224
+ "grad_norm": 0.566146969795227,
225
+ "learning_rate": 8.267558528428094e-05,
226
+ "loss": 0.7248,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.20711974110032363,
231
+ "grad_norm": 0.5496853590011597,
232
+ "learning_rate": 8.200668896321071e-05,
233
+ "loss": 0.7588,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.21359223300970873,
238
+ "grad_norm": 0.5420507192611694,
239
+ "learning_rate": 8.133779264214048e-05,
240
+ "loss": 0.7673,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.22006472491909385,
245
+ "grad_norm": 0.582582950592041,
246
+ "learning_rate": 8.066889632107024e-05,
247
+ "loss": 0.7341,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.22653721682847897,
252
+ "grad_norm": 0.5332103371620178,
253
+ "learning_rate": 8e-05,
254
+ "loss": 0.7518,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.23300970873786409,
259
+ "grad_norm": 0.5132858157157898,
260
+ "learning_rate": 7.933110367892977e-05,
261
+ "loss": 0.752,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.23948220064724918,
266
+ "grad_norm": 0.5356752276420593,
267
+ "learning_rate": 7.866220735785953e-05,
268
+ "loss": 0.7432,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.2459546925566343,
273
+ "grad_norm": 0.5592623353004456,
274
+ "learning_rate": 7.79933110367893e-05,
275
+ "loss": 0.7396,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.2524271844660194,
280
+ "grad_norm": 0.5103404521942139,
281
+ "learning_rate": 7.732441471571906e-05,
282
+ "loss": 0.7253,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.2588996763754045,
287
+ "grad_norm": 0.5380160212516785,
288
+ "learning_rate": 7.665551839464883e-05,
289
+ "loss": 0.7464,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.26537216828478966,
294
+ "grad_norm": 0.5373779535293579,
295
+ "learning_rate": 7.598662207357859e-05,
296
+ "loss": 0.7377,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.27184466019417475,
301
+ "grad_norm": 0.523170530796051,
302
+ "learning_rate": 7.531772575250837e-05,
303
+ "loss": 0.7413,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.2783171521035599,
308
+ "grad_norm": 0.5368985533714294,
309
+ "learning_rate": 7.464882943143813e-05,
310
+ "loss": 0.7227,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.284789644012945,
315
+ "grad_norm": 0.587778627872467,
316
+ "learning_rate": 7.39799331103679e-05,
317
+ "loss": 0.7289,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.2912621359223301,
322
+ "grad_norm": 0.5471896529197693,
323
+ "learning_rate": 7.331103678929766e-05,
324
+ "loss": 0.7301,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.2977346278317152,
329
+ "grad_norm": 0.5446394085884094,
330
+ "learning_rate": 7.264214046822742e-05,
331
+ "loss": 0.7325,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.3042071197411003,
336
+ "grad_norm": 0.5761464238166809,
337
+ "learning_rate": 7.197324414715719e-05,
338
+ "loss": 0.7583,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.3106796116504854,
343
+ "grad_norm": 0.5082632899284363,
344
+ "learning_rate": 7.130434782608696e-05,
345
+ "loss": 0.7382,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.31715210355987056,
350
+ "grad_norm": 0.5441333651542664,
351
+ "learning_rate": 7.063545150501673e-05,
352
+ "loss": 0.7411,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.32362459546925565,
357
+ "grad_norm": 0.5039440393447876,
358
+ "learning_rate": 6.996655518394649e-05,
359
+ "loss": 0.7332,
360
+ "step": 500
361
+ },
362
+ {
363
+ "epoch": 0.3300970873786408,
364
+ "grad_norm": 0.5474230051040649,
365
+ "learning_rate": 6.929765886287626e-05,
366
+ "loss": 0.7216,
367
+ "step": 510
368
+ },
369
+ {
370
+ "epoch": 0.3365695792880259,
371
+ "grad_norm": 0.5532674789428711,
372
+ "learning_rate": 6.862876254180602e-05,
373
+ "loss": 0.7389,
374
+ "step": 520
375
+ },
376
+ {
377
+ "epoch": 0.343042071197411,
378
+ "grad_norm": 0.5631316900253296,
379
+ "learning_rate": 6.79598662207358e-05,
380
+ "loss": 0.7065,
381
+ "step": 530
382
+ },
383
+ {
384
+ "epoch": 0.34951456310679613,
385
+ "grad_norm": 0.45850899815559387,
386
+ "learning_rate": 6.729096989966556e-05,
387
+ "loss": 0.7472,
388
+ "step": 540
389
+ },
390
+ {
391
+ "epoch": 0.3559870550161812,
392
+ "grad_norm": 0.48253533244132996,
393
+ "learning_rate": 6.662207357859532e-05,
394
+ "loss": 0.7312,
395
+ "step": 550
396
+ },
397
+ {
398
+ "epoch": 0.36245954692556637,
399
+ "grad_norm": 0.6004955172538757,
400
+ "learning_rate": 6.595317725752509e-05,
401
+ "loss": 0.7681,
402
+ "step": 560
403
+ },
404
+ {
405
+ "epoch": 0.36893203883495146,
406
+ "grad_norm": 0.517497181892395,
407
+ "learning_rate": 6.528428093645485e-05,
408
+ "loss": 0.7581,
409
+ "step": 570
410
+ },
411
+ {
412
+ "epoch": 0.37540453074433655,
413
+ "grad_norm": 0.5473213195800781,
414
+ "learning_rate": 6.461538461538462e-05,
415
+ "loss": 0.7459,
416
+ "step": 580
417
+ },
418
+ {
419
+ "epoch": 0.3818770226537217,
420
+ "grad_norm": 0.5596595406532288,
421
+ "learning_rate": 6.39464882943144e-05,
422
+ "loss": 0.7322,
423
+ "step": 590
424
+ },
425
+ {
426
+ "epoch": 0.3883495145631068,
427
+ "grad_norm": 0.5507789254188538,
428
+ "learning_rate": 6.327759197324416e-05,
429
+ "loss": 0.7324,
430
+ "step": 600
431
+ },
432
+ {
433
+ "epoch": 0.3948220064724919,
434
+ "grad_norm": 0.5934524536132812,
435
+ "learning_rate": 6.260869565217392e-05,
436
+ "loss": 0.7072,
437
+ "step": 610
438
+ },
439
+ {
440
+ "epoch": 0.40129449838187703,
441
+ "grad_norm": 0.5208216905593872,
442
+ "learning_rate": 6.193979933110368e-05,
443
+ "loss": 0.755,
444
+ "step": 620
445
+ },
446
+ {
447
+ "epoch": 0.4077669902912621,
448
+ "grad_norm": 0.506360650062561,
449
+ "learning_rate": 6.127090301003345e-05,
450
+ "loss": 0.7244,
451
+ "step": 630
452
+ },
453
+ {
454
+ "epoch": 0.41423948220064727,
455
+ "grad_norm": 0.5485314130783081,
456
+ "learning_rate": 6.0602006688963206e-05,
457
+ "loss": 0.7284,
458
+ "step": 640
459
+ },
460
+ {
461
+ "epoch": 0.42071197411003236,
462
+ "grad_norm": 0.5979824066162109,
463
+ "learning_rate": 5.9933110367892977e-05,
464
+ "loss": 0.7065,
465
+ "step": 650
466
+ },
467
+ {
468
+ "epoch": 0.42718446601941745,
469
+ "grad_norm": 0.5200499892234802,
470
+ "learning_rate": 5.926421404682274e-05,
471
+ "loss": 0.7124,
472
+ "step": 660
473
+ },
474
+ {
475
+ "epoch": 0.4336569579288026,
476
+ "grad_norm": 0.5502509474754333,
477
+ "learning_rate": 5.859531772575251e-05,
478
+ "loss": 0.7493,
479
+ "step": 670
480
+ },
481
+ {
482
+ "epoch": 0.4401294498381877,
483
+ "grad_norm": 0.4699322581291199,
484
+ "learning_rate": 5.7926421404682275e-05,
485
+ "loss": 0.7588,
486
+ "step": 680
487
+ },
488
+ {
489
+ "epoch": 0.44660194174757284,
490
+ "grad_norm": 0.4809620678424835,
491
+ "learning_rate": 5.725752508361204e-05,
492
+ "loss": 0.7463,
493
+ "step": 690
494
+ },
495
+ {
496
+ "epoch": 0.45307443365695793,
497
+ "grad_norm": 0.5801950693130493,
498
+ "learning_rate": 5.658862876254181e-05,
499
+ "loss": 0.7206,
500
+ "step": 700
501
+ },
502
+ {
503
+ "epoch": 0.459546925566343,
504
+ "grad_norm": 0.5444105267524719,
505
+ "learning_rate": 5.591973244147157e-05,
506
+ "loss": 0.7197,
507
+ "step": 710
508
+ },
509
+ {
510
+ "epoch": 0.46601941747572817,
511
+ "grad_norm": 0.4769653379917145,
512
+ "learning_rate": 5.525083612040134e-05,
513
+ "loss": 0.7359,
514
+ "step": 720
515
+ },
516
+ {
517
+ "epoch": 0.47249190938511326,
518
+ "grad_norm": 0.5103979706764221,
519
+ "learning_rate": 5.458193979933111e-05,
520
+ "loss": 0.769,
521
+ "step": 730
522
+ },
523
+ {
524
+ "epoch": 0.47896440129449835,
525
+ "grad_norm": 0.543662965297699,
526
+ "learning_rate": 5.391304347826087e-05,
527
+ "loss": 0.7618,
528
+ "step": 740
529
+ },
530
+ {
531
+ "epoch": 0.4854368932038835,
532
+ "grad_norm": 0.5364013314247131,
533
+ "learning_rate": 5.3244147157190635e-05,
534
+ "loss": 0.72,
535
+ "step": 750
536
+ },
537
+ {
538
+ "epoch": 0.4919093851132686,
539
+ "grad_norm": 0.6091406345367432,
540
+ "learning_rate": 5.2575250836120406e-05,
541
+ "loss": 0.6965,
542
+ "step": 760
543
+ },
544
+ {
545
+ "epoch": 0.49838187702265374,
546
+ "grad_norm": 0.5977626442909241,
547
+ "learning_rate": 5.190635451505017e-05,
548
+ "loss": 0.7172,
549
+ "step": 770
550
+ },
551
+ {
552
+ "epoch": 0.5048543689320388,
553
+ "grad_norm": 0.45984843373298645,
554
+ "learning_rate": 5.1237458193979934e-05,
555
+ "loss": 0.7391,
556
+ "step": 780
557
+ },
558
+ {
559
+ "epoch": 0.511326860841424,
560
+ "grad_norm": 0.49561065435409546,
561
+ "learning_rate": 5.0568561872909704e-05,
562
+ "loss": 0.7526,
563
+ "step": 790
564
+ },
565
+ {
566
+ "epoch": 0.517799352750809,
567
+ "grad_norm": 0.4905160367488861,
568
+ "learning_rate": 4.989966555183947e-05,
569
+ "loss": 0.7278,
570
+ "step": 800
571
+ },
572
+ {
573
+ "epoch": 0.5242718446601942,
574
+ "grad_norm": 0.5338721871376038,
575
+ "learning_rate": 4.923076923076924e-05,
576
+ "loss": 0.7297,
577
+ "step": 810
578
+ },
579
+ {
580
+ "epoch": 0.5307443365695793,
581
+ "grad_norm": 0.5137319564819336,
582
+ "learning_rate": 4.8561872909698996e-05,
583
+ "loss": 0.7863,
584
+ "step": 820
585
+ },
586
+ {
587
+ "epoch": 0.5372168284789643,
588
+ "grad_norm": 0.4649944603443146,
589
+ "learning_rate": 4.789297658862876e-05,
590
+ "loss": 0.7525,
591
+ "step": 830
592
+ },
593
+ {
594
+ "epoch": 0.5436893203883495,
595
+ "grad_norm": 0.5467334985733032,
596
+ "learning_rate": 4.722408026755853e-05,
597
+ "loss": 0.7144,
598
+ "step": 840
599
+ },
600
+ {
601
+ "epoch": 0.5501618122977346,
602
+ "grad_norm": 0.47389763593673706,
603
+ "learning_rate": 4.6555183946488294e-05,
604
+ "loss": 0.7217,
605
+ "step": 850
606
+ },
607
+ {
608
+ "epoch": 0.5566343042071198,
609
+ "grad_norm": 0.45604580640792847,
610
+ "learning_rate": 4.5886287625418065e-05,
611
+ "loss": 0.7243,
612
+ "step": 860
613
+ },
614
+ {
615
+ "epoch": 0.5631067961165048,
616
+ "grad_norm": 0.548736572265625,
617
+ "learning_rate": 4.521739130434783e-05,
618
+ "loss": 0.7078,
619
+ "step": 870
620
+ },
621
+ {
622
+ "epoch": 0.56957928802589,
623
+ "grad_norm": 0.5387297868728638,
624
+ "learning_rate": 4.454849498327759e-05,
625
+ "loss": 0.6778,
626
+ "step": 880
627
+ },
628
+ {
629
+ "epoch": 0.5760517799352751,
630
+ "grad_norm": 0.5060684680938721,
631
+ "learning_rate": 4.3879598662207363e-05,
632
+ "loss": 0.7088,
633
+ "step": 890
634
+ },
635
+ {
636
+ "epoch": 0.5825242718446602,
637
+ "grad_norm": 0.5241347551345825,
638
+ "learning_rate": 4.321070234113713e-05,
639
+ "loss": 0.7232,
640
+ "step": 900
641
+ },
642
+ {
643
+ "epoch": 0.5889967637540453,
644
+ "grad_norm": 0.5504106283187866,
645
+ "learning_rate": 4.254180602006689e-05,
646
+ "loss": 0.7118,
647
+ "step": 910
648
+ },
649
+ {
650
+ "epoch": 0.5954692556634305,
651
+ "grad_norm": 0.47307100892066956,
652
+ "learning_rate": 4.187290969899666e-05,
653
+ "loss": 0.7377,
654
+ "step": 920
655
+ },
656
+ {
657
+ "epoch": 0.6019417475728155,
658
+ "grad_norm": 0.5132865905761719,
659
+ "learning_rate": 4.1204013377926426e-05,
660
+ "loss": 0.7237,
661
+ "step": 930
662
+ },
663
+ {
664
+ "epoch": 0.6084142394822006,
665
+ "grad_norm": 0.5692107677459717,
666
+ "learning_rate": 4.053511705685619e-05,
667
+ "loss": 0.6933,
668
+ "step": 940
669
+ },
670
+ {
671
+ "epoch": 0.6148867313915858,
672
+ "grad_norm": 0.5088536143302917,
673
+ "learning_rate": 3.986622073578595e-05,
674
+ "loss": 0.7094,
675
+ "step": 950
676
+ },
677
+ {
678
+ "epoch": 0.6213592233009708,
679
+ "grad_norm": 0.47623276710510254,
680
+ "learning_rate": 3.919732441471572e-05,
681
+ "loss": 0.7219,
682
+ "step": 960
683
+ },
684
+ {
685
+ "epoch": 0.627831715210356,
686
+ "grad_norm": 0.48927319049835205,
687
+ "learning_rate": 3.852842809364549e-05,
688
+ "loss": 0.7324,
689
+ "step": 970
690
+ },
691
+ {
692
+ "epoch": 0.6343042071197411,
693
+ "grad_norm": 0.4764053523540497,
694
+ "learning_rate": 3.785953177257525e-05,
695
+ "loss": 0.7286,
696
+ "step": 980
697
+ },
698
+ {
699
+ "epoch": 0.6407766990291263,
700
+ "grad_norm": 0.4962717592716217,
701
+ "learning_rate": 3.7190635451505016e-05,
702
+ "loss": 0.7471,
703
+ "step": 990
704
+ },
705
+ {
706
+ "epoch": 0.6472491909385113,
707
+ "grad_norm": 0.5014144778251648,
708
+ "learning_rate": 3.6521739130434786e-05,
709
+ "loss": 0.6966,
710
+ "step": 1000
711
+ },
712
+ {
713
+ "epoch": 0.6537216828478964,
714
+ "grad_norm": 0.44654181599617004,
715
+ "learning_rate": 3.585284280936455e-05,
716
+ "loss": 0.7057,
717
+ "step": 1010
718
+ },
719
+ {
720
+ "epoch": 0.6601941747572816,
721
+ "grad_norm": 0.49160680174827576,
722
+ "learning_rate": 3.5183946488294314e-05,
723
+ "loss": 0.7016,
724
+ "step": 1020
725
+ },
726
+ {
727
+ "epoch": 0.6666666666666666,
728
+ "grad_norm": 0.5010440349578857,
729
+ "learning_rate": 3.4515050167224085e-05,
730
+ "loss": 0.7132,
731
+ "step": 1030
732
+ },
733
+ {
734
+ "epoch": 0.6731391585760518,
735
+ "grad_norm": 0.4979737102985382,
736
+ "learning_rate": 3.384615384615385e-05,
737
+ "loss": 0.7243,
738
+ "step": 1040
739
+ },
740
+ {
741
+ "epoch": 0.6796116504854369,
742
+ "grad_norm": 0.45698127150535583,
743
+ "learning_rate": 3.317725752508362e-05,
744
+ "loss": 0.7535,
745
+ "step": 1050
746
+ },
747
+ {
748
+ "epoch": 0.686084142394822,
749
+ "grad_norm": 0.47817227244377136,
750
+ "learning_rate": 3.250836120401338e-05,
751
+ "loss": 0.7749,
752
+ "step": 1060
753
+ },
754
+ {
755
+ "epoch": 0.6925566343042071,
756
+ "grad_norm": 0.4754548966884613,
757
+ "learning_rate": 3.183946488294314e-05,
758
+ "loss": 0.7229,
759
+ "step": 1070
760
+ },
761
+ {
762
+ "epoch": 0.6990291262135923,
763
+ "grad_norm": 0.5484136343002319,
764
+ "learning_rate": 3.117056856187291e-05,
765
+ "loss": 0.7191,
766
+ "step": 1080
767
+ },
768
+ {
769
+ "epoch": 0.7055016181229773,
770
+ "grad_norm": 0.5162729620933533,
771
+ "learning_rate": 3.0501672240802674e-05,
772
+ "loss": 0.7293,
773
+ "step": 1090
774
+ },
775
+ {
776
+ "epoch": 0.7119741100323624,
777
+ "grad_norm": 0.5177437663078308,
778
+ "learning_rate": 2.9832775919732442e-05,
779
+ "loss": 0.7254,
780
+ "step": 1100
781
+ },
782
+ {
783
+ "epoch": 0.7184466019417476,
784
+ "grad_norm": 0.5128371119499207,
785
+ "learning_rate": 2.916387959866221e-05,
786
+ "loss": 0.728,
787
+ "step": 1110
788
+ },
789
+ {
790
+ "epoch": 0.7249190938511327,
791
+ "grad_norm": 0.5296499729156494,
792
+ "learning_rate": 2.8494983277591973e-05,
793
+ "loss": 0.6976,
794
+ "step": 1120
795
+ },
796
+ {
797
+ "epoch": 0.7313915857605178,
798
+ "grad_norm": 0.5406925678253174,
799
+ "learning_rate": 2.782608695652174e-05,
800
+ "loss": 0.7313,
801
+ "step": 1130
802
+ },
803
+ {
804
+ "epoch": 0.7378640776699029,
805
+ "grad_norm": 0.5372524261474609,
806
+ "learning_rate": 2.7157190635451507e-05,
807
+ "loss": 0.7126,
808
+ "step": 1140
809
+ },
810
+ {
811
+ "epoch": 0.7443365695792881,
812
+ "grad_norm": 0.5130836367607117,
813
+ "learning_rate": 2.6488294314381275e-05,
814
+ "loss": 0.7178,
815
+ "step": 1150
816
+ },
817
+ {
818
+ "epoch": 0.7508090614886731,
819
+ "grad_norm": 0.636240541934967,
820
+ "learning_rate": 2.581939799331104e-05,
821
+ "loss": 0.6861,
822
+ "step": 1160
823
+ },
824
+ {
825
+ "epoch": 0.7572815533980582,
826
+ "grad_norm": 0.4897923767566681,
827
+ "learning_rate": 2.5150501672240806e-05,
828
+ "loss": 0.6889,
829
+ "step": 1170
830
+ },
831
+ {
832
+ "epoch": 0.7637540453074434,
833
+ "grad_norm": 0.4422953128814697,
834
+ "learning_rate": 2.448160535117057e-05,
835
+ "loss": 0.7314,
836
+ "step": 1180
837
+ },
838
+ {
839
+ "epoch": 0.7702265372168284,
840
+ "grad_norm": 0.4565708041191101,
841
+ "learning_rate": 2.3812709030100337e-05,
842
+ "loss": 0.7111,
843
+ "step": 1190
844
+ },
845
+ {
846
+ "epoch": 0.7766990291262136,
847
+ "grad_norm": 0.5054651498794556,
848
+ "learning_rate": 2.31438127090301e-05,
849
+ "loss": 0.742,
850
+ "step": 1200
851
+ },
852
+ {
853
+ "epoch": 0.7831715210355987,
854
+ "grad_norm": 0.5192699432373047,
855
+ "learning_rate": 2.2474916387959868e-05,
856
+ "loss": 0.7211,
857
+ "step": 1210
858
+ },
859
+ {
860
+ "epoch": 0.7896440129449838,
861
+ "grad_norm": 0.4859030842781067,
862
+ "learning_rate": 2.1806020066889632e-05,
863
+ "loss": 0.6962,
864
+ "step": 1220
865
+ },
866
+ {
867
+ "epoch": 0.7961165048543689,
868
+ "grad_norm": 0.4740995466709137,
869
+ "learning_rate": 2.11371237458194e-05,
870
+ "loss": 0.6868,
871
+ "step": 1230
872
+ },
873
+ {
874
+ "epoch": 0.8025889967637541,
875
+ "grad_norm": 0.4981386959552765,
876
+ "learning_rate": 2.0468227424749163e-05,
877
+ "loss": 0.7222,
878
+ "step": 1240
879
+ },
880
+ {
881
+ "epoch": 0.8090614886731392,
882
+ "grad_norm": 0.5139773488044739,
883
+ "learning_rate": 1.979933110367893e-05,
884
+ "loss": 0.7247,
885
+ "step": 1250
886
+ },
887
+ {
888
+ "epoch": 0.8155339805825242,
889
+ "grad_norm": 0.5492443442344666,
890
+ "learning_rate": 1.9130434782608697e-05,
891
+ "loss": 0.7195,
892
+ "step": 1260
893
+ },
894
+ {
895
+ "epoch": 0.8220064724919094,
896
+ "grad_norm": 0.46029913425445557,
897
+ "learning_rate": 1.8461538461538465e-05,
898
+ "loss": 0.6993,
899
+ "step": 1270
900
+ },
901
+ {
902
+ "epoch": 0.8284789644012945,
903
+ "grad_norm": 0.45861348509788513,
904
+ "learning_rate": 1.779264214046823e-05,
905
+ "loss": 0.7214,
906
+ "step": 1280
907
+ },
908
+ {
909
+ "epoch": 0.8349514563106796,
910
+ "grad_norm": 0.48776987195014954,
911
+ "learning_rate": 1.7123745819397992e-05,
912
+ "loss": 0.7554,
913
+ "step": 1290
914
+ },
915
+ {
916
+ "epoch": 0.8414239482200647,
917
+ "grad_norm": 0.44210392236709595,
918
+ "learning_rate": 1.645484949832776e-05,
919
+ "loss": 0.7155,
920
+ "step": 1300
921
+ },
922
+ {
923
+ "epoch": 0.8478964401294499,
924
+ "grad_norm": 0.45907357335090637,
925
+ "learning_rate": 1.5785953177257527e-05,
926
+ "loss": 0.7345,
927
+ "step": 1310
928
+ },
929
+ {
930
+ "epoch": 0.8543689320388349,
931
+ "grad_norm": 0.47504889965057373,
932
+ "learning_rate": 1.5117056856187292e-05,
933
+ "loss": 0.6969,
934
+ "step": 1320
935
+ },
936
+ {
937
+ "epoch": 0.86084142394822,
938
+ "grad_norm": 0.553654670715332,
939
+ "learning_rate": 1.4448160535117058e-05,
940
+ "loss": 0.6983,
941
+ "step": 1330
942
+ },
943
+ {
944
+ "epoch": 0.8673139158576052,
945
+ "grad_norm": 0.5136590003967285,
946
+ "learning_rate": 1.3779264214046825e-05,
947
+ "loss": 0.7423,
948
+ "step": 1340
949
+ },
950
+ {
951
+ "epoch": 0.8737864077669902,
952
+ "grad_norm": 0.44719818234443665,
953
+ "learning_rate": 1.3110367892976589e-05,
954
+ "loss": 0.6991,
955
+ "step": 1350
956
+ },
957
+ {
958
+ "epoch": 0.8802588996763754,
959
+ "grad_norm": 0.5215861797332764,
960
+ "learning_rate": 1.2441471571906355e-05,
961
+ "loss": 0.7171,
962
+ "step": 1360
963
+ },
964
+ {
965
+ "epoch": 0.8867313915857605,
966
+ "grad_norm": 0.48722243309020996,
967
+ "learning_rate": 1.177257525083612e-05,
968
+ "loss": 0.7072,
969
+ "step": 1370
970
+ },
971
+ {
972
+ "epoch": 0.8932038834951457,
973
+ "grad_norm": 0.46748003363609314,
974
+ "learning_rate": 1.1103678929765887e-05,
975
+ "loss": 0.7326,
976
+ "step": 1380
977
+ },
978
+ {
979
+ "epoch": 0.8996763754045307,
980
+ "grad_norm": 0.47341325879096985,
981
+ "learning_rate": 1.0434782608695651e-05,
982
+ "loss": 0.7437,
983
+ "step": 1390
984
+ },
985
+ {
986
+ "epoch": 0.9061488673139159,
987
+ "grad_norm": 0.7238942980766296,
988
+ "learning_rate": 9.765886287625419e-06,
989
+ "loss": 0.6979,
990
+ "step": 1400
991
+ },
992
+ {
993
+ "epoch": 0.912621359223301,
994
+ "grad_norm": 0.4732665717601776,
995
+ "learning_rate": 9.096989966555184e-06,
996
+ "loss": 0.7304,
997
+ "step": 1410
998
+ },
999
+ {
1000
+ "epoch": 0.919093851132686,
1001
+ "grad_norm": 0.5634979009628296,
1002
+ "learning_rate": 8.42809364548495e-06,
1003
+ "loss": 0.7154,
1004
+ "step": 1420
1005
+ },
1006
+ {
1007
+ "epoch": 0.9255663430420712,
1008
+ "grad_norm": 0.5499462485313416,
1009
+ "learning_rate": 7.759197324414715e-06,
1010
+ "loss": 0.7049,
1011
+ "step": 1430
1012
+ },
1013
+ {
1014
+ "epoch": 0.9320388349514563,
1015
+ "grad_norm": 0.4869629144668579,
1016
+ "learning_rate": 7.090301003344482e-06,
1017
+ "loss": 0.7045,
1018
+ "step": 1440
1019
+ },
1020
+ {
1021
+ "epoch": 0.9385113268608414,
1022
+ "grad_norm": 0.4725572168827057,
1023
+ "learning_rate": 6.421404682274247e-06,
1024
+ "loss": 0.7077,
1025
+ "step": 1450
1026
+ },
1027
+ {
1028
+ "epoch": 0.9449838187702265,
1029
+ "grad_norm": 0.46987947821617126,
1030
+ "learning_rate": 5.7525083612040135e-06,
1031
+ "loss": 0.7217,
1032
+ "step": 1460
1033
+ },
1034
+ {
1035
+ "epoch": 0.9514563106796117,
1036
+ "grad_norm": 0.5289896130561829,
1037
+ "learning_rate": 5.08361204013378e-06,
1038
+ "loss": 0.6953,
1039
+ "step": 1470
1040
+ },
1041
+ {
1042
+ "epoch": 0.9579288025889967,
1043
+ "grad_norm": 0.545849621295929,
1044
+ "learning_rate": 4.4147157190635455e-06,
1045
+ "loss": 0.7042,
1046
+ "step": 1480
1047
+ },
1048
+ {
1049
+ "epoch": 0.9644012944983819,
1050
+ "grad_norm": 0.556922197341919,
1051
+ "learning_rate": 3.745819397993311e-06,
1052
+ "loss": 0.7215,
1053
+ "step": 1490
1054
+ },
1055
+ {
1056
+ "epoch": 0.970873786407767,
1057
+ "grad_norm": 0.4752863645553589,
1058
+ "learning_rate": 3.0769230769230774e-06,
1059
+ "loss": 0.705,
1060
+ "step": 1500
1061
+ },
1062
+ {
1063
+ "epoch": 0.9773462783171522,
1064
+ "grad_norm": 0.461245059967041,
1065
+ "learning_rate": 2.408026755852843e-06,
1066
+ "loss": 0.6679,
1067
+ "step": 1510
1068
+ },
1069
+ {
1070
+ "epoch": 0.9838187702265372,
1071
+ "grad_norm": 0.45782899856567383,
1072
+ "learning_rate": 1.7391304347826088e-06,
1073
+ "loss": 0.7341,
1074
+ "step": 1520
1075
+ },
1076
+ {
1077
+ "epoch": 0.9902912621359223,
1078
+ "grad_norm": 0.5337499976158142,
1079
+ "learning_rate": 1.0702341137123745e-06,
1080
+ "loss": 0.7053,
1081
+ "step": 1530
1082
+ },
1083
+ {
1084
+ "epoch": 0.9967637540453075,
1085
+ "grad_norm": 0.4995047152042389,
1086
+ "learning_rate": 4.0133779264214045e-07,
1087
+ "loss": 0.7352,
1088
+ "step": 1540
1089
+ }
1090
+ ],
1091
+ "logging_steps": 10,
1092
+ "max_steps": 1545,
1093
+ "num_input_tokens_seen": 0,
1094
+ "num_train_epochs": 1,
1095
+ "save_steps": 500,
1096
+ "stateful_callbacks": {
1097
+ "TrainerControl": {
1098
+ "args": {
1099
+ "should_epoch_stop": false,
1100
+ "should_evaluate": false,
1101
+ "should_log": false,
1102
+ "should_save": true,
1103
+ "should_training_stop": true
1104
+ },
1105
+ "attributes": {}
1106
+ }
1107
+ },
1108
+ "total_flos": 5.6084401061206426e+17,
1109
+ "train_batch_size": 16,
1110
+ "trial_name": null,
1111
+ "trial_params": null
1112
+ }
checkpoint-1545/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3ca7cc1e4f7451b1bc8044e4e525255bb7182251d1959fd39f0cc8c4abb5629
3
+ size 5752
checkpoint-500/README.md ADDED
@@ -0,0 +1,210 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: unsloth/mistral-7b-v0.3-bnb-4bit
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:unsloth/mistral-7b-v0.3-bnb-4bit
7
+ - lora
8
+ - sft
9
+ - transformers
10
+ - trl
11
+ - unsloth
12
+ ---
13
+
14
+ # Model Card for Model ID
15
+
16
+ <!-- Provide a quick summary of what the model is/does. -->
17
+
18
+
19
+
20
+ ## Model Details
21
+
22
+ ### Model Description
23
+
24
+ <!-- Provide a longer summary of what this model is. -->
25
+
26
+
27
+
28
+ - **Developed by:** [More Information Needed]
29
+ - **Funded by [optional]:** [More Information Needed]
30
+ - **Shared by [optional]:** [More Information Needed]
31
+ - **Model type:** [More Information Needed]
32
+ - **Language(s) (NLP):** [More Information Needed]
33
+ - **License:** [More Information Needed]
34
+ - **Finetuned from model [optional]:** [More Information Needed]
35
+
36
+ ### Model Sources [optional]
37
+
38
+ <!-- Provide the basic links for the model. -->
39
+
40
+ - **Repository:** [More Information Needed]
41
+ - **Paper [optional]:** [More Information Needed]
42
+ - **Demo [optional]:** [More Information Needed]
43
+
44
+ ## Uses
45
+
46
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
47
+
48
+ ### Direct Use
49
+
50
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
51
+
52
+ [More Information Needed]
53
+
54
+ ### Downstream Use [optional]
55
+
56
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
57
+
58
+ [More Information Needed]
59
+
60
+ ### Out-of-Scope Use
61
+
62
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
63
+
64
+ [More Information Needed]
65
+
66
+ ## Bias, Risks, and Limitations
67
+
68
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
69
+
70
+ [More Information Needed]
71
+
72
+ ### Recommendations
73
+
74
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
75
+
76
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
77
+
78
+ ## How to Get Started with the Model
79
+
80
+ Use the code below to get started with the model.
81
+
82
+ [More Information Needed]
83
+
84
+ ## Training Details
85
+
86
+ ### Training Data
87
+
88
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
89
+
90
+ [More Information Needed]
91
+
92
+ ### Training Procedure
93
+
94
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
95
+
96
+ #### Preprocessing [optional]
97
+
98
+ [More Information Needed]
99
+
100
+
101
+ #### Training Hyperparameters
102
+
103
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
104
+
105
+ #### Speeds, Sizes, Times [optional]
106
+
107
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
108
+
109
+ [More Information Needed]
110
+
111
+ ## Evaluation
112
+
113
+ <!-- This section describes the evaluation protocols and provides the results. -->
114
+
115
+ ### Testing Data, Factors & Metrics
116
+
117
+ #### Testing Data
118
+
119
+ <!-- This should link to a Dataset Card if possible. -->
120
+
121
+ [More Information Needed]
122
+
123
+ #### Factors
124
+
125
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
126
+
127
+ [More Information Needed]
128
+
129
+ #### Metrics
130
+
131
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
132
+
133
+ [More Information Needed]
134
+
135
+ ### Results
136
+
137
+ [More Information Needed]
138
+
139
+ #### Summary
140
+
141
+
142
+
143
+ ## Model Examination [optional]
144
+
145
+ <!-- Relevant interpretability work for the model goes here -->
146
+
147
+ [More Information Needed]
148
+
149
+ ## Environmental Impact
150
+
151
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
152
+
153
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
154
+
155
+ - **Hardware Type:** [More Information Needed]
156
+ - **Hours used:** [More Information Needed]
157
+ - **Cloud Provider:** [More Information Needed]
158
+ - **Compute Region:** [More Information Needed]
159
+ - **Carbon Emitted:** [More Information Needed]
160
+
161
+ ## Technical Specifications [optional]
162
+
163
+ ### Model Architecture and Objective
164
+
165
+ [More Information Needed]
166
+
167
+ ### Compute Infrastructure
168
+
169
+ [More Information Needed]
170
+
171
+ #### Hardware
172
+
173
+ [More Information Needed]
174
+
175
+ #### Software
176
+
177
+ [More Information Needed]
178
+
179
+ ## Citation [optional]
180
+
181
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
182
+
183
+ **BibTeX:**
184
+
185
+ [More Information Needed]
186
+
187
+ **APA:**
188
+
189
+ [More Information Needed]
190
+
191
+ ## Glossary [optional]
192
+
193
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
194
+
195
+ [More Information Needed]
196
+
197
+ ## More Information [optional]
198
+
199
+ [More Information Needed]
200
+
201
+ ## Model Card Authors [optional]
202
+
203
+ [More Information Needed]
204
+
205
+ ## Model Card Contact
206
+
207
+ [More Information Needed]
208
+ ### Framework versions
209
+
210
+ - PEFT 0.18.0
checkpoint-500/adapter_config.json ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": {
6
+ "base_model_class": "MistralForCausalLM",
7
+ "parent_library": "transformers.models.mistral.modeling_mistral",
8
+ "unsloth_fixed": true
9
+ },
10
+ "base_model_name_or_path": "unsloth/mistral-7b-v0.3-bnb-4bit",
11
+ "bias": "none",
12
+ "corda_config": null,
13
+ "ensure_weight_tying": false,
14
+ "eva_config": null,
15
+ "exclude_modules": null,
16
+ "fan_in_fan_out": false,
17
+ "inference_mode": true,
18
+ "init_lora_weights": true,
19
+ "layer_replication": null,
20
+ "layers_pattern": null,
21
+ "layers_to_transform": null,
22
+ "loftq_config": {},
23
+ "lora_alpha": 32,
24
+ "lora_bias": false,
25
+ "lora_dropout": 0,
26
+ "megatron_config": null,
27
+ "megatron_core": "megatron.core",
28
+ "modules_to_save": null,
29
+ "peft_type": "LORA",
30
+ "peft_version": "0.18.0",
31
+ "qalora_group_size": 16,
32
+ "r": 16,
33
+ "rank_pattern": {},
34
+ "revision": null,
35
+ "target_modules": [
36
+ "k_proj",
37
+ "gate_proj",
38
+ "o_proj",
39
+ "v_proj",
40
+ "q_proj",
41
+ "down_proj",
42
+ "up_proj"
43
+ ],
44
+ "target_parameters": null,
45
+ "task_type": "CAUSAL_LM",
46
+ "trainable_token_indices": null,
47
+ "use_dora": false,
48
+ "use_qalora": false,
49
+ "use_rslora": false
50
+ }
checkpoint-500/adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ab362747e291a9f164191d1bff01fbf8fdccea0ac254b85e96fbcce72506d62
3
+ size 167832240
checkpoint-500/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82435f7758710e1a50a60db848e3468a6fc0088a1eae673dcdf6f3f279d150bb
3
+ size 85723732
checkpoint-500/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:608fccb6c056ce88cdfd5355e6be2046f4d107a24a87c6b0d2c3b200ce6bb4ea
3
+ size 14244
checkpoint-500/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f34a1507a9a22bae2552c642334196ef352b592d1f99eb28a2b3765bb56a7f7e
3
+ size 1064
checkpoint-500/special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[control_768]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
checkpoint-500/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-500/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37f00374dea48658ee8f5d0f21895b9bc55cb0103939607c8185bfd1c6ca1f89
3
+ size 587404
checkpoint-500/tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-500/trainer_state.json ADDED
@@ -0,0 +1,384 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 0.32362459546925565,
6
+ "eval_steps": 500,
7
+ "global_step": 500,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.006472491909385114,
14
+ "grad_norm": 2.0135042667388916,
15
+ "learning_rate": 1.8e-05,
16
+ "loss": 1.5311,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.012944983818770227,
21
+ "grad_norm": 1.0551478862762451,
22
+ "learning_rate": 3.8e-05,
23
+ "loss": 1.2229,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.019417475728155338,
28
+ "grad_norm": 1.101570725440979,
29
+ "learning_rate": 5.8e-05,
30
+ "loss": 0.9845,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.025889967637540454,
35
+ "grad_norm": 0.9952124357223511,
36
+ "learning_rate": 7.800000000000001e-05,
37
+ "loss": 0.8599,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.032362459546925564,
42
+ "grad_norm": 0.8940704464912415,
43
+ "learning_rate": 9.8e-05,
44
+ "loss": 0.8401,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.038834951456310676,
49
+ "grad_norm": 0.8487522602081299,
50
+ "learning_rate": 9.93979933110368e-05,
51
+ "loss": 0.776,
52
+ "step": 60
53
+ },
54
+ {
55
+ "epoch": 0.045307443365695796,
56
+ "grad_norm": 0.7864758968353271,
57
+ "learning_rate": 9.872909698996656e-05,
58
+ "loss": 0.8153,
59
+ "step": 70
60
+ },
61
+ {
62
+ "epoch": 0.05177993527508091,
63
+ "grad_norm": 0.7295967936515808,
64
+ "learning_rate": 9.806020066889633e-05,
65
+ "loss": 0.8251,
66
+ "step": 80
67
+ },
68
+ {
69
+ "epoch": 0.05825242718446602,
70
+ "grad_norm": 0.7075174450874329,
71
+ "learning_rate": 9.739130434782609e-05,
72
+ "loss": 0.749,
73
+ "step": 90
74
+ },
75
+ {
76
+ "epoch": 0.06472491909385113,
77
+ "grad_norm": 0.6422586441040039,
78
+ "learning_rate": 9.672240802675586e-05,
79
+ "loss": 0.7428,
80
+ "step": 100
81
+ },
82
+ {
83
+ "epoch": 0.07119741100323625,
84
+ "grad_norm": 0.6745761632919312,
85
+ "learning_rate": 9.605351170568563e-05,
86
+ "loss": 0.7806,
87
+ "step": 110
88
+ },
89
+ {
90
+ "epoch": 0.07766990291262135,
91
+ "grad_norm": 0.6769530177116394,
92
+ "learning_rate": 9.53846153846154e-05,
93
+ "loss": 0.796,
94
+ "step": 120
95
+ },
96
+ {
97
+ "epoch": 0.08414239482200647,
98
+ "grad_norm": 0.5150454044342041,
99
+ "learning_rate": 9.471571906354515e-05,
100
+ "loss": 0.7486,
101
+ "step": 130
102
+ },
103
+ {
104
+ "epoch": 0.09061488673139159,
105
+ "grad_norm": 0.6074182391166687,
106
+ "learning_rate": 9.404682274247491e-05,
107
+ "loss": 0.7783,
108
+ "step": 140
109
+ },
110
+ {
111
+ "epoch": 0.0970873786407767,
112
+ "grad_norm": 0.6745702624320984,
113
+ "learning_rate": 9.337792642140469e-05,
114
+ "loss": 0.7503,
115
+ "step": 150
116
+ },
117
+ {
118
+ "epoch": 0.10355987055016182,
119
+ "grad_norm": 0.6088436841964722,
120
+ "learning_rate": 9.270903010033445e-05,
121
+ "loss": 0.7771,
122
+ "step": 160
123
+ },
124
+ {
125
+ "epoch": 0.11003236245954692,
126
+ "grad_norm": 0.6351744532585144,
127
+ "learning_rate": 9.204013377926422e-05,
128
+ "loss": 0.7554,
129
+ "step": 170
130
+ },
131
+ {
132
+ "epoch": 0.11650485436893204,
133
+ "grad_norm": 0.5281522274017334,
134
+ "learning_rate": 9.137123745819398e-05,
135
+ "loss": 0.7635,
136
+ "step": 180
137
+ },
138
+ {
139
+ "epoch": 0.12297734627831715,
140
+ "grad_norm": 0.6833154559135437,
141
+ "learning_rate": 9.070234113712374e-05,
142
+ "loss": 0.753,
143
+ "step": 190
144
+ },
145
+ {
146
+ "epoch": 0.12944983818770225,
147
+ "grad_norm": 0.5272049307823181,
148
+ "learning_rate": 9.003344481605351e-05,
149
+ "loss": 0.7384,
150
+ "step": 200
151
+ },
152
+ {
153
+ "epoch": 0.13592233009708737,
154
+ "grad_norm": 0.6461851596832275,
155
+ "learning_rate": 8.936454849498329e-05,
156
+ "loss": 0.7548,
157
+ "step": 210
158
+ },
159
+ {
160
+ "epoch": 0.1423948220064725,
161
+ "grad_norm": 0.5979950428009033,
162
+ "learning_rate": 8.869565217391305e-05,
163
+ "loss": 0.7534,
164
+ "step": 220
165
+ },
166
+ {
167
+ "epoch": 0.1488673139158576,
168
+ "grad_norm": 0.6396514773368835,
169
+ "learning_rate": 8.802675585284281e-05,
170
+ "loss": 0.7436,
171
+ "step": 230
172
+ },
173
+ {
174
+ "epoch": 0.1553398058252427,
175
+ "grad_norm": 0.6104384064674377,
176
+ "learning_rate": 8.735785953177258e-05,
177
+ "loss": 0.7345,
178
+ "step": 240
179
+ },
180
+ {
181
+ "epoch": 0.16181229773462782,
182
+ "grad_norm": 0.5535375475883484,
183
+ "learning_rate": 8.668896321070234e-05,
184
+ "loss": 0.7293,
185
+ "step": 250
186
+ },
187
+ {
188
+ "epoch": 0.16828478964401294,
189
+ "grad_norm": 0.5984886884689331,
190
+ "learning_rate": 8.602006688963212e-05,
191
+ "loss": 0.7173,
192
+ "step": 260
193
+ },
194
+ {
195
+ "epoch": 0.17475728155339806,
196
+ "grad_norm": 0.517716646194458,
197
+ "learning_rate": 8.535117056856188e-05,
198
+ "loss": 0.7694,
199
+ "step": 270
200
+ },
201
+ {
202
+ "epoch": 0.18122977346278318,
203
+ "grad_norm": 0.5802426934242249,
204
+ "learning_rate": 8.468227424749165e-05,
205
+ "loss": 0.7505,
206
+ "step": 280
207
+ },
208
+ {
209
+ "epoch": 0.18770226537216828,
210
+ "grad_norm": 0.5289272665977478,
211
+ "learning_rate": 8.401337792642141e-05,
212
+ "loss": 0.7678,
213
+ "step": 290
214
+ },
215
+ {
216
+ "epoch": 0.1941747572815534,
217
+ "grad_norm": 0.5293733477592468,
218
+ "learning_rate": 8.334448160535117e-05,
219
+ "loss": 0.7242,
220
+ "step": 300
221
+ },
222
+ {
223
+ "epoch": 0.20064724919093851,
224
+ "grad_norm": 0.566146969795227,
225
+ "learning_rate": 8.267558528428094e-05,
226
+ "loss": 0.7248,
227
+ "step": 310
228
+ },
229
+ {
230
+ "epoch": 0.20711974110032363,
231
+ "grad_norm": 0.5496853590011597,
232
+ "learning_rate": 8.200668896321071e-05,
233
+ "loss": 0.7588,
234
+ "step": 320
235
+ },
236
+ {
237
+ "epoch": 0.21359223300970873,
238
+ "grad_norm": 0.5420507192611694,
239
+ "learning_rate": 8.133779264214048e-05,
240
+ "loss": 0.7673,
241
+ "step": 330
242
+ },
243
+ {
244
+ "epoch": 0.22006472491909385,
245
+ "grad_norm": 0.582582950592041,
246
+ "learning_rate": 8.066889632107024e-05,
247
+ "loss": 0.7341,
248
+ "step": 340
249
+ },
250
+ {
251
+ "epoch": 0.22653721682847897,
252
+ "grad_norm": 0.5332103371620178,
253
+ "learning_rate": 8e-05,
254
+ "loss": 0.7518,
255
+ "step": 350
256
+ },
257
+ {
258
+ "epoch": 0.23300970873786409,
259
+ "grad_norm": 0.5132858157157898,
260
+ "learning_rate": 7.933110367892977e-05,
261
+ "loss": 0.752,
262
+ "step": 360
263
+ },
264
+ {
265
+ "epoch": 0.23948220064724918,
266
+ "grad_norm": 0.5356752276420593,
267
+ "learning_rate": 7.866220735785953e-05,
268
+ "loss": 0.7432,
269
+ "step": 370
270
+ },
271
+ {
272
+ "epoch": 0.2459546925566343,
273
+ "grad_norm": 0.5592623353004456,
274
+ "learning_rate": 7.79933110367893e-05,
275
+ "loss": 0.7396,
276
+ "step": 380
277
+ },
278
+ {
279
+ "epoch": 0.2524271844660194,
280
+ "grad_norm": 0.5103404521942139,
281
+ "learning_rate": 7.732441471571906e-05,
282
+ "loss": 0.7253,
283
+ "step": 390
284
+ },
285
+ {
286
+ "epoch": 0.2588996763754045,
287
+ "grad_norm": 0.5380160212516785,
288
+ "learning_rate": 7.665551839464883e-05,
289
+ "loss": 0.7464,
290
+ "step": 400
291
+ },
292
+ {
293
+ "epoch": 0.26537216828478966,
294
+ "grad_norm": 0.5373779535293579,
295
+ "learning_rate": 7.598662207357859e-05,
296
+ "loss": 0.7377,
297
+ "step": 410
298
+ },
299
+ {
300
+ "epoch": 0.27184466019417475,
301
+ "grad_norm": 0.523170530796051,
302
+ "learning_rate": 7.531772575250837e-05,
303
+ "loss": 0.7413,
304
+ "step": 420
305
+ },
306
+ {
307
+ "epoch": 0.2783171521035599,
308
+ "grad_norm": 0.5368985533714294,
309
+ "learning_rate": 7.464882943143813e-05,
310
+ "loss": 0.7227,
311
+ "step": 430
312
+ },
313
+ {
314
+ "epoch": 0.284789644012945,
315
+ "grad_norm": 0.587778627872467,
316
+ "learning_rate": 7.39799331103679e-05,
317
+ "loss": 0.7289,
318
+ "step": 440
319
+ },
320
+ {
321
+ "epoch": 0.2912621359223301,
322
+ "grad_norm": 0.5471896529197693,
323
+ "learning_rate": 7.331103678929766e-05,
324
+ "loss": 0.7301,
325
+ "step": 450
326
+ },
327
+ {
328
+ "epoch": 0.2977346278317152,
329
+ "grad_norm": 0.5446394085884094,
330
+ "learning_rate": 7.264214046822742e-05,
331
+ "loss": 0.7325,
332
+ "step": 460
333
+ },
334
+ {
335
+ "epoch": 0.3042071197411003,
336
+ "grad_norm": 0.5761464238166809,
337
+ "learning_rate": 7.197324414715719e-05,
338
+ "loss": 0.7583,
339
+ "step": 470
340
+ },
341
+ {
342
+ "epoch": 0.3106796116504854,
343
+ "grad_norm": 0.5082632899284363,
344
+ "learning_rate": 7.130434782608696e-05,
345
+ "loss": 0.7382,
346
+ "step": 480
347
+ },
348
+ {
349
+ "epoch": 0.31715210355987056,
350
+ "grad_norm": 0.5441333651542664,
351
+ "learning_rate": 7.063545150501673e-05,
352
+ "loss": 0.7411,
353
+ "step": 490
354
+ },
355
+ {
356
+ "epoch": 0.32362459546925565,
357
+ "grad_norm": 0.5039440393447876,
358
+ "learning_rate": 6.996655518394649e-05,
359
+ "loss": 0.7332,
360
+ "step": 500
361
+ }
362
+ ],
363
+ "logging_steps": 10,
364
+ "max_steps": 1545,
365
+ "num_input_tokens_seen": 0,
366
+ "num_train_epochs": 1,
367
+ "save_steps": 500,
368
+ "stateful_callbacks": {
369
+ "TrainerControl": {
370
+ "args": {
371
+ "should_epoch_stop": false,
372
+ "should_evaluate": false,
373
+ "should_log": false,
374
+ "should_save": true,
375
+ "should_training_stop": false
376
+ },
377
+ "attributes": {}
378
+ }
379
+ },
380
+ "total_flos": 1.803063236717445e+17,
381
+ "train_batch_size": 16,
382
+ "trial_name": null,
383
+ "trial_params": null
384
+ }