asun17904 commited on
Commit
7e88ada
·
verified ·
1 Parent(s): 12d1976

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/gemma-2-2b-it
3
+ library_name: peft
4
+ pipeline_tag: text-generation
5
+ tags:
6
+ - base_model:adapter:google/gemma-2-2b-it
7
+ - lora
8
+ - transformers
9
+ ---
10
+
11
+ # Model Card for Model ID
12
+
13
+ <!-- Provide a quick summary of what the model is/does. -->
14
+
15
+
16
+
17
+ ## Model Details
18
+
19
+ ### Model Description
20
+
21
+ <!-- Provide a longer summary of what this model is. -->
22
+
23
+
24
+
25
+ - **Developed by:** [More Information Needed]
26
+ - **Funded by [optional]:** [More Information Needed]
27
+ - **Shared by [optional]:** [More Information Needed]
28
+ - **Model type:** [More Information Needed]
29
+ - **Language(s) (NLP):** [More Information Needed]
30
+ - **License:** [More Information Needed]
31
+ - **Finetuned from model [optional]:** [More Information Needed]
32
+
33
+ ### Model Sources [optional]
34
+
35
+ <!-- Provide the basic links for the model. -->
36
+
37
+ - **Repository:** [More Information Needed]
38
+ - **Paper [optional]:** [More Information Needed]
39
+ - **Demo [optional]:** [More Information Needed]
40
+
41
+ ## Uses
42
+
43
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
44
+
45
+ ### Direct Use
46
+
47
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
48
+
49
+ [More Information Needed]
50
+
51
+ ### Downstream Use [optional]
52
+
53
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
54
+
55
+ [More Information Needed]
56
+
57
+ ### Out-of-Scope Use
58
+
59
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
60
+
61
+ [More Information Needed]
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ [More Information Needed]
68
+
69
+ ### Recommendations
70
+
71
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
72
+
73
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
74
+
75
+ ## How to Get Started with the Model
76
+
77
+ Use the code below to get started with the model.
78
+
79
+ [More Information Needed]
80
+
81
+ ## Training Details
82
+
83
+ ### Training Data
84
+
85
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
86
+
87
+ [More Information Needed]
88
+
89
+ ### Training Procedure
90
+
91
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
92
+
93
+ #### Preprocessing [optional]
94
+
95
+ [More Information Needed]
96
+
97
+
98
+ #### Training Hyperparameters
99
+
100
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
101
+
102
+ #### Speeds, Sizes, Times [optional]
103
+
104
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
105
+
106
+ [More Information Needed]
107
+
108
+ ## Evaluation
109
+
110
+ <!-- This section describes the evaluation protocols and provides the results. -->
111
+
112
+ ### Testing Data, Factors & Metrics
113
+
114
+ #### Testing Data
115
+
116
+ <!-- This should link to a Dataset Card if possible. -->
117
+
118
+ [More Information Needed]
119
+
120
+ #### Factors
121
+
122
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
123
+
124
+ [More Information Needed]
125
+
126
+ #### Metrics
127
+
128
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
129
+
130
+ [More Information Needed]
131
+
132
+ ### Results
133
+
134
+ [More Information Needed]
135
+
136
+ #### Summary
137
+
138
+
139
+
140
+ ## Model Examination [optional]
141
+
142
+ <!-- Relevant interpretability work for the model goes here -->
143
+
144
+ [More Information Needed]
145
+
146
+ ## Environmental Impact
147
+
148
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
149
+
150
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
151
+
152
+ - **Hardware Type:** [More Information Needed]
153
+ - **Hours used:** [More Information Needed]
154
+ - **Cloud Provider:** [More Information Needed]
155
+ - **Compute Region:** [More Information Needed]
156
+ - **Carbon Emitted:** [More Information Needed]
157
+
158
+ ## Technical Specifications [optional]
159
+
160
+ ### Model Architecture and Objective
161
+
162
+ [More Information Needed]
163
+
164
+ ### Compute Infrastructure
165
+
166
+ [More Information Needed]
167
+
168
+ #### Hardware
169
+
170
+ [More Information Needed]
171
+
172
+ #### Software
173
+
174
+ [More Information Needed]
175
+
176
+ ## Citation [optional]
177
+
178
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
179
+
180
+ **BibTeX:**
181
+
182
+ [More Information Needed]
183
+
184
+ **APA:**
185
+
186
+ [More Information Needed]
187
+
188
+ ## Glossary [optional]
189
+
190
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
191
+
192
+ [More Information Needed]
193
+
194
+ ## More Information [optional]
195
+
196
+ [More Information Needed]
197
+
198
+ ## Model Card Authors [optional]
199
+
200
+ [More Information Needed]
201
+
202
+ ## Model Card Contact
203
+
204
+ [More Information Needed]
205
+ ### Framework versions
206
+
207
+ - PEFT 0.18.0
adapter_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alora_invocation_tokens": null,
3
+ "alpha_pattern": {},
4
+ "arrow_config": null,
5
+ "auto_mapping": null,
6
+ "base_model_name_or_path": "google/gemma-2-2b-it",
7
+ "bias": "none",
8
+ "corda_config": null,
9
+ "ensure_weight_tying": false,
10
+ "eva_config": null,
11
+ "exclude_modules": null,
12
+ "fan_in_fan_out": false,
13
+ "inference_mode": true,
14
+ "init_lora_weights": true,
15
+ "layer_replication": null,
16
+ "layers_pattern": null,
17
+ "layers_to_transform": null,
18
+ "loftq_config": {},
19
+ "lora_alpha": 16,
20
+ "lora_bias": false,
21
+ "lora_dropout": 0.05,
22
+ "megatron_config": null,
23
+ "megatron_core": "megatron.core",
24
+ "modules_to_save": null,
25
+ "peft_type": "LORA",
26
+ "peft_version": "0.18.0",
27
+ "qalora_group_size": 16,
28
+ "r": 8,
29
+ "rank_pattern": {},
30
+ "revision": null,
31
+ "target_modules": [
32
+ "gate_proj",
33
+ "k_proj",
34
+ "o_proj",
35
+ "up_proj",
36
+ "down_proj",
37
+ "v_proj",
38
+ "q_proj"
39
+ ],
40
+ "target_parameters": null,
41
+ "task_type": "CAUSAL_LM",
42
+ "trainable_token_indices": null,
43
+ "use_dora": false,
44
+ "use_qalora": false,
45
+ "use_rslora": false
46
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2c55ce2e0bc5d9cbf6f59cd73f14b4b3afa1f733444c69ef299cee3b560933f
3
+ size 4760230520
optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e2bc7bdc8f99bfb534fdb34907c405450f788a98bc0cfc3f3bce77d5faed9a8
3
+ size 83435239
pytorch_model_fsdp.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10084d923a553ce6bf51ee77d5dafd3361ad8f63bbdfeacbb21dbd51fec6b2df
3
+ size 41669389
rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:038b7a53449691b1a82168faf5c866687dee9dcdf3251f8a7b2c72be550e1ed8
3
+ size 14853
rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f948428f5e3951179e5c18b82d3db76bcebe7dfa8329e0407d4449a9974a582b
3
+ size 14853
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:294ef3359a016c0d29e847800f90df41811358f65e37e7a827e762cbcbde13cd
3
+ size 1465
trainer_state.json ADDED
@@ -0,0 +1,636 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 50,
7
+ "global_step": 708,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.0423728813559322,
14
+ "grad_norm": 1.7643470764160156,
15
+ "learning_rate": 4.936440677966102e-05,
16
+ "loss": 1.7478,
17
+ "step": 10
18
+ },
19
+ {
20
+ "epoch": 0.0847457627118644,
21
+ "grad_norm": 1.352023959159851,
22
+ "learning_rate": 4.8658192090395484e-05,
23
+ "loss": 1.2556,
24
+ "step": 20
25
+ },
26
+ {
27
+ "epoch": 0.1271186440677966,
28
+ "grad_norm": 1.205060601234436,
29
+ "learning_rate": 4.795197740112994e-05,
30
+ "loss": 1.1919,
31
+ "step": 30
32
+ },
33
+ {
34
+ "epoch": 0.1694915254237288,
35
+ "grad_norm": 1.019239902496338,
36
+ "learning_rate": 4.724576271186441e-05,
37
+ "loss": 1.061,
38
+ "step": 40
39
+ },
40
+ {
41
+ "epoch": 0.211864406779661,
42
+ "grad_norm": 1.0925650596618652,
43
+ "learning_rate": 4.6539548022598875e-05,
44
+ "loss": 1.1092,
45
+ "step": 50
46
+ },
47
+ {
48
+ "epoch": 0.211864406779661,
49
+ "eval_loss": 1.0438729524612427,
50
+ "eval_runtime": 22.5506,
51
+ "eval_samples_per_second": 95.385,
52
+ "eval_steps_per_second": 1.508,
53
+ "step": 50
54
+ },
55
+ {
56
+ "epoch": 0.2542372881355932,
57
+ "grad_norm": 0.9862654209136963,
58
+ "learning_rate": 4.5833333333333334e-05,
59
+ "loss": 1.0876,
60
+ "step": 60
61
+ },
62
+ {
63
+ "epoch": 0.2966101694915254,
64
+ "grad_norm": 0.9759798645973206,
65
+ "learning_rate": 4.51271186440678e-05,
66
+ "loss": 1.0237,
67
+ "step": 70
68
+ },
69
+ {
70
+ "epoch": 0.3389830508474576,
71
+ "grad_norm": 1.0690054893493652,
72
+ "learning_rate": 4.442090395480226e-05,
73
+ "loss": 1.0092,
74
+ "step": 80
75
+ },
76
+ {
77
+ "epoch": 0.3813559322033898,
78
+ "grad_norm": 1.1084680557250977,
79
+ "learning_rate": 4.3714689265536725e-05,
80
+ "loss": 1.0229,
81
+ "step": 90
82
+ },
83
+ {
84
+ "epoch": 0.423728813559322,
85
+ "grad_norm": 1.081882357597351,
86
+ "learning_rate": 4.300847457627119e-05,
87
+ "loss": 1.0233,
88
+ "step": 100
89
+ },
90
+ {
91
+ "epoch": 0.423728813559322,
92
+ "eval_loss": 1.003989815711975,
93
+ "eval_runtime": 22.3839,
94
+ "eval_samples_per_second": 96.096,
95
+ "eval_steps_per_second": 1.519,
96
+ "step": 100
97
+ },
98
+ {
99
+ "epoch": 0.4661016949152542,
100
+ "grad_norm": 1.1931794881820679,
101
+ "learning_rate": 4.230225988700565e-05,
102
+ "loss": 0.9807,
103
+ "step": 110
104
+ },
105
+ {
106
+ "epoch": 0.5084745762711864,
107
+ "grad_norm": 1.1798979043960571,
108
+ "learning_rate": 4.1596045197740115e-05,
109
+ "loss": 1.037,
110
+ "step": 120
111
+ },
112
+ {
113
+ "epoch": 0.5508474576271186,
114
+ "grad_norm": 1.236763596534729,
115
+ "learning_rate": 4.088983050847458e-05,
116
+ "loss": 1.0088,
117
+ "step": 130
118
+ },
119
+ {
120
+ "epoch": 0.5932203389830508,
121
+ "grad_norm": 1.031708836555481,
122
+ "learning_rate": 4.018361581920904e-05,
123
+ "loss": 0.9605,
124
+ "step": 140
125
+ },
126
+ {
127
+ "epoch": 0.635593220338983,
128
+ "grad_norm": 1.0881413221359253,
129
+ "learning_rate": 3.9477401129943506e-05,
130
+ "loss": 0.9669,
131
+ "step": 150
132
+ },
133
+ {
134
+ "epoch": 0.635593220338983,
135
+ "eval_loss": 0.9852223992347717,
136
+ "eval_runtime": 22.645,
137
+ "eval_samples_per_second": 94.988,
138
+ "eval_steps_per_second": 1.501,
139
+ "step": 150
140
+ },
141
+ {
142
+ "epoch": 0.6779661016949152,
143
+ "grad_norm": 1.0248509645462036,
144
+ "learning_rate": 3.877118644067797e-05,
145
+ "loss": 0.9584,
146
+ "step": 160
147
+ },
148
+ {
149
+ "epoch": 0.7203389830508474,
150
+ "grad_norm": 1.1003658771514893,
151
+ "learning_rate": 3.806497175141243e-05,
152
+ "loss": 1.0191,
153
+ "step": 170
154
+ },
155
+ {
156
+ "epoch": 0.7627118644067796,
157
+ "grad_norm": 0.9632079005241394,
158
+ "learning_rate": 3.735875706214689e-05,
159
+ "loss": 1.0048,
160
+ "step": 180
161
+ },
162
+ {
163
+ "epoch": 0.8050847457627118,
164
+ "grad_norm": 1.2166016101837158,
165
+ "learning_rate": 3.665254237288136e-05,
166
+ "loss": 1.0035,
167
+ "step": 190
168
+ },
169
+ {
170
+ "epoch": 0.847457627118644,
171
+ "grad_norm": 1.132070779800415,
172
+ "learning_rate": 3.594632768361582e-05,
173
+ "loss": 0.9939,
174
+ "step": 200
175
+ },
176
+ {
177
+ "epoch": 0.847457627118644,
178
+ "eval_loss": 0.9768873453140259,
179
+ "eval_runtime": 23.2377,
180
+ "eval_samples_per_second": 92.565,
181
+ "eval_steps_per_second": 1.463,
182
+ "step": 200
183
+ },
184
+ {
185
+ "epoch": 0.8898305084745762,
186
+ "grad_norm": 1.0186951160430908,
187
+ "learning_rate": 3.524011299435028e-05,
188
+ "loss": 0.9749,
189
+ "step": 210
190
+ },
191
+ {
192
+ "epoch": 0.9322033898305084,
193
+ "grad_norm": 1.2069514989852905,
194
+ "learning_rate": 3.4533898305084746e-05,
195
+ "loss": 0.9756,
196
+ "step": 220
197
+ },
198
+ {
199
+ "epoch": 0.9745762711864406,
200
+ "grad_norm": 1.1548621654510498,
201
+ "learning_rate": 3.382768361581921e-05,
202
+ "loss": 0.9723,
203
+ "step": 230
204
+ },
205
+ {
206
+ "epoch": 1.0169491525423728,
207
+ "grad_norm": 1.0338218212127686,
208
+ "learning_rate": 3.312146892655367e-05,
209
+ "loss": 0.9304,
210
+ "step": 240
211
+ },
212
+ {
213
+ "epoch": 1.0593220338983051,
214
+ "grad_norm": 1.1850943565368652,
215
+ "learning_rate": 3.241525423728814e-05,
216
+ "loss": 0.9512,
217
+ "step": 250
218
+ },
219
+ {
220
+ "epoch": 1.0593220338983051,
221
+ "eval_loss": 0.9682268500328064,
222
+ "eval_runtime": 22.9025,
223
+ "eval_samples_per_second": 93.92,
224
+ "eval_steps_per_second": 1.485,
225
+ "step": 250
226
+ },
227
+ {
228
+ "epoch": 1.1016949152542372,
229
+ "grad_norm": 1.0937676429748535,
230
+ "learning_rate": 3.17090395480226e-05,
231
+ "loss": 0.9241,
232
+ "step": 260
233
+ },
234
+ {
235
+ "epoch": 1.1440677966101696,
236
+ "grad_norm": 1.296743631362915,
237
+ "learning_rate": 3.100282485875706e-05,
238
+ "loss": 0.9508,
239
+ "step": 270
240
+ },
241
+ {
242
+ "epoch": 1.1864406779661016,
243
+ "grad_norm": 1.145988941192627,
244
+ "learning_rate": 3.0296610169491528e-05,
245
+ "loss": 0.9629,
246
+ "step": 280
247
+ },
248
+ {
249
+ "epoch": 1.228813559322034,
250
+ "grad_norm": 1.2783371210098267,
251
+ "learning_rate": 2.959039548022599e-05,
252
+ "loss": 0.9621,
253
+ "step": 290
254
+ },
255
+ {
256
+ "epoch": 1.271186440677966,
257
+ "grad_norm": 1.3129595518112183,
258
+ "learning_rate": 2.8884180790960453e-05,
259
+ "loss": 0.9667,
260
+ "step": 300
261
+ },
262
+ {
263
+ "epoch": 1.271186440677966,
264
+ "eval_loss": 0.9666466116905212,
265
+ "eval_runtime": 23.5102,
266
+ "eval_samples_per_second": 91.492,
267
+ "eval_steps_per_second": 1.446,
268
+ "step": 300
269
+ },
270
+ {
271
+ "epoch": 1.3135593220338984,
272
+ "grad_norm": 1.2464239597320557,
273
+ "learning_rate": 2.817796610169492e-05,
274
+ "loss": 0.9359,
275
+ "step": 310
276
+ },
277
+ {
278
+ "epoch": 1.3559322033898304,
279
+ "grad_norm": 1.3320660591125488,
280
+ "learning_rate": 2.747175141242938e-05,
281
+ "loss": 0.9232,
282
+ "step": 320
283
+ },
284
+ {
285
+ "epoch": 1.3983050847457628,
286
+ "grad_norm": 1.3213363885879517,
287
+ "learning_rate": 2.6765536723163843e-05,
288
+ "loss": 0.9064,
289
+ "step": 330
290
+ },
291
+ {
292
+ "epoch": 1.4406779661016949,
293
+ "grad_norm": 1.4874022006988525,
294
+ "learning_rate": 2.605932203389831e-05,
295
+ "loss": 0.9367,
296
+ "step": 340
297
+ },
298
+ {
299
+ "epoch": 1.4830508474576272,
300
+ "grad_norm": 1.308205246925354,
301
+ "learning_rate": 2.535310734463277e-05,
302
+ "loss": 0.9058,
303
+ "step": 350
304
+ },
305
+ {
306
+ "epoch": 1.4830508474576272,
307
+ "eval_loss": 0.964046061038971,
308
+ "eval_runtime": 22.9381,
309
+ "eval_samples_per_second": 93.774,
310
+ "eval_steps_per_second": 1.482,
311
+ "step": 350
312
+ },
313
+ {
314
+ "epoch": 1.5254237288135593,
315
+ "grad_norm": 1.3957818746566772,
316
+ "learning_rate": 2.464689265536723e-05,
317
+ "loss": 0.9232,
318
+ "step": 360
319
+ },
320
+ {
321
+ "epoch": 1.5677966101694916,
322
+ "grad_norm": 1.3040454387664795,
323
+ "learning_rate": 2.3940677966101697e-05,
324
+ "loss": 0.9376,
325
+ "step": 370
326
+ },
327
+ {
328
+ "epoch": 1.6101694915254239,
329
+ "grad_norm": 1.4122378826141357,
330
+ "learning_rate": 2.323446327683616e-05,
331
+ "loss": 0.9115,
332
+ "step": 380
333
+ },
334
+ {
335
+ "epoch": 1.652542372881356,
336
+ "grad_norm": 1.3442506790161133,
337
+ "learning_rate": 2.252824858757062e-05,
338
+ "loss": 0.9599,
339
+ "step": 390
340
+ },
341
+ {
342
+ "epoch": 1.694915254237288,
343
+ "grad_norm": 1.5409705638885498,
344
+ "learning_rate": 2.1822033898305087e-05,
345
+ "loss": 0.9258,
346
+ "step": 400
347
+ },
348
+ {
349
+ "epoch": 1.694915254237288,
350
+ "eval_loss": 0.9589577913284302,
351
+ "eval_runtime": 22.6608,
352
+ "eval_samples_per_second": 94.922,
353
+ "eval_steps_per_second": 1.5,
354
+ "step": 400
355
+ },
356
+ {
357
+ "epoch": 1.7372881355932204,
358
+ "grad_norm": 1.315023422241211,
359
+ "learning_rate": 2.111581920903955e-05,
360
+ "loss": 0.9088,
361
+ "step": 410
362
+ },
363
+ {
364
+ "epoch": 1.7796610169491527,
365
+ "grad_norm": 1.4532520771026611,
366
+ "learning_rate": 2.0409604519774012e-05,
367
+ "loss": 0.9351,
368
+ "step": 420
369
+ },
370
+ {
371
+ "epoch": 1.8220338983050848,
372
+ "grad_norm": 1.3708842992782593,
373
+ "learning_rate": 1.9703389830508475e-05,
374
+ "loss": 0.95,
375
+ "step": 430
376
+ },
377
+ {
378
+ "epoch": 1.8644067796610169,
379
+ "grad_norm": 1.3897417783737183,
380
+ "learning_rate": 1.899717514124294e-05,
381
+ "loss": 0.8943,
382
+ "step": 440
383
+ },
384
+ {
385
+ "epoch": 1.9067796610169492,
386
+ "grad_norm": 1.2795928716659546,
387
+ "learning_rate": 1.8290960451977403e-05,
388
+ "loss": 0.8856,
389
+ "step": 450
390
+ },
391
+ {
392
+ "epoch": 1.9067796610169492,
393
+ "eval_loss": 0.9577222466468811,
394
+ "eval_runtime": 23.1528,
395
+ "eval_samples_per_second": 92.905,
396
+ "eval_steps_per_second": 1.469,
397
+ "step": 450
398
+ },
399
+ {
400
+ "epoch": 1.9491525423728815,
401
+ "grad_norm": 1.2920231819152832,
402
+ "learning_rate": 1.7584745762711865e-05,
403
+ "loss": 0.9415,
404
+ "step": 460
405
+ },
406
+ {
407
+ "epoch": 1.9915254237288136,
408
+ "grad_norm": 1.3519489765167236,
409
+ "learning_rate": 1.687853107344633e-05,
410
+ "loss": 0.9708,
411
+ "step": 470
412
+ },
413
+ {
414
+ "epoch": 2.0338983050847457,
415
+ "grad_norm": 1.3836076259613037,
416
+ "learning_rate": 1.617231638418079e-05,
417
+ "loss": 0.8801,
418
+ "step": 480
419
+ },
420
+ {
421
+ "epoch": 2.0762711864406778,
422
+ "grad_norm": 1.478256344795227,
423
+ "learning_rate": 1.5466101694915256e-05,
424
+ "loss": 0.8908,
425
+ "step": 490
426
+ },
427
+ {
428
+ "epoch": 2.1186440677966103,
429
+ "grad_norm": 1.571032166481018,
430
+ "learning_rate": 1.475988700564972e-05,
431
+ "loss": 0.9013,
432
+ "step": 500
433
+ },
434
+ {
435
+ "epoch": 2.1186440677966103,
436
+ "eval_loss": 0.9608597755432129,
437
+ "eval_runtime": 22.5872,
438
+ "eval_samples_per_second": 95.231,
439
+ "eval_steps_per_second": 1.505,
440
+ "step": 500
441
+ },
442
+ {
443
+ "epoch": 2.1610169491525424,
444
+ "grad_norm": 1.466437578201294,
445
+ "learning_rate": 1.4053672316384181e-05,
446
+ "loss": 0.8926,
447
+ "step": 510
448
+ },
449
+ {
450
+ "epoch": 2.2033898305084745,
451
+ "grad_norm": 1.6073771715164185,
452
+ "learning_rate": 1.3347457627118645e-05,
453
+ "loss": 0.8717,
454
+ "step": 520
455
+ },
456
+ {
457
+ "epoch": 2.2457627118644066,
458
+ "grad_norm": 1.5890705585479736,
459
+ "learning_rate": 1.2641242937853107e-05,
460
+ "loss": 0.8633,
461
+ "step": 530
462
+ },
463
+ {
464
+ "epoch": 2.288135593220339,
465
+ "grad_norm": 1.6749075651168823,
466
+ "learning_rate": 1.1935028248587572e-05,
467
+ "loss": 0.8941,
468
+ "step": 540
469
+ },
470
+ {
471
+ "epoch": 2.330508474576271,
472
+ "grad_norm": 1.6727330684661865,
473
+ "learning_rate": 1.1228813559322036e-05,
474
+ "loss": 0.8929,
475
+ "step": 550
476
+ },
477
+ {
478
+ "epoch": 2.330508474576271,
479
+ "eval_loss": 0.9595548510551453,
480
+ "eval_runtime": 22.9206,
481
+ "eval_samples_per_second": 93.846,
482
+ "eval_steps_per_second": 1.483,
483
+ "step": 550
484
+ },
485
+ {
486
+ "epoch": 2.3728813559322033,
487
+ "grad_norm": 1.5609760284423828,
488
+ "learning_rate": 1.0522598870056498e-05,
489
+ "loss": 0.8298,
490
+ "step": 560
491
+ },
492
+ {
493
+ "epoch": 2.415254237288136,
494
+ "grad_norm": 1.7489114999771118,
495
+ "learning_rate": 9.81638418079096e-06,
496
+ "loss": 0.8618,
497
+ "step": 570
498
+ },
499
+ {
500
+ "epoch": 2.457627118644068,
501
+ "grad_norm": 1.7662324905395508,
502
+ "learning_rate": 9.110169491525423e-06,
503
+ "loss": 0.8932,
504
+ "step": 580
505
+ },
506
+ {
507
+ "epoch": 2.5,
508
+ "grad_norm": 1.9018930196762085,
509
+ "learning_rate": 8.403954802259887e-06,
510
+ "loss": 0.8743,
511
+ "step": 590
512
+ },
513
+ {
514
+ "epoch": 2.542372881355932,
515
+ "grad_norm": 1.6308236122131348,
516
+ "learning_rate": 7.697740112994351e-06,
517
+ "loss": 0.8752,
518
+ "step": 600
519
+ },
520
+ {
521
+ "epoch": 2.542372881355932,
522
+ "eval_loss": 0.9617831707000732,
523
+ "eval_runtime": 23.2374,
524
+ "eval_samples_per_second": 92.566,
525
+ "eval_steps_per_second": 1.463,
526
+ "step": 600
527
+ },
528
+ {
529
+ "epoch": 2.584745762711864,
530
+ "grad_norm": 1.5900993347167969,
531
+ "learning_rate": 6.991525423728814e-06,
532
+ "loss": 0.8653,
533
+ "step": 610
534
+ },
535
+ {
536
+ "epoch": 2.6271186440677967,
537
+ "grad_norm": 1.8414711952209473,
538
+ "learning_rate": 6.285310734463278e-06,
539
+ "loss": 0.8877,
540
+ "step": 620
541
+ },
542
+ {
543
+ "epoch": 2.669491525423729,
544
+ "grad_norm": 1.7117115259170532,
545
+ "learning_rate": 5.57909604519774e-06,
546
+ "loss": 0.8691,
547
+ "step": 630
548
+ },
549
+ {
550
+ "epoch": 2.711864406779661,
551
+ "grad_norm": 1.690845012664795,
552
+ "learning_rate": 4.872881355932204e-06,
553
+ "loss": 0.8728,
554
+ "step": 640
555
+ },
556
+ {
557
+ "epoch": 2.7542372881355934,
558
+ "grad_norm": 1.727298378944397,
559
+ "learning_rate": 4.166666666666667e-06,
560
+ "loss": 0.9005,
561
+ "step": 650
562
+ },
563
+ {
564
+ "epoch": 2.7542372881355934,
565
+ "eval_loss": 0.959761381149292,
566
+ "eval_runtime": 22.8543,
567
+ "eval_samples_per_second": 94.118,
568
+ "eval_steps_per_second": 1.488,
569
+ "step": 650
570
+ },
571
+ {
572
+ "epoch": 2.7966101694915255,
573
+ "grad_norm": 1.6035046577453613,
574
+ "learning_rate": 3.46045197740113e-06,
575
+ "loss": 0.9181,
576
+ "step": 660
577
+ },
578
+ {
579
+ "epoch": 2.8389830508474576,
580
+ "grad_norm": 1.7235994338989258,
581
+ "learning_rate": 2.7542372881355934e-06,
582
+ "loss": 0.9003,
583
+ "step": 670
584
+ },
585
+ {
586
+ "epoch": 2.8813559322033897,
587
+ "grad_norm": 1.7914451360702515,
588
+ "learning_rate": 2.0480225988700563e-06,
589
+ "loss": 0.951,
590
+ "step": 680
591
+ },
592
+ {
593
+ "epoch": 2.923728813559322,
594
+ "grad_norm": 1.8638156652450562,
595
+ "learning_rate": 1.3418079096045198e-06,
596
+ "loss": 0.8739,
597
+ "step": 690
598
+ },
599
+ {
600
+ "epoch": 2.9661016949152543,
601
+ "grad_norm": 1.6826245784759521,
602
+ "learning_rate": 6.355932203389831e-07,
603
+ "loss": 0.8375,
604
+ "step": 700
605
+ },
606
+ {
607
+ "epoch": 2.9661016949152543,
608
+ "eval_loss": 0.9599277973175049,
609
+ "eval_runtime": 22.8971,
610
+ "eval_samples_per_second": 93.942,
611
+ "eval_steps_per_second": 1.485,
612
+ "step": 700
613
+ }
614
+ ],
615
+ "logging_steps": 10,
616
+ "max_steps": 708,
617
+ "num_input_tokens_seen": 0,
618
+ "num_train_epochs": 3,
619
+ "save_steps": 50,
620
+ "stateful_callbacks": {
621
+ "TrainerControl": {
622
+ "args": {
623
+ "should_epoch_stop": false,
624
+ "should_evaluate": false,
625
+ "should_log": false,
626
+ "should_save": true,
627
+ "should_training_stop": true
628
+ },
629
+ "attributes": {}
630
+ }
631
+ },
632
+ "total_flos": 1.8604011452104704e+16,
633
+ "train_batch_size": 32,
634
+ "trial_name": null,
635
+ "trial_params": null
636
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45230dc52c8a9b80d460a2f194beb3f6064dc863151313ebd149ad62df34e59e
3
+ size 6097