irishprancer commited on
Commit
c9ec126
·
verified ·
1 Parent(s): 29e5752

Training in progress, step 150, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -22,8 +22,8 @@
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
- "q_proj",
26
- "k_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
+ "k_proj",
26
+ "q_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c571c3d9bc97035512218123eb16c4a5ac39adeb400b8f87dca66f9daff0f4c
3
  size 527048968
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abee9defa2f4343dad896a46c39f88dc5e2d2b650491c0c4e7d2178312355299
3
  size 527048968
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4bb2b5192e62889a659f5c32729030ddd962b6ab8d9ab4f6e8f06e1b2f7f86d0
3
  size 1054135994
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e158809030e00ab57a06e7f58d9209551ec0bf02f8cf1e23626ea2cae56c82d
3
  size 1054135994
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:037d13720220086c05f76f1146cd4356e8b9d075b5d306338df00d366045e1c1
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5feb56512e955691dc9bb9a1e37b9dd590e06a961d7d94560b679e2730b03194
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f4c00d522bdde510099aafe1617b13d114dce17a17b44e05876f016f4e4d7af
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cf3f988e8fed2daa2e801eb1f19b681872781cf57f0fb7b896e859a12cfe2bb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,302 +1,157 @@
1
  {
2
- "best_metric": 0.7310147881507874,
3
- "best_model_checkpoint": "./output/checkpoint-300",
4
- "epoch": 13.043478260869565,
5
  "eval_steps": 150,
6
- "global_step": 300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.43478260869565216,
13
- "grad_norm": 1.5021581649780273,
14
  "learning_rate": 3e-06,
15
- "loss": 0.9064,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.8695652173913043,
20
- "grad_norm": 1.6793084144592285,
21
  "learning_rate": 6e-06,
22
- "loss": 0.9027,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 1.3043478260869565,
27
- "grad_norm": 1.7296501398086548,
28
  "learning_rate": 9e-06,
29
- "loss": 0.9003,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.7391304347826086,
34
- "grad_norm": 1.4458428621292114,
35
  "learning_rate": 1.2e-05,
36
- "loss": 0.9092,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 2.1739130434782608,
41
- "grad_norm": 1.351685643196106,
42
  "learning_rate": 1.5e-05,
43
- "loss": 0.8363,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 2.608695652173913,
48
- "grad_norm": 2.0312559604644775,
49
  "learning_rate": 1.8e-05,
50
  "loss": 0.8894,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 3.0434782608695654,
55
- "grad_norm": 1.4843778610229492,
56
  "learning_rate": 2.1e-05,
57
- "loss": 0.8913,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 3.4782608695652173,
62
- "grad_norm": 1.729393720626831,
63
  "learning_rate": 2.4e-05,
64
- "loss": 0.8231,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 3.9130434782608696,
69
- "grad_norm": 1.4239307641983032,
70
  "learning_rate": 2.7000000000000002e-05,
71
- "loss": 0.8528,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 4.3478260869565215,
76
- "grad_norm": 1.3656258583068848,
77
  "learning_rate": 3e-05,
78
- "loss": 0.865,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 4.782608695652174,
83
- "grad_norm": 2.198690176010132,
84
  "learning_rate": 2.999999702723963e-05,
85
- "loss": 0.8228,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 5.217391304347826,
90
- "grad_norm": 1.0728951692581177,
91
  "learning_rate": 2.9999988108959687e-05,
92
- "loss": 0.7655,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 5.6521739130434785,
97
- "grad_norm": 1.5677502155303955,
98
  "learning_rate": 2.9999973245163716e-05,
99
- "loss": 0.7412,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 6.086956521739131,
104
- "grad_norm": 1.9072229862213135,
105
  "learning_rate": 2.99999524358576e-05,
106
- "loss": 0.7657,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 6.521739130434782,
111
- "grad_norm": 1.1221002340316772,
112
  "learning_rate": 2.9999925681049593e-05,
113
  "loss": 0.7857,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 6.521739130434782,
118
- "eval_loss": 0.7962104678153992,
119
- "eval_runtime": 0.4685,
120
- "eval_samples_per_second": 21.344,
121
- "eval_steps_per_second": 21.344,
122
  "step": 150
123
  },
124
  {
125
  "epoch": 6.521739130434782,
126
- "eval_loss": 0.8606523275375366,
127
- "eval_runtime": 0.3993,
128
- "eval_samples_per_second": 25.042,
129
- "eval_steps_per_second": 25.042,
130
  "step": 150
131
  },
132
  {
133
  "epoch": 6.521739130434782,
134
- "eval_loss": 0.7962104678153992,
135
- "eval_runtime": 0.3965,
136
- "eval_samples_per_second": 25.221,
137
- "eval_steps_per_second": 25.221,
138
  "step": 150
139
  },
140
  {
141
  "epoch": 6.521739130434782,
142
- "eval_loss": 0.8606523275375366,
143
- "eval_runtime": 0.4092,
144
- "eval_samples_per_second": 24.439,
145
- "eval_steps_per_second": 24.439,
146
  "step": 150
147
  },
148
  {
149
  "epoch": 6.521739130434782,
150
- "eval_loss": 0.8619200587272644,
151
- "eval_runtime": 0.3955,
152
- "eval_samples_per_second": 25.287,
153
- "eval_steps_per_second": 25.287,
154
  "step": 150
155
- },
156
- {
157
- "epoch": 6.956521739130435,
158
- "grad_norm": 1.540989637374878,
159
- "learning_rate": 2.9999892980750297e-05,
160
- "loss": 0.6586,
161
- "step": 160
162
- },
163
- {
164
- "epoch": 7.391304347826087,
165
- "grad_norm": 1.345400094985962,
166
- "learning_rate": 2.9999854334972675e-05,
167
- "loss": 0.739,
168
- "step": 170
169
- },
170
- {
171
- "epoch": 7.826086956521739,
172
- "grad_norm": 1.725967526435852,
173
- "learning_rate": 2.999980974373204e-05,
174
- "loss": 0.7291,
175
- "step": 180
176
- },
177
- {
178
- "epoch": 8.26086956521739,
179
- "grad_norm": 1.5401579141616821,
180
- "learning_rate": 2.9999759207046075e-05,
181
- "loss": 0.6245,
182
- "step": 190
183
- },
184
- {
185
- "epoch": 8.695652173913043,
186
- "grad_norm": 1.7425003051757812,
187
- "learning_rate": 2.9999702724934804e-05,
188
- "loss": 0.6765,
189
- "step": 200
190
- },
191
- {
192
- "epoch": 9.130434782608695,
193
- "grad_norm": 1.0419254302978516,
194
- "learning_rate": 2.999964029742062e-05,
195
- "loss": 0.6524,
196
- "step": 210
197
- },
198
- {
199
- "epoch": 9.565217391304348,
200
- "grad_norm": 1.2119251489639282,
201
- "learning_rate": 2.9999571924528263e-05,
202
- "loss": 0.5593,
203
- "step": 220
204
- },
205
- {
206
- "epoch": 10.0,
207
- "grad_norm": 1.5276358127593994,
208
- "learning_rate": 2.9999497606284837e-05,
209
- "loss": 0.756,
210
- "step": 230
211
- },
212
- {
213
- "epoch": 10.434782608695652,
214
- "grad_norm": 1.413993000984192,
215
- "learning_rate": 2.9999417342719796e-05,
216
- "loss": 0.7116,
217
- "step": 240
218
- },
219
- {
220
- "epoch": 10.869565217391305,
221
- "grad_norm": 0.9750322699546814,
222
- "learning_rate": 2.9999331133864956e-05,
223
- "loss": 0.5897,
224
- "step": 250
225
- },
226
- {
227
- "epoch": 11.304347826086957,
228
- "grad_norm": 1.1936322450637817,
229
- "learning_rate": 2.9999238979754485e-05,
230
- "loss": 0.6547,
231
- "step": 260
232
- },
233
- {
234
- "epoch": 11.73913043478261,
235
- "grad_norm": 1.0489903688430786,
236
- "learning_rate": 2.999914088042492e-05,
237
- "loss": 0.6475,
238
- "step": 270
239
- },
240
- {
241
- "epoch": 12.173913043478262,
242
- "grad_norm": 1.3138858079910278,
243
- "learning_rate": 2.9999036835915132e-05,
244
- "loss": 0.5937,
245
- "step": 280
246
- },
247
- {
248
- "epoch": 12.608695652173914,
249
- "grad_norm": 1.0842424631118774,
250
- "learning_rate": 2.9998926846266365e-05,
251
- "loss": 0.6326,
252
- "step": 290
253
- },
254
- {
255
- "epoch": 13.043478260869565,
256
- "grad_norm": 1.3883247375488281,
257
- "learning_rate": 2.9998810911522213e-05,
258
- "loss": 0.5806,
259
- "step": 300
260
- },
261
- {
262
- "epoch": 13.043478260869565,
263
- "eval_loss": 0.7310147881507874,
264
- "eval_runtime": 0.4443,
265
- "eval_samples_per_second": 22.51,
266
- "eval_steps_per_second": 22.51,
267
- "step": 300
268
- },
269
- {
270
- "epoch": 13.043478260869565,
271
- "eval_loss": 0.8606523275375366,
272
- "eval_runtime": 0.4404,
273
- "eval_samples_per_second": 22.708,
274
- "eval_steps_per_second": 22.708,
275
- "step": 300
276
- },
277
- {
278
- "epoch": 13.043478260869565,
279
- "eval_loss": 0.7310147881507874,
280
- "eval_runtime": 0.4391,
281
- "eval_samples_per_second": 22.773,
282
- "eval_steps_per_second": 22.773,
283
- "step": 300
284
- },
285
- {
286
- "epoch": 13.043478260869565,
287
- "eval_loss": 0.7758375406265259,
288
- "eval_runtime": 0.4396,
289
- "eval_samples_per_second": 22.75,
290
- "eval_steps_per_second": 22.75,
291
- "step": 300
292
- },
293
- {
294
- "epoch": 13.043478260869565,
295
- "eval_loss": 0.860957145690918,
296
- "eval_runtime": 0.4444,
297
- "eval_samples_per_second": 22.5,
298
- "eval_steps_per_second": 22.5,
299
- "step": 300
300
  }
301
  ],
302
  "logging_steps": 10,
@@ -316,7 +171,7 @@
316
  "attributes": {}
317
  }
318
  },
319
- "total_flos": 7724643094462464.0,
320
  "train_batch_size": 4,
321
  "trial_name": null,
322
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.7949807047843933,
3
+ "best_model_checkpoint": "./output/checkpoint-150",
4
+ "epoch": 6.521739130434782,
5
  "eval_steps": 150,
6
+ "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
  "epoch": 0.43478260869565216,
13
+ "grad_norm": 1.5021635293960571,
14
  "learning_rate": 3e-06,
15
+ "loss": 0.906,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 0.8695652173913043,
20
+ "grad_norm": 1.687072992324829,
21
  "learning_rate": 6e-06,
22
+ "loss": 0.9025,
23
  "step": 20
24
  },
25
  {
26
  "epoch": 1.3043478260869565,
27
+ "grad_norm": 1.7296812534332275,
28
  "learning_rate": 9e-06,
29
+ "loss": 0.9001,
30
  "step": 30
31
  },
32
  {
33
  "epoch": 1.7391304347826086,
34
+ "grad_norm": 1.4459028244018555,
35
  "learning_rate": 1.2e-05,
36
+ "loss": 0.909,
37
  "step": 40
38
  },
39
  {
40
  "epoch": 2.1739130434782608,
41
+ "grad_norm": 1.351724624633789,
42
  "learning_rate": 1.5e-05,
43
+ "loss": 0.8361,
44
  "step": 50
45
  },
46
  {
47
  "epoch": 2.608695652173913,
48
+ "grad_norm": 2.0466818809509277,
49
  "learning_rate": 1.8e-05,
50
  "loss": 0.8894,
51
  "step": 60
52
  },
53
  {
54
  "epoch": 3.0434782608695654,
55
+ "grad_norm": 1.4843230247497559,
56
  "learning_rate": 2.1e-05,
57
+ "loss": 0.8912,
58
  "step": 70
59
  },
60
  {
61
  "epoch": 3.4782608695652173,
62
+ "grad_norm": 1.7293753623962402,
63
  "learning_rate": 2.4e-05,
64
+ "loss": 0.8237,
65
  "step": 80
66
  },
67
  {
68
  "epoch": 3.9130434782608696,
69
+ "grad_norm": 1.424095869064331,
70
  "learning_rate": 2.7000000000000002e-05,
71
+ "loss": 0.8527,
72
  "step": 90
73
  },
74
  {
75
  "epoch": 4.3478260869565215,
76
+ "grad_norm": 1.3656634092330933,
77
  "learning_rate": 3e-05,
78
+ "loss": 0.8649,
79
  "step": 100
80
  },
81
  {
82
  "epoch": 4.782608695652174,
83
+ "grad_norm": 2.198690891265869,
84
  "learning_rate": 2.999999702723963e-05,
85
+ "loss": 0.8224,
86
  "step": 110
87
  },
88
  {
89
  "epoch": 5.217391304347826,
90
+ "grad_norm": 1.0726526975631714,
91
  "learning_rate": 2.9999988108959687e-05,
92
+ "loss": 0.7651,
93
  "step": 120
94
  },
95
  {
96
  "epoch": 5.6521739130434785,
97
+ "grad_norm": 1.5673664808273315,
98
  "learning_rate": 2.9999973245163716e-05,
99
+ "loss": 0.7415,
100
  "step": 130
101
  },
102
  {
103
  "epoch": 6.086956521739131,
104
+ "grad_norm": 1.9072725772857666,
105
  "learning_rate": 2.99999524358576e-05,
106
+ "loss": 0.7655,
107
  "step": 140
108
  },
109
  {
110
  "epoch": 6.521739130434782,
111
+ "grad_norm": 1.1216552257537842,
112
  "learning_rate": 2.9999925681049593e-05,
113
  "loss": 0.7857,
114
  "step": 150
115
  },
116
  {
117
  "epoch": 6.521739130434782,
118
+ "eval_loss": 0.7949807047843933,
119
+ "eval_runtime": 0.4775,
120
+ "eval_samples_per_second": 20.944,
121
+ "eval_steps_per_second": 20.944,
122
  "step": 150
123
  },
124
  {
125
  "epoch": 6.521739130434782,
126
+ "eval_loss": 0.8609212040901184,
127
+ "eval_runtime": 0.417,
128
+ "eval_samples_per_second": 23.983,
129
+ "eval_steps_per_second": 23.983,
130
  "step": 150
131
  },
132
  {
133
  "epoch": 6.521739130434782,
134
+ "eval_loss": 0.7949807047843933,
135
+ "eval_runtime": 0.4175,
136
+ "eval_samples_per_second": 23.952,
137
+ "eval_steps_per_second": 23.952,
138
  "step": 150
139
  },
140
  {
141
  "epoch": 6.521739130434782,
142
+ "eval_loss": 0.8609212040901184,
143
+ "eval_runtime": 0.4278,
144
+ "eval_samples_per_second": 23.374,
145
+ "eval_steps_per_second": 23.374,
146
  "step": 150
147
  },
148
  {
149
  "epoch": 6.521739130434782,
150
+ "eval_loss": 0.8608071208000183,
151
+ "eval_runtime": 0.4109,
152
+ "eval_samples_per_second": 24.338,
153
+ "eval_steps_per_second": 24.338,
154
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
155
  }
156
  ],
157
  "logging_steps": 10,
 
171
  "attributes": {}
172
  }
173
  },
174
+ "total_flos": 3894839614291968.0,
175
  "train_batch_size": 4,
176
  "trial_name": null,
177
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:003a11451b42f43906137235259ca60836693a9072ec318195749b927967a9e1
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1943b73dd12e0d46a3b26b62916dc0f34ed8d7e7a1a2985c54b619124519181f
3
  size 5368