slseanwu commited on
Commit
63c15e1
·
1 Parent(s): d9fd963

update to 5K steps

Browse files
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:be5a2cf78fc790af947f41c8f48603a500ccbc795cc1eef8fde9ede469227887
3
  size 4986380064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2ccf0728fbe58d56a20b6542840e576d33993d11d88e154175abf5b2b49df27b
3
  size 4986380064
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9aae98a11ed47e21fc54ba54c9fd57b0b7e38ad60b4530df163d94d434b14b0
3
  size 399532808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3490aab8dec625c1fbd420912aec0d3f5ccdc1fd0f6def94ebfff3c5ae62b25a
3
  size 399532808
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2962276faa60e57e48f24d0bcb0ffdf0d13f6f6ffd1675b63acf4be21b2611f3
3
  size 2699039674
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e8930ffbbd210d89b64f0d202a87a3741110a1fd51376dfc97da56cac9823b1
3
  size 2699039674
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9089141e94f72d9fcce2d7c290b3b5179d0af129b907289193048dd665d4fe8
3
- size 14180
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fddbe326fadb3ce8c449c2642e8a1dd894d5eabee5fd4473ac2c38deccaf19b
3
+ size 14244
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c69e416a1c1c7cfbd03505995e73f15a3c32b1cc0b0dc09a821d33b4e92286e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2be67f6aac8e482bb2022409709d8774ffb125292c0c9cf025c0ae747f3a6d57
3
  size 1064
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 1.721170395869191,
5
  "eval_steps": 500,
6
- "global_step": 1000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -93,6 +93,350 @@
93
  "eval_samples_per_second": 15.496,
94
  "eval_steps_per_second": 3.874,
95
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
96
  }
97
  ],
98
  "logging_steps": 100,
@@ -100,7 +444,7 @@
100
  "num_input_tokens_seen": 0,
101
  "num_train_epochs": 9,
102
  "save_steps": 500,
103
- "total_flos": 2.5172343959165338e+17,
104
  "train_batch_size": 4,
105
  "trial_name": null,
106
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 8.605851979345955,
5
  "eval_steps": 500,
6
+ "global_step": 5000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
93
  "eval_samples_per_second": 15.496,
94
  "eval_steps_per_second": 3.874,
95
  "step": 1000
96
+ },
97
+ {
98
+ "epoch": 1.89,
99
+ "grad_norm": 0.9721308946609497,
100
+ "learning_rate": 1.7333333333333336e-05,
101
+ "loss": 0.2547,
102
+ "step": 1100
103
+ },
104
+ {
105
+ "epoch": 2.07,
106
+ "grad_norm": 1.9259058237075806,
107
+ "learning_rate": 1.688888888888889e-05,
108
+ "loss": 0.2175,
109
+ "step": 1200
110
+ },
111
+ {
112
+ "epoch": 2.24,
113
+ "grad_norm": 1.3556184768676758,
114
+ "learning_rate": 1.6444444444444444e-05,
115
+ "loss": 0.1587,
116
+ "step": 1300
117
+ },
118
+ {
119
+ "epoch": 2.41,
120
+ "grad_norm": 1.4878870248794556,
121
+ "learning_rate": 1.6000000000000003e-05,
122
+ "loss": 0.1517,
123
+ "step": 1400
124
+ },
125
+ {
126
+ "epoch": 2.58,
127
+ "grad_norm": 1.0709530115127563,
128
+ "learning_rate": 1.555555555555556e-05,
129
+ "loss": 0.151,
130
+ "step": 1500
131
+ },
132
+ {
133
+ "epoch": 2.58,
134
+ "eval_loss": 0.23017294704914093,
135
+ "eval_runtime": 64.5369,
136
+ "eval_samples_per_second": 15.495,
137
+ "eval_steps_per_second": 3.874,
138
+ "step": 1500
139
+ },
140
+ {
141
+ "epoch": 2.75,
142
+ "grad_norm": 1.185964822769165,
143
+ "learning_rate": 1.5111111111111112e-05,
144
+ "loss": 0.1584,
145
+ "step": 1600
146
+ },
147
+ {
148
+ "epoch": 2.93,
149
+ "grad_norm": 1.543461561203003,
150
+ "learning_rate": 1.4666666666666666e-05,
151
+ "loss": 0.16,
152
+ "step": 1700
153
+ },
154
+ {
155
+ "epoch": 3.1,
156
+ "grad_norm": 1.1960724592208862,
157
+ "learning_rate": 1.4222222222222224e-05,
158
+ "loss": 0.1081,
159
+ "step": 1800
160
+ },
161
+ {
162
+ "epoch": 3.27,
163
+ "grad_norm": 0.9588227868080139,
164
+ "learning_rate": 1.377777777777778e-05,
165
+ "loss": 0.0811,
166
+ "step": 1900
167
+ },
168
+ {
169
+ "epoch": 3.44,
170
+ "grad_norm": 1.2663285732269287,
171
+ "learning_rate": 1.3333333333333333e-05,
172
+ "loss": 0.0825,
173
+ "step": 2000
174
+ },
175
+ {
176
+ "epoch": 3.44,
177
+ "eval_loss": 0.2604709267616272,
178
+ "eval_runtime": 64.5026,
179
+ "eval_samples_per_second": 15.503,
180
+ "eval_steps_per_second": 3.876,
181
+ "step": 2000
182
+ },
183
+ {
184
+ "epoch": 3.61,
185
+ "grad_norm": 1.2963746786117554,
186
+ "learning_rate": 1.288888888888889e-05,
187
+ "loss": 0.0833,
188
+ "step": 2100
189
+ },
190
+ {
191
+ "epoch": 3.79,
192
+ "grad_norm": 1.282989263534546,
193
+ "learning_rate": 1.2444444444444446e-05,
194
+ "loss": 0.0811,
195
+ "step": 2200
196
+ },
197
+ {
198
+ "epoch": 3.96,
199
+ "grad_norm": 1.1624544858932495,
200
+ "learning_rate": 1.2e-05,
201
+ "loss": 0.0786,
202
+ "step": 2300
203
+ },
204
+ {
205
+ "epoch": 4.13,
206
+ "grad_norm": 1.0563055276870728,
207
+ "learning_rate": 1.1555555555555556e-05,
208
+ "loss": 0.0496,
209
+ "step": 2400
210
+ },
211
+ {
212
+ "epoch": 4.3,
213
+ "grad_norm": 1.2742573022842407,
214
+ "learning_rate": 1.1111111111111113e-05,
215
+ "loss": 0.0398,
216
+ "step": 2500
217
+ },
218
+ {
219
+ "epoch": 4.3,
220
+ "eval_loss": 0.3238039016723633,
221
+ "eval_runtime": 64.5078,
222
+ "eval_samples_per_second": 15.502,
223
+ "eval_steps_per_second": 3.876,
224
+ "step": 2500
225
+ },
226
+ {
227
+ "epoch": 4.48,
228
+ "grad_norm": 1.1727477312088013,
229
+ "learning_rate": 1.0666666666666667e-05,
230
+ "loss": 0.0401,
231
+ "step": 2600
232
+ },
233
+ {
234
+ "epoch": 4.65,
235
+ "grad_norm": 1.1197234392166138,
236
+ "learning_rate": 1.0222222222222223e-05,
237
+ "loss": 0.042,
238
+ "step": 2700
239
+ },
240
+ {
241
+ "epoch": 4.82,
242
+ "grad_norm": 0.9968566298484802,
243
+ "learning_rate": 9.777777777777779e-06,
244
+ "loss": 0.0431,
245
+ "step": 2800
246
+ },
247
+ {
248
+ "epoch": 4.99,
249
+ "grad_norm": 0.8295079469680786,
250
+ "learning_rate": 9.333333333333334e-06,
251
+ "loss": 0.0422,
252
+ "step": 2900
253
+ },
254
+ {
255
+ "epoch": 5.16,
256
+ "grad_norm": 0.4814624786376953,
257
+ "learning_rate": 8.888888888888888e-06,
258
+ "loss": 0.0191,
259
+ "step": 3000
260
+ },
261
+ {
262
+ "epoch": 5.16,
263
+ "eval_loss": 0.3760123550891876,
264
+ "eval_runtime": 64.4488,
265
+ "eval_samples_per_second": 15.516,
266
+ "eval_steps_per_second": 3.879,
267
+ "step": 3000
268
+ },
269
+ {
270
+ "epoch": 5.34,
271
+ "grad_norm": 1.421600341796875,
272
+ "learning_rate": 8.444444444444446e-06,
273
+ "loss": 0.0189,
274
+ "step": 3100
275
+ },
276
+ {
277
+ "epoch": 5.51,
278
+ "grad_norm": 0.989109218120575,
279
+ "learning_rate": 8.000000000000001e-06,
280
+ "loss": 0.0192,
281
+ "step": 3200
282
+ },
283
+ {
284
+ "epoch": 5.68,
285
+ "grad_norm": 0.7597835063934326,
286
+ "learning_rate": 7.555555555555556e-06,
287
+ "loss": 0.02,
288
+ "step": 3300
289
+ },
290
+ {
291
+ "epoch": 5.85,
292
+ "grad_norm": 0.9319747686386108,
293
+ "learning_rate": 7.111111111111112e-06,
294
+ "loss": 0.0212,
295
+ "step": 3400
296
+ },
297
+ {
298
+ "epoch": 6.02,
299
+ "grad_norm": 0.8922848105430603,
300
+ "learning_rate": 6.666666666666667e-06,
301
+ "loss": 0.02,
302
+ "step": 3500
303
+ },
304
+ {
305
+ "epoch": 6.02,
306
+ "eval_loss": 0.34943887591362,
307
+ "eval_runtime": 64.4584,
308
+ "eval_samples_per_second": 15.514,
309
+ "eval_steps_per_second": 3.878,
310
+ "step": 3500
311
+ },
312
+ {
313
+ "epoch": 6.2,
314
+ "grad_norm": 0.4450746178627014,
315
+ "learning_rate": 6.222222222222223e-06,
316
+ "loss": 0.0084,
317
+ "step": 3600
318
+ },
319
+ {
320
+ "epoch": 6.37,
321
+ "grad_norm": 0.4923359751701355,
322
+ "learning_rate": 5.777777777777778e-06,
323
+ "loss": 0.0093,
324
+ "step": 3700
325
+ },
326
+ {
327
+ "epoch": 6.54,
328
+ "grad_norm": 0.7262879014015198,
329
+ "learning_rate": 5.333333333333334e-06,
330
+ "loss": 0.0083,
331
+ "step": 3800
332
+ },
333
+ {
334
+ "epoch": 6.71,
335
+ "grad_norm": 0.7225219011306763,
336
+ "learning_rate": 4.888888888888889e-06,
337
+ "loss": 0.0101,
338
+ "step": 3900
339
+ },
340
+ {
341
+ "epoch": 6.88,
342
+ "grad_norm": 0.6946544647216797,
343
+ "learning_rate": 4.444444444444444e-06,
344
+ "loss": 0.0089,
345
+ "step": 4000
346
+ },
347
+ {
348
+ "epoch": 6.88,
349
+ "eval_loss": 0.3759503960609436,
350
+ "eval_runtime": 64.5688,
351
+ "eval_samples_per_second": 15.487,
352
+ "eval_steps_per_second": 3.872,
353
+ "step": 4000
354
+ },
355
+ {
356
+ "epoch": 7.06,
357
+ "grad_norm": 0.4038643538951874,
358
+ "learning_rate": 4.000000000000001e-06,
359
+ "loss": 0.0066,
360
+ "step": 4100
361
+ },
362
+ {
363
+ "epoch": 7.23,
364
+ "grad_norm": 0.6771142482757568,
365
+ "learning_rate": 3.555555555555556e-06,
366
+ "loss": 0.0047,
367
+ "step": 4200
368
+ },
369
+ {
370
+ "epoch": 7.4,
371
+ "grad_norm": 0.3253248631954193,
372
+ "learning_rate": 3.1111111111111116e-06,
373
+ "loss": 0.0035,
374
+ "step": 4300
375
+ },
376
+ {
377
+ "epoch": 7.57,
378
+ "grad_norm": 0.11476542800664902,
379
+ "learning_rate": 2.666666666666667e-06,
380
+ "loss": 0.0035,
381
+ "step": 4400
382
+ },
383
+ {
384
+ "epoch": 7.75,
385
+ "grad_norm": 0.12515957653522491,
386
+ "learning_rate": 2.222222222222222e-06,
387
+ "loss": 0.0036,
388
+ "step": 4500
389
+ },
390
+ {
391
+ "epoch": 7.75,
392
+ "eval_loss": 0.3938386142253876,
393
+ "eval_runtime": 64.4918,
394
+ "eval_samples_per_second": 15.506,
395
+ "eval_steps_per_second": 3.876,
396
+ "step": 4500
397
+ },
398
+ {
399
+ "epoch": 7.92,
400
+ "grad_norm": 0.31758418679237366,
401
+ "learning_rate": 1.777777777777778e-06,
402
+ "loss": 0.0042,
403
+ "step": 4600
404
+ },
405
+ {
406
+ "epoch": 8.09,
407
+ "grad_norm": 0.08976474404335022,
408
+ "learning_rate": 1.3333333333333334e-06,
409
+ "loss": 0.0035,
410
+ "step": 4700
411
+ },
412
+ {
413
+ "epoch": 8.26,
414
+ "grad_norm": 0.1867213398218155,
415
+ "learning_rate": 8.88888888888889e-07,
416
+ "loss": 0.0026,
417
+ "step": 4800
418
+ },
419
+ {
420
+ "epoch": 8.43,
421
+ "grad_norm": 0.10856834053993225,
422
+ "learning_rate": 4.444444444444445e-07,
423
+ "loss": 0.0018,
424
+ "step": 4900
425
+ },
426
+ {
427
+ "epoch": 8.61,
428
+ "grad_norm": 0.15036863088607788,
429
+ "learning_rate": 0.0,
430
+ "loss": 0.0019,
431
+ "step": 5000
432
+ },
433
+ {
434
+ "epoch": 8.61,
435
+ "eval_loss": 0.4249526262283325,
436
+ "eval_runtime": 64.5498,
437
+ "eval_samples_per_second": 15.492,
438
+ "eval_steps_per_second": 3.873,
439
+ "step": 5000
440
  }
441
  ],
442
  "logging_steps": 100,
 
444
  "num_input_tokens_seen": 0,
445
  "num_train_epochs": 9,
446
  "save_steps": 500,
447
+ "total_flos": 1.258569996863275e+18,
448
  "train_batch_size": 4,
449
  "trial_name": null,
450
  "trial_params": null