File size: 9,459 Bytes
0bb9041
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 8.330556481172943,
  "eval_steps": 500,
  "global_step": 25000,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.16661112962345884,
      "grad_norm": 2.0231502056121826,
      "learning_rate": 2.776852160390981e-06,
      "loss": 8.4699,
      "step": 500
    },
    {
      "epoch": 0.3332222592469177,
      "grad_norm": 1.640729546546936,
      "learning_rate": 5.553704320781962e-06,
      "loss": 7.446,
      "step": 1000
    },
    {
      "epoch": 0.49983338887037654,
      "grad_norm": 1.6770328283309937,
      "learning_rate": 8.330556481172942e-06,
      "loss": 6.6034,
      "step": 1500
    },
    {
      "epoch": 0.6664445184938353,
      "grad_norm": 2.338146924972534,
      "learning_rate": 1.1107408641563923e-05,
      "loss": 5.9787,
      "step": 2000
    },
    {
      "epoch": 0.8330556481172943,
      "grad_norm": 2.557213068008423,
      "learning_rate": 1.3884260801954905e-05,
      "loss": 5.5732,
      "step": 2500
    },
    {
      "epoch": 0.9996667777407531,
      "grad_norm": 1.7546786069869995,
      "learning_rate": 1.6661112962345883e-05,
      "loss": 5.2881,
      "step": 3000
    },
    {
      "epoch": 1.166277907364212,
      "grad_norm": 1.7593083381652832,
      "learning_rate": 1.9437965122736867e-05,
      "loss": 5.0771,
      "step": 3500
    },
    {
      "epoch": 1.3328890369876707,
      "grad_norm": 1.860612392425537,
      "learning_rate": 2.2214817283127847e-05,
      "loss": 4.8991,
      "step": 4000
    },
    {
      "epoch": 1.4995001666111296,
      "grad_norm": 2.036371946334839,
      "learning_rate": 2.4991669443518826e-05,
      "loss": 4.7465,
      "step": 4500
    },
    {
      "epoch": 1.6661112962345883,
      "grad_norm": 1.8470864295959473,
      "learning_rate": 2.776852160390981e-05,
      "loss": 4.5925,
      "step": 5000
    },
    {
      "epoch": 1.8327224258580475,
      "grad_norm": 1.9861804246902466,
      "learning_rate": 3.0545373764300786e-05,
      "loss": 4.4767,
      "step": 5500
    },
    {
      "epoch": 1.9993335554815062,
      "grad_norm": 2.018094062805176,
      "learning_rate": 3.3322225924691766e-05,
      "loss": 4.3708,
      "step": 6000
    },
    {
      "epoch": 2.165944685104965,
      "grad_norm": 1.8410370349884033,
      "learning_rate": 3.609352438076197e-05,
      "loss": 4.2741,
      "step": 6500
    },
    {
      "epoch": 2.332555814728424,
      "grad_norm": 1.7712337970733643,
      "learning_rate": 3.887037654115295e-05,
      "loss": 4.192,
      "step": 7000
    },
    {
      "epoch": 2.4991669443518827,
      "grad_norm": 1.7352423667907715,
      "learning_rate": 4.164722870154393e-05,
      "loss": 4.1192,
      "step": 7500
    },
    {
      "epoch": 2.6657780739753414,
      "grad_norm": 1.8459163904190063,
      "learning_rate": 4.442408086193491e-05,
      "loss": 4.0351,
      "step": 8000
    },
    {
      "epoch": 2.8323892035988005,
      "grad_norm": 2.009974956512451,
      "learning_rate": 4.719537931800511e-05,
      "loss": 3.9709,
      "step": 8500
    },
    {
      "epoch": 2.9990003332222592,
      "grad_norm": 1.7068768739700317,
      "learning_rate": 4.997223147839609e-05,
      "loss": 3.899,
      "step": 9000
    },
    {
      "epoch": 3.165611462845718,
      "grad_norm": 1.6234753131866455,
      "learning_rate": 5.274908363878708e-05,
      "loss": 3.8237,
      "step": 9500
    },
    {
      "epoch": 3.332222592469177,
      "grad_norm": 1.786528468132019,
      "learning_rate": 5.552593579917805e-05,
      "loss": 3.7704,
      "step": 10000
    },
    {
      "epoch": 3.4988337220926358,
      "grad_norm": 1.7539526224136353,
      "learning_rate": 5.829723425524826e-05,
      "loss": 3.7183,
      "step": 10500
    },
    {
      "epoch": 3.665444851716095,
      "grad_norm": 1.6715501546859741,
      "learning_rate": 6.107408641563923e-05,
      "loss": 3.6671,
      "step": 11000
    },
    {
      "epoch": 3.8320559813395536,
      "grad_norm": 1.7561357021331787,
      "learning_rate": 6.385093857603022e-05,
      "loss": 3.6214,
      "step": 11500
    },
    {
      "epoch": 3.9986671109630123,
      "grad_norm": 1.5936201810836792,
      "learning_rate": 6.662779073642119e-05,
      "loss": 3.5782,
      "step": 12000
    },
    {
      "epoch": 4.1652782405864714,
      "grad_norm": 1.7363481521606445,
      "learning_rate": 6.940464289681218e-05,
      "loss": 3.5291,
      "step": 12500
    },
    {
      "epoch": 4.33188937020993,
      "grad_norm": 1.5552008152008057,
      "learning_rate": 7.217594135288238e-05,
      "loss": 3.5017,
      "step": 13000
    },
    {
      "epoch": 4.498500499833389,
      "grad_norm": 1.4608962535858154,
      "learning_rate": 7.495279351327336e-05,
      "loss": 3.4685,
      "step": 13500
    },
    {
      "epoch": 4.665111629456848,
      "grad_norm": 1.470942735671997,
      "learning_rate": 7.772964567366434e-05,
      "loss": 3.4407,
      "step": 14000
    },
    {
      "epoch": 4.831722759080306,
      "grad_norm": 1.4744900465011597,
      "learning_rate": 8.050649783405532e-05,
      "loss": 3.4156,
      "step": 14500
    },
    {
      "epoch": 4.998333888703765,
      "grad_norm": 1.449549913406372,
      "learning_rate": 8.32833499944463e-05,
      "loss": 3.3933,
      "step": 15000
    },
    {
      "epoch": 5.1649450183272245,
      "grad_norm": 1.4103004932403564,
      "learning_rate": 8.606020215483728e-05,
      "loss": 3.355,
      "step": 15500
    },
    {
      "epoch": 5.331556147950683,
      "grad_norm": 1.3490815162658691,
      "learning_rate": 8.883705431522826e-05,
      "loss": 3.335,
      "step": 16000
    },
    {
      "epoch": 5.498167277574142,
      "grad_norm": 1.3762222528457642,
      "learning_rate": 9.161390647561924e-05,
      "loss": 3.3183,
      "step": 16500
    },
    {
      "epoch": 5.664778407197601,
      "grad_norm": 1.3008846044540405,
      "learning_rate": 9.438520493168944e-05,
      "loss": 3.3013,
      "step": 17000
    },
    {
      "epoch": 5.831389536821059,
      "grad_norm": 1.3132846355438232,
      "learning_rate": 9.716205709208043e-05,
      "loss": 3.2846,
      "step": 17500
    },
    {
      "epoch": 5.9980006664445185,
      "grad_norm": 1.194407343864441,
      "learning_rate": 9.99389092524714e-05,
      "loss": 3.2706,
      "step": 18000
    },
    {
      "epoch": 6.164611796067978,
      "grad_norm": 1.1745057106018066,
      "learning_rate": 9.996657887827809e-05,
      "loss": 3.2324,
      "step": 18500
    },
    {
      "epoch": 6.331222925691436,
      "grad_norm": 1.1769245862960815,
      "learning_rate": 9.986361454903414e-05,
      "loss": 3.2237,
      "step": 19000
    },
    {
      "epoch": 6.497834055314895,
      "grad_norm": 1.1529046297073364,
      "learning_rate": 9.969082137309689e-05,
      "loss": 3.2103,
      "step": 19500
    },
    {
      "epoch": 6.664445184938354,
      "grad_norm": 1.1225526332855225,
      "learning_rate": 9.944857739312099e-05,
      "loss": 3.1936,
      "step": 20000
    },
    {
      "epoch": 6.831056314561812,
      "grad_norm": 1.1167833805084229,
      "learning_rate": 9.913722118348934e-05,
      "loss": 3.1818,
      "step": 20500
    },
    {
      "epoch": 6.9976674441852715,
      "grad_norm": 1.093309998512268,
      "learning_rate": 9.875801618300155e-05,
      "loss": 3.1733,
      "step": 21000
    },
    {
      "epoch": 7.164278573808731,
      "grad_norm": 1.1539673805236816,
      "learning_rate": 9.83099727084671e-05,
      "loss": 3.1342,
      "step": 21500
    },
    {
      "epoch": 7.330889703432189,
      "grad_norm": 1.1930997371673584,
      "learning_rate": 9.779440838491111e-05,
      "loss": 3.1356,
      "step": 22000
    },
    {
      "epoch": 7.497500833055648,
      "grad_norm": 1.0827438831329346,
      "learning_rate": 9.721204379524642e-05,
      "loss": 3.1231,
      "step": 22500
    },
    {
      "epoch": 7.664111962679107,
      "grad_norm": 1.0850555896759033,
      "learning_rate": 9.656505485544595e-05,
      "loss": 3.1126,
      "step": 23000
    },
    {
      "epoch": 7.8307230923025655,
      "grad_norm": 1.0119590759277344,
      "learning_rate": 9.585175299579849e-05,
      "loss": 3.1049,
      "step": 23500
    },
    {
      "epoch": 7.997334221926025,
      "grad_norm": 1.0027412176132202,
      "learning_rate": 9.507436603906176e-05,
      "loss": 3.1004,
      "step": 24000
    },
    {
      "epoch": 8.163945351549483,
      "grad_norm": 1.0899256467819214,
      "learning_rate": 9.423398050680459e-05,
      "loss": 3.0591,
      "step": 24500
    },
    {
      "epoch": 8.330556481172943,
      "grad_norm": 1.0424436330795288,
      "learning_rate": 9.333363626660517e-05,
      "loss": 3.0582,
      "step": 25000
    }
  ],
  "logging_steps": 500,
  "max_steps": 60020,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 20,
  "save_steps": 1000,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 1.046659236626432e+17,
  "train_batch_size": 16,
  "trial_name": null,
  "trial_params": null
}