File size: 8,276 Bytes
1010106
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
{
  "best_global_step": null,
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 5.0,
  "eval_steps": 500,
  "global_step": 21930,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.11399908800729594,
      "grad_norm": 1.4812531471252441,
      "learning_rate": 1.9544915640674875e-05,
      "loss": 0.1287,
      "step": 500
    },
    {
      "epoch": 0.22799817601459188,
      "grad_norm": 2.198870897293091,
      "learning_rate": 1.9088919288645692e-05,
      "loss": 0.1045,
      "step": 1000
    },
    {
      "epoch": 0.34199726402188785,
      "grad_norm": 1.3517733812332153,
      "learning_rate": 1.863292293661651e-05,
      "loss": 0.0895,
      "step": 1500
    },
    {
      "epoch": 0.45599635202918376,
      "grad_norm": 1.352580189704895,
      "learning_rate": 1.8176926584587324e-05,
      "loss": 0.0766,
      "step": 2000
    },
    {
      "epoch": 0.5699954400364797,
      "grad_norm": 1.960016131401062,
      "learning_rate": 1.772093023255814e-05,
      "loss": 0.0713,
      "step": 2500
    },
    {
      "epoch": 0.6839945280437757,
      "grad_norm": 0.5876455903053284,
      "learning_rate": 1.7264933880528958e-05,
      "loss": 0.0657,
      "step": 3000
    },
    {
      "epoch": 0.7979936160510716,
      "grad_norm": 0.9299254417419434,
      "learning_rate": 1.6808937528499772e-05,
      "loss": 0.0623,
      "step": 3500
    },
    {
      "epoch": 0.9119927040583675,
      "grad_norm": 1.3892526626586914,
      "learning_rate": 1.635294117647059e-05,
      "loss": 0.0621,
      "step": 4000
    },
    {
      "epoch": 1.0259917920656634,
      "grad_norm": 1.7030184268951416,
      "learning_rate": 1.5896944824441403e-05,
      "loss": 0.0585,
      "step": 4500
    },
    {
      "epoch": 1.1399908800729595,
      "grad_norm": 0.9616603255271912,
      "learning_rate": 1.5440948472412224e-05,
      "loss": 0.0448,
      "step": 5000
    },
    {
      "epoch": 1.2539899680802553,
      "grad_norm": 1.245125412940979,
      "learning_rate": 1.4984952120383038e-05,
      "loss": 0.0429,
      "step": 5500
    },
    {
      "epoch": 1.3679890560875512,
      "grad_norm": 1.2846943140029907,
      "learning_rate": 1.4528955768353854e-05,
      "loss": 0.0408,
      "step": 6000
    },
    {
      "epoch": 1.4819881440948472,
      "grad_norm": 1.0010948181152344,
      "learning_rate": 1.407295941632467e-05,
      "loss": 0.0386,
      "step": 6500
    },
    {
      "epoch": 1.5959872321021433,
      "grad_norm": 1.1298205852508545,
      "learning_rate": 1.3616963064295488e-05,
      "loss": 0.0364,
      "step": 7000
    },
    {
      "epoch": 1.7099863201094392,
      "grad_norm": 0.5932161211967468,
      "learning_rate": 1.3160966712266304e-05,
      "loss": 0.0344,
      "step": 7500
    },
    {
      "epoch": 1.823985408116735,
      "grad_norm": 1.195202112197876,
      "learning_rate": 1.270497036023712e-05,
      "loss": 0.033,
      "step": 8000
    },
    {
      "epoch": 1.937984496124031,
      "grad_norm": 1.2507325410842896,
      "learning_rate": 1.2248974008207935e-05,
      "loss": 0.0321,
      "step": 8500
    },
    {
      "epoch": 2.0519835841313268,
      "grad_norm": 2.5148258209228516,
      "learning_rate": 1.1792977656178753e-05,
      "loss": 0.0278,
      "step": 9000
    },
    {
      "epoch": 2.165982672138623,
      "grad_norm": 0.7329283356666565,
      "learning_rate": 1.1336981304149568e-05,
      "loss": 0.0233,
      "step": 9500
    },
    {
      "epoch": 2.279981760145919,
      "grad_norm": 0.42915332317352295,
      "learning_rate": 1.0880984952120384e-05,
      "loss": 0.0226,
      "step": 10000
    },
    {
      "epoch": 2.3939808481532148,
      "grad_norm": 0.7090346813201904,
      "learning_rate": 1.04249886000912e-05,
      "loss": 0.0223,
      "step": 10500
    },
    {
      "epoch": 2.5079799361605106,
      "grad_norm": 0.6480665802955627,
      "learning_rate": 9.968992248062017e-06,
      "loss": 0.0204,
      "step": 11000
    },
    {
      "epoch": 2.621979024167807,
      "grad_norm": 1.7029985189437866,
      "learning_rate": 9.512995896032832e-06,
      "loss": 0.0221,
      "step": 11500
    },
    {
      "epoch": 2.7359781121751023,
      "grad_norm": 0.9722331166267395,
      "learning_rate": 9.056999544003648e-06,
      "loss": 0.0208,
      "step": 12000
    },
    {
      "epoch": 2.8499772001823986,
      "grad_norm": 0.3082011938095093,
      "learning_rate": 8.601003191974465e-06,
      "loss": 0.0204,
      "step": 12500
    },
    {
      "epoch": 2.9639762881896945,
      "grad_norm": 0.33128827810287476,
      "learning_rate": 8.145006839945281e-06,
      "loss": 0.0207,
      "step": 13000
    },
    {
      "epoch": 3.0779753761969904,
      "grad_norm": 0.7698410749435425,
      "learning_rate": 7.689010487916098e-06,
      "loss": 0.0171,
      "step": 13500
    },
    {
      "epoch": 3.191974464204286,
      "grad_norm": 0.7875366806983948,
      "learning_rate": 7.233014135886913e-06,
      "loss": 0.0156,
      "step": 14000
    },
    {
      "epoch": 3.305973552211582,
      "grad_norm": 0.6971263289451599,
      "learning_rate": 6.77701778385773e-06,
      "loss": 0.0157,
      "step": 14500
    },
    {
      "epoch": 3.4199726402188784,
      "grad_norm": 0.464100182056427,
      "learning_rate": 6.321021431828546e-06,
      "loss": 0.0143,
      "step": 15000
    },
    {
      "epoch": 3.5339717282261742,
      "grad_norm": 0.5803436636924744,
      "learning_rate": 5.8650250797993626e-06,
      "loss": 0.0147,
      "step": 15500
    },
    {
      "epoch": 3.64797081623347,
      "grad_norm": 0.7879688739776611,
      "learning_rate": 5.409028727770178e-06,
      "loss": 0.0137,
      "step": 16000
    },
    {
      "epoch": 3.761969904240766,
      "grad_norm": 1.0888534784317017,
      "learning_rate": 4.953032375740995e-06,
      "loss": 0.0139,
      "step": 16500
    },
    {
      "epoch": 3.875968992248062,
      "grad_norm": 0.3893286883831024,
      "learning_rate": 4.49703602371181e-06,
      "loss": 0.0137,
      "step": 17000
    },
    {
      "epoch": 3.989968080255358,
      "grad_norm": 0.4917721748352051,
      "learning_rate": 4.041039671682627e-06,
      "loss": 0.0136,
      "step": 17500
    },
    {
      "epoch": 4.1039671682626535,
      "grad_norm": 1.25238835811615,
      "learning_rate": 3.585043319653443e-06,
      "loss": 0.0117,
      "step": 18000
    },
    {
      "epoch": 4.21796625626995,
      "grad_norm": 0.9064317941665649,
      "learning_rate": 3.129046967624259e-06,
      "loss": 0.0108,
      "step": 18500
    },
    {
      "epoch": 4.331965344277246,
      "grad_norm": 0.4522368311882019,
      "learning_rate": 2.6730506155950754e-06,
      "loss": 0.0105,
      "step": 19000
    },
    {
      "epoch": 4.4459644322845415,
      "grad_norm": 1.0190762281417847,
      "learning_rate": 2.217054263565892e-06,
      "loss": 0.0108,
      "step": 19500
    },
    {
      "epoch": 4.559963520291838,
      "grad_norm": 0.23909200727939606,
      "learning_rate": 1.7610579115367079e-06,
      "loss": 0.0114,
      "step": 20000
    },
    {
      "epoch": 4.673962608299133,
      "grad_norm": 0.8679990768432617,
      "learning_rate": 1.3050615595075241e-06,
      "loss": 0.0105,
      "step": 20500
    },
    {
      "epoch": 4.7879616963064295,
      "grad_norm": 0.4421948790550232,
      "learning_rate": 8.490652074783402e-07,
      "loss": 0.0109,
      "step": 21000
    },
    {
      "epoch": 4.901960784313726,
      "grad_norm": 1.2229384183883667,
      "learning_rate": 3.930688554491564e-07,
      "loss": 0.0103,
      "step": 21500
    }
  ],
  "logging_steps": 500,
  "max_steps": 21930,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 5,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 2352410932543488.0,
  "train_batch_size": 16,
  "trial_name": null,
  "trial_params": null
}