File size: 9,935 Bytes
109c8cb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
{
  "best_global_step": 24806,
  "best_metric": 0.13123206794261932,
  "best_model_checkpoint": "Model1-v1-Rival/checkpoint-24806",
  "epoch": 2.0,
  "eval_steps": 500,
  "global_step": 24806,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.04031282754172378,
      "grad_norm": 35.80685043334961,
      "learning_rate": 2.011285771866183e-07,
      "loss": 0.7492,
      "step": 500
    },
    {
      "epoch": 0.08062565508344756,
      "grad_norm": 27.830657958984375,
      "learning_rate": 4.0266021765417167e-07,
      "loss": 0.564,
      "step": 1000
    },
    {
      "epoch": 0.12093848262517133,
      "grad_norm": 22.993253707885742,
      "learning_rate": 6.041918581217252e-07,
      "loss": 0.497,
      "step": 1500
    },
    {
      "epoch": 0.16125131016689512,
      "grad_norm": 31.184377670288086,
      "learning_rate": 8.057234985892785e-07,
      "loss": 0.4362,
      "step": 2000
    },
    {
      "epoch": 0.2015641377086189,
      "grad_norm": 31.91282844543457,
      "learning_rate": 9.991937290033595e-07,
      "loss": 0.41,
      "step": 2500
    },
    {
      "epoch": 0.24187696525034266,
      "grad_norm": 42.18364715576172,
      "learning_rate": 9.767973124300112e-07,
      "loss": 0.3815,
      "step": 3000
    },
    {
      "epoch": 0.28218979279206646,
      "grad_norm": 31.624061584472656,
      "learning_rate": 9.544008958566627e-07,
      "loss": 0.3568,
      "step": 3500
    },
    {
      "epoch": 0.32250262033379024,
      "grad_norm": 36.55485534667969,
      "learning_rate": 9.320044792833146e-07,
      "loss": 0.341,
      "step": 4000
    },
    {
      "epoch": 0.362815447875514,
      "grad_norm": 34.942264556884766,
      "learning_rate": 9.096080627099664e-07,
      "loss": 0.3293,
      "step": 4500
    },
    {
      "epoch": 0.4031282754172378,
      "grad_norm": 30.162443161010742,
      "learning_rate": 8.872116461366181e-07,
      "loss": 0.3184,
      "step": 5000
    },
    {
      "epoch": 0.44344110295896155,
      "grad_norm": 42.151607513427734,
      "learning_rate": 8.648152295632698e-07,
      "loss": 0.3149,
      "step": 5500
    },
    {
      "epoch": 0.4837539305006853,
      "grad_norm": 21.629465103149414,
      "learning_rate": 8.424188129899216e-07,
      "loss": 0.2969,
      "step": 6000
    },
    {
      "epoch": 0.5240667580424091,
      "grad_norm": 45.252803802490234,
      "learning_rate": 8.200223964165734e-07,
      "loss": 0.2962,
      "step": 6500
    },
    {
      "epoch": 0.5643795855841329,
      "grad_norm": 31.193838119506836,
      "learning_rate": 7.97625979843225e-07,
      "loss": 0.2858,
      "step": 7000
    },
    {
      "epoch": 0.6046924131258566,
      "grad_norm": 96.9237289428711,
      "learning_rate": 7.752295632698768e-07,
      "loss": 0.268,
      "step": 7500
    },
    {
      "epoch": 0.6450052406675805,
      "grad_norm": 21.470144271850586,
      "learning_rate": 7.528331466965285e-07,
      "loss": 0.2622,
      "step": 8000
    },
    {
      "epoch": 0.6853180682093042,
      "grad_norm": 35.70467758178711,
      "learning_rate": 7.304367301231803e-07,
      "loss": 0.2741,
      "step": 8500
    },
    {
      "epoch": 0.725630895751028,
      "grad_norm": 19.73318862915039,
      "learning_rate": 7.080403135498319e-07,
      "loss": 0.2626,
      "step": 9000
    },
    {
      "epoch": 0.7659437232927517,
      "grad_norm": 23.529918670654297,
      "learning_rate": 6.856438969764838e-07,
      "loss": 0.2518,
      "step": 9500
    },
    {
      "epoch": 0.8062565508344756,
      "grad_norm": 41.51838302612305,
      "learning_rate": 6.632474804031355e-07,
      "loss": 0.2495,
      "step": 10000
    },
    {
      "epoch": 0.8465693783761993,
      "grad_norm": 14.217453956604004,
      "learning_rate": 6.408510638297872e-07,
      "loss": 0.2437,
      "step": 10500
    },
    {
      "epoch": 0.8868822059179231,
      "grad_norm": 18.972307205200195,
      "learning_rate": 6.184546472564389e-07,
      "loss": 0.2482,
      "step": 11000
    },
    {
      "epoch": 0.9271950334596468,
      "grad_norm": 47.28855514526367,
      "learning_rate": 5.960582306830907e-07,
      "loss": 0.2419,
      "step": 11500
    },
    {
      "epoch": 0.9675078610013707,
      "grad_norm": 48.309165954589844,
      "learning_rate": 5.736618141097424e-07,
      "loss": 0.2421,
      "step": 12000
    },
    {
      "epoch": 1.0,
      "eval_accuracy": 0.9340860036526648,
      "eval_loss": 0.17090356349945068,
      "eval_model_preparation_time": 0.0026,
      "eval_runtime": 109.2506,
      "eval_samples_per_second": 441.087,
      "eval_steps_per_second": 55.139,
      "step": 12403
    },
    {
      "epoch": 1.0078206885430945,
      "grad_norm": 47.087913513183594,
      "learning_rate": 5.512653975363942e-07,
      "loss": 0.2372,
      "step": 12500
    },
    {
      "epoch": 1.0481335160848182,
      "grad_norm": 40.675697326660156,
      "learning_rate": 5.288689809630459e-07,
      "loss": 0.2249,
      "step": 13000
    },
    {
      "epoch": 1.088446343626542,
      "grad_norm": 46.77396774291992,
      "learning_rate": 5.064725643896977e-07,
      "loss": 0.2207,
      "step": 13500
    },
    {
      "epoch": 1.1287591711682659,
      "grad_norm": 69.7470703125,
      "learning_rate": 4.840761478163493e-07,
      "loss": 0.2276,
      "step": 14000
    },
    {
      "epoch": 1.1690719987099896,
      "grad_norm": 35.32638168334961,
      "learning_rate": 4.6167973124300107e-07,
      "loss": 0.2235,
      "step": 14500
    },
    {
      "epoch": 1.2093848262517133,
      "grad_norm": 63.080291748046875,
      "learning_rate": 4.3928331466965287e-07,
      "loss": 0.2219,
      "step": 15000
    },
    {
      "epoch": 1.249697653793437,
      "grad_norm": 45.029239654541016,
      "learning_rate": 4.1688689809630457e-07,
      "loss": 0.2184,
      "step": 15500
    },
    {
      "epoch": 1.2900104813351607,
      "grad_norm": 20.416336059570312,
      "learning_rate": 3.944904815229563e-07,
      "loss": 0.2081,
      "step": 16000
    },
    {
      "epoch": 1.3303233088768847,
      "grad_norm": 28.367952346801758,
      "learning_rate": 3.72094064949608e-07,
      "loss": 0.2157,
      "step": 16500
    },
    {
      "epoch": 1.3706361364186084,
      "grad_norm": 13.548885345458984,
      "learning_rate": 3.4969764837625977e-07,
      "loss": 0.212,
      "step": 17000
    },
    {
      "epoch": 1.4109489639603323,
      "grad_norm": 94.12673950195312,
      "learning_rate": 3.273012318029115e-07,
      "loss": 0.2075,
      "step": 17500
    },
    {
      "epoch": 1.451261791502056,
      "grad_norm": 38.12338638305664,
      "learning_rate": 3.049048152295632e-07,
      "loss": 0.2148,
      "step": 18000
    },
    {
      "epoch": 1.4915746190437797,
      "grad_norm": 45.156455993652344,
      "learning_rate": 2.82508398656215e-07,
      "loss": 0.2075,
      "step": 18500
    },
    {
      "epoch": 1.5318874465855035,
      "grad_norm": 18.083568572998047,
      "learning_rate": 2.601119820828667e-07,
      "loss": 0.2092,
      "step": 19000
    },
    {
      "epoch": 1.5722002741272272,
      "grad_norm": 27.93409538269043,
      "learning_rate": 2.3771556550951847e-07,
      "loss": 0.205,
      "step": 19500
    },
    {
      "epoch": 1.612513101668951,
      "grad_norm": 39.768192291259766,
      "learning_rate": 2.153191489361702e-07,
      "loss": 0.2039,
      "step": 20000
    },
    {
      "epoch": 1.6528259292106748,
      "grad_norm": 28.78324317932129,
      "learning_rate": 1.9292273236282195e-07,
      "loss": 0.204,
      "step": 20500
    },
    {
      "epoch": 1.6931387567523988,
      "grad_norm": 94.95115661621094,
      "learning_rate": 1.7052631578947368e-07,
      "loss": 0.2154,
      "step": 21000
    },
    {
      "epoch": 1.7334515842941225,
      "grad_norm": 51.76600646972656,
      "learning_rate": 1.4812989921612543e-07,
      "loss": 0.208,
      "step": 21500
    },
    {
      "epoch": 1.7737644118358462,
      "grad_norm": 24.089155197143555,
      "learning_rate": 1.2573348264277713e-07,
      "loss": 0.2071,
      "step": 22000
    },
    {
      "epoch": 1.81407723937757,
      "grad_norm": 46.96974182128906,
      "learning_rate": 1.0333706606942888e-07,
      "loss": 0.198,
      "step": 22500
    },
    {
      "epoch": 1.8543900669192936,
      "grad_norm": 66.37223815917969,
      "learning_rate": 8.094064949608062e-08,
      "loss": 0.2015,
      "step": 23000
    },
    {
      "epoch": 1.8947028944610174,
      "grad_norm": 42.39664077758789,
      "learning_rate": 5.854423292273236e-08,
      "loss": 0.1982,
      "step": 23500
    },
    {
      "epoch": 1.9350157220027413,
      "grad_norm": 42.68307113647461,
      "learning_rate": 3.61478163493841e-08,
      "loss": 0.1958,
      "step": 24000
    },
    {
      "epoch": 1.975328549544465,
      "grad_norm": 17.30171012878418,
      "learning_rate": 1.3751399776035833e-08,
      "loss": 0.2024,
      "step": 24500
    },
    {
      "epoch": 2.0,
      "eval_accuracy": 0.9520795284741823,
      "eval_loss": 0.13123206794261932,
      "eval_model_preparation_time": 0.0026,
      "eval_runtime": 100.8699,
      "eval_samples_per_second": 477.734,
      "eval_steps_per_second": 59.72,
      "step": 24806
    }
  ],
  "logging_steps": 500,
  "max_steps": 24806,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 2,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 6.6484442538990895e+19,
  "train_batch_size": 32,
  "trial_name": null,
  "trial_params": null
}