File size: 7,250 Bytes
161775f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 3.0,
  "eval_steps": 500,
  "global_step": 387,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.07751937984496124,
      "grad_norm": 2.245417356491089,
      "learning_rate": 5e-05,
      "loss": 2.1087,
      "step": 10
    },
    {
      "epoch": 0.15503875968992248,
      "grad_norm": 3.3328239917755127,
      "learning_rate": 0.0001,
      "loss": 1.7547,
      "step": 20
    },
    {
      "epoch": 0.23255813953488372,
      "grad_norm": 1.6743243932724,
      "learning_rate": 9.993684783030088e-05,
      "loss": 1.3868,
      "step": 30
    },
    {
      "epoch": 0.31007751937984496,
      "grad_norm": 1.585747480392456,
      "learning_rate": 9.974755084906502e-05,
      "loss": 1.1455,
      "step": 40
    },
    {
      "epoch": 0.3875968992248062,
      "grad_norm": 1.2781128883361816,
      "learning_rate": 9.94325872368957e-05,
      "loss": 1.1394,
      "step": 50
    },
    {
      "epoch": 0.46511627906976744,
      "grad_norm": 1.3732120990753174,
      "learning_rate": 9.899275261921234e-05,
      "loss": 1.0942,
      "step": 60
    },
    {
      "epoch": 0.5426356589147286,
      "grad_norm": 1.4817951917648315,
      "learning_rate": 9.842915805643155e-05,
      "loss": 0.9934,
      "step": 70
    },
    {
      "epoch": 0.6201550387596899,
      "grad_norm": 1.261801838874817,
      "learning_rate": 9.774322723733216e-05,
      "loss": 1.0093,
      "step": 80
    },
    {
      "epoch": 0.6976744186046512,
      "grad_norm": 1.801243782043457,
      "learning_rate": 9.693669288269372e-05,
      "loss": 1.0686,
      "step": 90
    },
    {
      "epoch": 0.7751937984496124,
      "grad_norm": 1.719342827796936,
      "learning_rate": 9.601159236829352e-05,
      "loss": 0.9973,
      "step": 100
    },
    {
      "epoch": 0.8527131782945736,
      "grad_norm": 1.5797449350357056,
      "learning_rate": 9.497026257831855e-05,
      "loss": 1.0002,
      "step": 110
    },
    {
      "epoch": 0.9302325581395349,
      "grad_norm": 1.790971279144287,
      "learning_rate": 9.381533400219318e-05,
      "loss": 0.9484,
      "step": 120
    },
    {
      "epoch": 1.0077519379844961,
      "grad_norm": 1.6127315759658813,
      "learning_rate": 9.254972408973461e-05,
      "loss": 0.979,
      "step": 130
    },
    {
      "epoch": 1.0852713178294573,
      "grad_norm": 1.4869861602783203,
      "learning_rate": 9.117662988142138e-05,
      "loss": 0.8899,
      "step": 140
    },
    {
      "epoch": 1.1627906976744187,
      "grad_norm": 2.101341962814331,
      "learning_rate": 8.969951993239177e-05,
      "loss": 0.9493,
      "step": 150
    },
    {
      "epoch": 1.2403100775193798,
      "grad_norm": 1.7181503772735596,
      "learning_rate": 8.81221255505724e-05,
      "loss": 0.8715,
      "step": 160
    },
    {
      "epoch": 1.3178294573643412,
      "grad_norm": 2.006377696990967,
      "learning_rate": 8.644843137107059e-05,
      "loss": 0.8846,
      "step": 170
    },
    {
      "epoch": 1.3953488372093024,
      "grad_norm": 2.0157504081726074,
      "learning_rate": 8.468266529064025e-05,
      "loss": 0.9292,
      "step": 180
    },
    {
      "epoch": 1.4728682170542635,
      "grad_norm": 1.704633355140686,
      "learning_rate": 8.282928778764783e-05,
      "loss": 0.8883,
      "step": 190
    },
    {
      "epoch": 1.550387596899225,
      "grad_norm": 1.7320111989974976,
      "learning_rate": 8.089298065451672e-05,
      "loss": 0.874,
      "step": 200
    },
    {
      "epoch": 1.627906976744186,
      "grad_norm": 2.369300127029419,
      "learning_rate": 7.887863517111338e-05,
      "loss": 0.8688,
      "step": 210
    },
    {
      "epoch": 1.7054263565891472,
      "grad_norm": 1.5535732507705688,
      "learning_rate": 7.679133974894983e-05,
      "loss": 0.8726,
      "step": 220
    },
    {
      "epoch": 1.7829457364341086,
      "grad_norm": 1.9749231338500977,
      "learning_rate": 7.463636707741458e-05,
      "loss": 0.8252,
      "step": 230
    },
    {
      "epoch": 1.8604651162790697,
      "grad_norm": 2.261054277420044,
      "learning_rate": 7.241916080450163e-05,
      "loss": 0.846,
      "step": 240
    },
    {
      "epoch": 1.937984496124031,
      "grad_norm": 2.019524097442627,
      "learning_rate": 7.014532178568314e-05,
      "loss": 0.8104,
      "step": 250
    },
    {
      "epoch": 2.0155038759689923,
      "grad_norm": 1.8253782987594604,
      "learning_rate": 6.782059393566253e-05,
      "loss": 0.8436,
      "step": 260
    },
    {
      "epoch": 2.0930232558139537,
      "grad_norm": 1.92261803150177,
      "learning_rate": 6.545084971874738e-05,
      "loss": 0.8007,
      "step": 270
    },
    {
      "epoch": 2.1705426356589146,
      "grad_norm": 1.959991455078125,
      "learning_rate": 6.304207531449486e-05,
      "loss": 0.8093,
      "step": 280
    },
    {
      "epoch": 2.248062015503876,
      "grad_norm": 2.0819852352142334,
      "learning_rate": 6.0600355496102745e-05,
      "loss": 0.7626,
      "step": 290
    },
    {
      "epoch": 2.3255813953488373,
      "grad_norm": 2.1029396057128906,
      "learning_rate": 5.813185825974419e-05,
      "loss": 0.7995,
      "step": 300
    },
    {
      "epoch": 2.4031007751937983,
      "grad_norm": 2.149883508682251,
      "learning_rate": 5.564281924367408e-05,
      "loss": 0.7867,
      "step": 310
    },
    {
      "epoch": 2.4806201550387597,
      "grad_norm": 2.2461025714874268,
      "learning_rate": 5.313952597646568e-05,
      "loss": 0.778,
      "step": 320
    },
    {
      "epoch": 2.558139534883721,
      "grad_norm": 2.1659789085388184,
      "learning_rate": 5.062830199416764e-05,
      "loss": 0.8143,
      "step": 330
    },
    {
      "epoch": 2.6356589147286824,
      "grad_norm": 2.3656270503997803,
      "learning_rate": 4.811549086650327e-05,
      "loss": 0.8017,
      "step": 340
    },
    {
      "epoch": 2.7131782945736433,
      "grad_norm": 2.3753740787506104,
      "learning_rate": 4.560744017246284e-05,
      "loss": 0.7666,
      "step": 350
    },
    {
      "epoch": 2.7906976744186047,
      "grad_norm": 2.221393585205078,
      "learning_rate": 4.31104854657681e-05,
      "loss": 0.7679,
      "step": 360
    },
    {
      "epoch": 2.8682170542635657,
      "grad_norm": 2.018345594406128,
      "learning_rate": 4.063093427071376e-05,
      "loss": 0.7507,
      "step": 370
    },
    {
      "epoch": 2.945736434108527,
      "grad_norm": 2.199131965637207,
      "learning_rate": 3.817505014881378e-05,
      "loss": 0.7413,
      "step": 380
    }
  ],
  "logging_steps": 10,
  "max_steps": 645,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 5,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 4.029055780857446e+16,
  "train_batch_size": 4,
  "trial_name": null,
  "trial_params": null
}