File size: 15,279 Bytes
8d4d182
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.995910949568378,
  "eval_steps": 400,
  "global_step": 137,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.03634711494775102,
      "grad_norm": 42.18056900627306,
      "learning_rate": 2.857142857142857e-07,
      "logits/chosen": -10.936590194702148,
      "logits/rejected": -11.232436180114746,
      "logps/chosen": -0.6059076189994812,
      "logps/rejected": -0.5706321001052856,
      "loss": 5.4752,
      "rewards/accuracies": 0.3125,
      "rewards/chosen": -6.05907678604126,
      "rewards/margins": -0.352755606174469,
      "rewards/rejected": -5.7063212394714355,
      "step": 5
    },
    {
      "epoch": 0.07269422989550205,
      "grad_norm": 33.128716826319426,
      "learning_rate": 5.714285714285714e-07,
      "logits/chosen": -10.911026954650879,
      "logits/rejected": -11.181764602661133,
      "logps/chosen": -0.6222352385520935,
      "logps/rejected": -0.5984331965446472,
      "loss": 5.5068,
      "rewards/accuracies": 0.3187499940395355,
      "rewards/chosen": -6.222352504730225,
      "rewards/margins": -0.23802027106285095,
      "rewards/rejected": -5.984332084655762,
      "step": 10
    },
    {
      "epoch": 0.10904134484325306,
      "grad_norm": 50.848111590558474,
      "learning_rate": 7.998695344323425e-07,
      "logits/chosen": -10.209379196166992,
      "logits/rejected": -10.468820571899414,
      "logps/chosen": -0.6080077290534973,
      "logps/rejected": -0.5699423551559448,
      "loss": 5.3874,
      "rewards/accuracies": 0.2874999940395355,
      "rewards/chosen": -6.080076694488525,
      "rewards/margins": -0.3806529641151428,
      "rewards/rejected": -5.699423789978027,
      "step": 15
    },
    {
      "epoch": 0.1453884597910041,
      "grad_norm": 35.145738479410355,
      "learning_rate": 7.953121695121394e-07,
      "logits/chosen": -9.792947769165039,
      "logits/rejected": -10.061820030212402,
      "logps/chosen": -0.5358205437660217,
      "logps/rejected": -0.5112504363059998,
      "loss": 5.3654,
      "rewards/accuracies": 0.3499999940395355,
      "rewards/chosen": -5.358205318450928,
      "rewards/margins": -0.24570095539093018,
      "rewards/rejected": -5.112504482269287,
      "step": 20
    },
    {
      "epoch": 0.18173557473875512,
      "grad_norm": 35.43905001965967,
      "learning_rate": 7.843163833184991e-07,
      "logits/chosen": -9.215612411499023,
      "logits/rejected": -9.5372314453125,
      "logps/chosen": -0.47783246636390686,
      "logps/rejected": -0.4458232522010803,
      "loss": 5.2677,
      "rewards/accuracies": 0.33125001192092896,
      "rewards/chosen": -4.778324604034424,
      "rewards/margins": -0.32009226083755493,
      "rewards/rejected": -4.4582319259643555,
      "step": 25
    },
    {
      "epoch": 0.21808268968650613,
      "grad_norm": 34.872469426364724,
      "learning_rate": 7.670612634414511e-07,
      "logits/chosen": -8.819177627563477,
      "logits/rejected": -8.965360641479492,
      "logps/chosen": -0.49288082122802734,
      "logps/rejected": -0.4752427041530609,
      "loss": 5.1863,
      "rewards/accuracies": 0.34375,
      "rewards/chosen": -4.928807735443115,
      "rewards/margins": -0.17638036608695984,
      "rewards/rejected": -4.752427577972412,
      "step": 30
    },
    {
      "epoch": 0.25442980463425713,
      "grad_norm": 41.683131083762184,
      "learning_rate": 7.438278427948805e-07,
      "logits/chosen": -8.983338356018066,
      "logits/rejected": -9.358332633972168,
      "logps/chosen": -0.45715832710266113,
      "logps/rejected": -0.4428199231624603,
      "loss": 5.1479,
      "rewards/accuracies": 0.3187499940395355,
      "rewards/chosen": -4.571582794189453,
      "rewards/margins": -0.14338405430316925,
      "rewards/rejected": -4.428199291229248,
      "step": 35
    },
    {
      "epoch": 0.2907769195820082,
      "grad_norm": 35.49350724475253,
      "learning_rate": 7.149945224533862e-07,
      "logits/chosen": -8.473846435546875,
      "logits/rejected": -8.634288787841797,
      "logps/chosen": -0.49668779969215393,
      "logps/rejected": -0.4822346568107605,
      "loss": 5.1059,
      "rewards/accuracies": 0.4124999940395355,
      "rewards/chosen": -4.9668779373168945,
      "rewards/margins": -0.14453084766864777,
      "rewards/rejected": -4.8223466873168945,
      "step": 40
    },
    {
      "epoch": 0.3271240345297592,
      "grad_norm": 35.201815027229166,
      "learning_rate": 6.810309086608129e-07,
      "logits/chosen": -9.485041618347168,
      "logits/rejected": -9.641453742980957,
      "logps/chosen": -0.4795761704444885,
      "logps/rejected": -0.4812702536582947,
      "loss": 5.0687,
      "rewards/accuracies": 0.3812499940395355,
      "rewards/chosen": -4.795762062072754,
      "rewards/margins": 0.016940664499998093,
      "rewards/rejected": -4.812702655792236,
      "step": 45
    },
    {
      "epoch": 0.36347114947751025,
      "grad_norm": 41.62812201333068,
      "learning_rate": 6.424901643866552e-07,
      "logits/chosen": -9.312422752380371,
      "logits/rejected": -9.527371406555176,
      "logps/chosen": -0.5089246034622192,
      "logps/rejected": -0.5194448232650757,
      "loss": 5.0107,
      "rewards/accuracies": 0.4124999940395355,
      "rewards/chosen": -5.0892462730407715,
      "rewards/margins": 0.10520193725824356,
      "rewards/rejected": -5.194448471069336,
      "step": 50
    },
    {
      "epoch": 0.39981826442526125,
      "grad_norm": 34.69469413711179,
      "learning_rate": 6e-07,
      "logits/chosen": -9.82011604309082,
      "logits/rejected": -9.918914794921875,
      "logps/chosen": -0.5091612935066223,
      "logps/rejected": -0.5071789026260376,
      "loss": 4.9611,
      "rewards/accuracies": 0.4749999940395355,
      "rewards/chosen": -5.091612815856934,
      "rewards/margins": -0.019824281334877014,
      "rewards/rejected": -5.071788311004639,
      "step": 55
    },
    {
      "epoch": 0.43616537937301225,
      "grad_norm": 39.20092819974629,
      "learning_rate": 5.542524497952543e-07,
      "logits/chosen": -10.392576217651367,
      "logits/rejected": -10.515856742858887,
      "logps/chosen": -0.49750223755836487,
      "logps/rejected": -0.5498560070991516,
      "loss": 4.7725,
      "rewards/accuracies": 0.4625000059604645,
      "rewards/chosen": -4.975022315979004,
      "rewards/margins": 0.5235382318496704,
      "rewards/rejected": -5.498560428619385,
      "step": 60
    },
    {
      "epoch": 0.4725124943207633,
      "grad_norm": 41.76932417294435,
      "learning_rate": 5.059926008786647e-07,
      "logits/chosen": -9.944269180297852,
      "logits/rejected": -10.23642635345459,
      "logps/chosen": -0.5527420043945312,
      "logps/rejected": -0.6039702296257019,
      "loss": 4.9196,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -5.5274200439453125,
      "rewards/margins": 0.5122829675674438,
      "rewards/rejected": -6.039702415466309,
      "step": 65
    },
    {
      "epoch": 0.5088596092685143,
      "grad_norm": 42.78642756929462,
      "learning_rate": 4.5600645798745166e-07,
      "logits/chosen": -10.92486572265625,
      "logits/rejected": -11.130925178527832,
      "logps/chosen": -0.5507891774177551,
      "logps/rejected": -0.5964936017990112,
      "loss": 4.7685,
      "rewards/accuracies": 0.5062500238418579,
      "rewards/chosen": -5.50789213180542,
      "rewards/margins": 0.4570443034172058,
      "rewards/rejected": -5.964935779571533,
      "step": 70
    },
    {
      "epoch": 0.5452067242162654,
      "grad_norm": 43.736655032116566,
      "learning_rate": 4.051081418863895e-07,
      "logits/chosen": -10.675444602966309,
      "logits/rejected": -10.829668045043945,
      "logps/chosen": -0.6147274374961853,
      "logps/rejected": -0.6446117758750916,
      "loss": 4.8607,
      "rewards/accuracies": 0.581250011920929,
      "rewards/chosen": -6.147274494171143,
      "rewards/margins": 0.29884377121925354,
      "rewards/rejected": -6.446118354797363,
      "step": 75
    },
    {
      "epoch": 0.5815538391640164,
      "grad_norm": 50.47780593376421,
      "learning_rate": 3.541266298406398e-07,
      "logits/chosen": -11.719624519348145,
      "logits/rejected": -11.85960578918457,
      "logps/chosen": -0.6017196774482727,
      "logps/rejected": -0.6400548815727234,
      "loss": 4.7166,
      "rewards/accuracies": 0.5,
      "rewards/chosen": -6.017197608947754,
      "rewards/margins": 0.38335245847702026,
      "rewards/rejected": -6.400549411773682,
      "step": 80
    },
    {
      "epoch": 0.6179009541117674,
      "grad_norm": 49.58320001105474,
      "learning_rate": 3.0389225412181565e-07,
      "logits/chosen": -11.57774543762207,
      "logits/rejected": -11.787096977233887,
      "logps/chosen": -0.639508843421936,
      "logps/rejected": -0.7087122797966003,
      "loss": 4.6635,
      "rewards/accuracies": 0.59375,
      "rewards/chosen": -6.395088195800781,
      "rewards/margins": 0.6920341849327087,
      "rewards/rejected": -7.087122917175293,
      "step": 85
    },
    {
      "epoch": 0.6542480690595184,
      "grad_norm": 60.72521586951035,
      "learning_rate": 2.5522317844515273e-07,
      "logits/chosen": -11.539664268493652,
      "logits/rejected": -11.73334789276123,
      "logps/chosen": -0.6763519644737244,
      "logps/rejected": -0.719671905040741,
      "loss": 4.6877,
      "rewards/accuracies": 0.543749988079071,
      "rewards/chosen": -6.763520240783691,
      "rewards/margins": 0.4331996440887451,
      "rewards/rejected": -7.196720123291016,
      "step": 90
    },
    {
      "epoch": 0.6905951840072694,
      "grad_norm": 73.18876407608641,
      "learning_rate": 2.0891207259509476e-07,
      "logits/chosen": -11.891935348510742,
      "logits/rejected": -11.978799819946289,
      "logps/chosen": -0.6321767568588257,
      "logps/rejected": -0.7210510969161987,
      "loss": 4.6661,
      "rewards/accuracies": 0.5625,
      "rewards/chosen": -6.321768283843994,
      "rewards/margins": 0.8887429237365723,
      "rewards/rejected": -7.210511207580566,
      "step": 95
    },
    {
      "epoch": 0.7269422989550205,
      "grad_norm": 58.29758563166122,
      "learning_rate": 1.6571320226872206e-07,
      "logits/chosen": -12.378867149353027,
      "logits/rejected": -12.450654983520508,
      "logps/chosen": -0.7091072797775269,
      "logps/rejected": -0.7949134111404419,
      "loss": 4.5318,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": -7.091072082519531,
      "rewards/margins": 0.8580614924430847,
      "rewards/rejected": -7.949133396148682,
      "step": 100
    },
    {
      "epoch": 0.7632894139027715,
      "grad_norm": 65.14186021974389,
      "learning_rate": 1.2633014440382787e-07,
      "logits/chosen": -12.120596885681152,
      "logits/rejected": -12.198526382446289,
      "logps/chosen": -0.7462261319160461,
      "logps/rejected": -0.8084705471992493,
      "loss": 4.5075,
      "rewards/accuracies": 0.5375000238418579,
      "rewards/chosen": -7.462261199951172,
      "rewards/margins": 0.622443437576294,
      "rewards/rejected": -8.084705352783203,
      "step": 105
    },
    {
      "epoch": 0.7996365288505225,
      "grad_norm": 66.21038830050585,
      "learning_rate": 9.14043280712228e-08,
      "logits/chosen": -12.642437934875488,
      "logits/rejected": -12.874890327453613,
      "logps/chosen": -0.7560291290283203,
      "logps/rejected": -0.8245932459831238,
      "loss": 4.6235,
      "rewards/accuracies": 0.5874999761581421,
      "rewards/chosen": -7.560290336608887,
      "rewards/margins": 0.6856421828269958,
      "rewards/rejected": -8.245932579040527,
      "step": 110
    },
    {
      "epoch": 0.8359836437982735,
      "grad_norm": 58.78661492817544,
      "learning_rate": 6.150458756494239e-08,
      "logits/chosen": -12.5349760055542,
      "logits/rejected": -12.633232116699219,
      "logps/chosen": -0.7407893538475037,
      "logps/rejected": -0.8003439903259277,
      "loss": 4.537,
      "rewards/accuracies": 0.5375000238418579,
      "rewards/chosen": -7.407894134521484,
      "rewards/margins": 0.5955458879470825,
      "rewards/rejected": -8.003438949584961,
      "step": 115
    },
    {
      "epoch": 0.8723307587460245,
      "grad_norm": 72.85901824273444,
      "learning_rate": 3.711789783843522e-08,
      "logits/chosen": -12.179682731628418,
      "logits/rejected": -12.421493530273438,
      "logps/chosen": -0.7077151536941528,
      "logps/rejected": -0.8436015248298645,
      "loss": 4.41,
      "rewards/accuracies": 0.6187499761581421,
      "rewards/chosen": -7.077151298522949,
      "rewards/margins": 1.358863115310669,
      "rewards/rejected": -8.436014175415039,
      "step": 120
    },
    {
      "epoch": 0.9086778736937755,
      "grad_norm": 81.44112780212329,
      "learning_rate": 1.8641443178027784e-08,
      "logits/chosen": -12.844386100769043,
      "logits/rejected": -13.204521179199219,
      "logps/chosen": -0.7584520578384399,
      "logps/rejected": -0.8550151586532593,
      "loss": 4.4737,
      "rewards/accuracies": 0.6312500238418579,
      "rewards/chosen": -7.584519863128662,
      "rewards/margins": 0.965630829334259,
      "rewards/rejected": -8.550151824951172,
      "step": 125
    },
    {
      "epoch": 0.9450249886415266,
      "grad_norm": 63.15208705033616,
      "learning_rate": 6.376148290617145e-09,
      "logits/chosen": -13.006174087524414,
      "logits/rejected": -13.145883560180664,
      "logps/chosen": -0.7715423703193665,
      "logps/rejected": -0.8375405073165894,
      "loss": 4.6746,
      "rewards/accuracies": 0.6187499761581421,
      "rewards/chosen": -7.715423583984375,
      "rewards/margins": 0.6599816679954529,
      "rewards/rejected": -8.375406265258789,
      "step": 130
    },
    {
      "epoch": 0.9813721035892776,
      "grad_norm": 69.30502594164692,
      "learning_rate": 5.217771643080127e-10,
      "logits/chosen": -12.759674072265625,
      "logits/rejected": -12.864163398742676,
      "logps/chosen": -0.7980765700340271,
      "logps/rejected": -0.8657040596008301,
      "loss": 4.3661,
      "rewards/accuracies": 0.612500011920929,
      "rewards/chosen": -7.980767250061035,
      "rewards/margins": 0.6762741804122925,
      "rewards/rejected": -8.6570405960083,
      "step": 135
    },
    {
      "epoch": 0.995910949568378,
      "step": 137,
      "total_flos": 0.0,
      "train_loss": 4.865393144370866,
      "train_runtime": 3024.5765,
      "train_samples_per_second": 5.821,
      "train_steps_per_second": 0.045
    }
  ],
  "logging_steps": 5,
  "max_steps": 137,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 1000000,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}