File size: 49,828 Bytes
d38fc29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.9989094874591058,
  "eval_steps": 400,
  "global_step": 458,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.010905125408942203,
      "grad_norm": 64.1503528533801,
      "learning_rate": 8.695652173913042e-08,
      "logits/chosen": -10.94892692565918,
      "logits/rejected": -10.934142112731934,
      "logps/chosen": -0.8995465040206909,
      "logps/rejected": -0.8591210246086121,
      "loss": 5.7639,
      "rewards/accuracies": 0.48750001192092896,
      "rewards/chosen": -8.995465278625488,
      "rewards/margins": -0.40425485372543335,
      "rewards/rejected": -8.59121036529541,
      "step": 5
    },
    {
      "epoch": 0.021810250817884406,
      "grad_norm": 111.15906224445196,
      "learning_rate": 1.7391304347826085e-07,
      "logits/chosen": -9.873212814331055,
      "logits/rejected": -9.899297714233398,
      "logps/chosen": -1.1384799480438232,
      "logps/rejected": -1.158919095993042,
      "loss": 6.1082,
      "rewards/accuracies": 0.518750011920929,
      "rewards/chosen": -11.384800910949707,
      "rewards/margins": 0.20439176261425018,
      "rewards/rejected": -11.589191436767578,
      "step": 10
    },
    {
      "epoch": 0.03271537622682661,
      "grad_norm": 96.8211465004871,
      "learning_rate": 2.608695652173913e-07,
      "logits/chosen": -10.422977447509766,
      "logits/rejected": -10.321990013122559,
      "logps/chosen": -0.8700820803642273,
      "logps/rejected": -0.840315043926239,
      "loss": 5.5576,
      "rewards/accuracies": 0.48124998807907104,
      "rewards/chosen": -8.700819969177246,
      "rewards/margins": -0.29766958951950073,
      "rewards/rejected": -8.40315055847168,
      "step": 15
    },
    {
      "epoch": 0.04362050163576881,
      "grad_norm": 59.01122752402013,
      "learning_rate": 3.478260869565217e-07,
      "logits/chosen": -9.897313117980957,
      "logits/rejected": -9.771203994750977,
      "logps/chosen": -0.9016021490097046,
      "logps/rejected": -0.9813264608383179,
      "loss": 5.4183,
      "rewards/accuracies": 0.574999988079071,
      "rewards/chosen": -9.016023635864258,
      "rewards/margins": 0.797242283821106,
      "rewards/rejected": -9.813264846801758,
      "step": 20
    },
    {
      "epoch": 0.05452562704471101,
      "grad_norm": 77.6935326260568,
      "learning_rate": 4.3478260869565214e-07,
      "logits/chosen": -10.004715919494629,
      "logits/rejected": -10.082869529724121,
      "logps/chosen": -0.7554069757461548,
      "logps/rejected": -0.7040703296661377,
      "loss": 5.4042,
      "rewards/accuracies": 0.5375000238418579,
      "rewards/chosen": -7.554069519042969,
      "rewards/margins": -0.5133668780326843,
      "rewards/rejected": -7.040703773498535,
      "step": 25
    },
    {
      "epoch": 0.06543075245365322,
      "grad_norm": 53.25393883515416,
      "learning_rate": 5.217391304347826e-07,
      "logits/chosen": -9.147699356079102,
      "logits/rejected": -9.047504425048828,
      "logps/chosen": -0.6732769012451172,
      "logps/rejected": -0.7553219199180603,
      "loss": 5.079,
      "rewards/accuracies": 0.5562499761581421,
      "rewards/chosen": -6.732769012451172,
      "rewards/margins": 0.8204498291015625,
      "rewards/rejected": -7.553219795227051,
      "step": 30
    },
    {
      "epoch": 0.07633587786259542,
      "grad_norm": 49.376076232037846,
      "learning_rate": 6.08695652173913e-07,
      "logits/chosen": -9.227429389953613,
      "logits/rejected": -9.0322847366333,
      "logps/chosen": -0.6311928033828735,
      "logps/rejected": -0.6472190022468567,
      "loss": 4.9879,
      "rewards/accuracies": 0.543749988079071,
      "rewards/chosen": -6.311927795410156,
      "rewards/margins": 0.160261332988739,
      "rewards/rejected": -6.472189426422119,
      "step": 35
    },
    {
      "epoch": 0.08724100327153762,
      "grad_norm": 71.82957946446516,
      "learning_rate": 6.956521739130434e-07,
      "logits/chosen": -8.839519500732422,
      "logits/rejected": -8.598153114318848,
      "logps/chosen": -0.5917301177978516,
      "logps/rejected": -0.6095324158668518,
      "loss": 4.9341,
      "rewards/accuracies": 0.543749988079071,
      "rewards/chosen": -5.917301177978516,
      "rewards/margins": 0.1780225783586502,
      "rewards/rejected": -6.09532356262207,
      "step": 40
    },
    {
      "epoch": 0.09814612868047982,
      "grad_norm": 62.13252777111053,
      "learning_rate": 7.826086956521739e-07,
      "logits/chosen": -8.288870811462402,
      "logits/rejected": -8.250112533569336,
      "logps/chosen": -0.5544935464859009,
      "logps/rejected": -0.5844836235046387,
      "loss": 4.6875,
      "rewards/accuracies": 0.5375000238418579,
      "rewards/chosen": -5.5449347496032715,
      "rewards/margins": 0.29990124702453613,
      "rewards/rejected": -5.844836235046387,
      "step": 45
    },
    {
      "epoch": 0.10905125408942203,
      "grad_norm": 50.37809623047656,
      "learning_rate": 7.998139534493406e-07,
      "logits/chosen": -8.140437126159668,
      "logits/rejected": -7.910015106201172,
      "logps/chosen": -0.638617753982544,
      "logps/rejected": -0.7139040231704712,
      "loss": 4.5683,
      "rewards/accuracies": 0.574999988079071,
      "rewards/chosen": -6.386178493499756,
      "rewards/margins": 0.7528623938560486,
      "rewards/rejected": -7.139040470123291,
      "step": 50
    },
    {
      "epoch": 0.11995637949836423,
      "grad_norm": 61.0124851348064,
      "learning_rate": 7.990584359406726e-07,
      "logits/chosen": -9.38193416595459,
      "logits/rejected": -9.200593948364258,
      "logps/chosen": -0.6021402478218079,
      "logps/rejected": -0.7378085255622864,
      "loss": 4.6702,
      "rewards/accuracies": 0.5562499761581421,
      "rewards/chosen": -6.021402359008789,
      "rewards/margins": 1.3566832542419434,
      "rewards/rejected": -7.378085136413574,
      "step": 55
    },
    {
      "epoch": 0.13086150490730644,
      "grad_norm": 58.34041763150138,
      "learning_rate": 7.977229168076815e-07,
      "logits/chosen": -8.994357109069824,
      "logits/rejected": -9.095643043518066,
      "logps/chosen": -0.6254938244819641,
      "logps/rejected": -0.7731696963310242,
      "loss": 4.4672,
      "rewards/accuracies": 0.606249988079071,
      "rewards/chosen": -6.254937648773193,
      "rewards/margins": 1.4767591953277588,
      "rewards/rejected": -7.731697082519531,
      "step": 60
    },
    {
      "epoch": 0.14176663031624864,
      "grad_norm": 54.06832565592651,
      "learning_rate": 7.95809337127655e-07,
      "logits/chosen": -9.56706428527832,
      "logits/rejected": -9.66978931427002,
      "logps/chosen": -0.6539489030838013,
      "logps/rejected": -0.7899436950683594,
      "loss": 4.4539,
      "rewards/accuracies": 0.625,
      "rewards/chosen": -6.539488315582275,
      "rewards/margins": 1.359948992729187,
      "rewards/rejected": -7.899436950683594,
      "step": 65
    },
    {
      "epoch": 0.15267175572519084,
      "grad_norm": 61.12821185450308,
      "learning_rate": 7.933204781457008e-07,
      "logits/chosen": -9.499263763427734,
      "logits/rejected": -9.496051788330078,
      "logps/chosen": -0.6623054146766663,
      "logps/rejected": -0.7747452259063721,
      "loss": 4.4992,
      "rewards/accuracies": 0.65625,
      "rewards/chosen": -6.623053550720215,
      "rewards/margins": 1.1243983507156372,
      "rewards/rejected": -7.747453212738037,
      "step": 70
    },
    {
      "epoch": 0.16357688113413305,
      "grad_norm": 51.861021015777716,
      "learning_rate": 7.902599572324151e-07,
      "logits/chosen": -9.391631126403809,
      "logits/rejected": -9.375069618225098,
      "logps/chosen": -0.6535338163375854,
      "logps/rejected": -0.8060415983200073,
      "loss": 4.4979,
      "rewards/accuracies": 0.6000000238418579,
      "rewards/chosen": -6.535338401794434,
      "rewards/margins": 1.5250780582427979,
      "rewards/rejected": -8.060416221618652,
      "step": 75
    },
    {
      "epoch": 0.17448200654307525,
      "grad_norm": 70.13525940500534,
      "learning_rate": 7.86632222626304e-07,
      "logits/chosen": -9.956920623779297,
      "logits/rejected": -9.78042984008789,
      "logps/chosen": -0.6164920926094055,
      "logps/rejected": -0.7590925097465515,
      "loss": 4.429,
      "rewards/accuracies": 0.65625,
      "rewards/chosen": -6.164921760559082,
      "rewards/margins": 1.4260046482086182,
      "rewards/rejected": -7.590925693511963,
      "step": 80
    },
    {
      "epoch": 0.18538713195201745,
      "grad_norm": 73.7268648843484,
      "learning_rate": 7.824425469686014e-07,
      "logits/chosen": -9.828500747680664,
      "logits/rejected": -9.678921699523926,
      "logps/chosen": -0.6361075043678284,
      "logps/rejected": -0.7591807246208191,
      "loss": 4.438,
      "rewards/accuracies": 0.65625,
      "rewards/chosen": -6.361075401306152,
      "rewards/margins": 1.2307318449020386,
      "rewards/rejected": -7.591805934906006,
      "step": 85
    },
    {
      "epoch": 0.19629225736095965,
      "grad_norm": 69.62622029011087,
      "learning_rate": 7.776970196398794e-07,
      "logits/chosen": -9.843175888061523,
      "logits/rejected": -9.792081832885742,
      "logps/chosen": -0.6243542432785034,
      "logps/rejected": -0.7286983132362366,
      "loss": 4.3737,
      "rewards/accuracies": 0.668749988079071,
      "rewards/chosen": -6.2435431480407715,
      "rewards/margins": 1.0434401035308838,
      "rewards/rejected": -7.286982536315918,
      "step": 90
    },
    {
      "epoch": 0.20719738276990185,
      "grad_norm": 66.25031731585119,
      "learning_rate": 7.724025379095878e-07,
      "logits/chosen": -10.194116592407227,
      "logits/rejected": -10.006689071655273,
      "logps/chosen": -0.6633736491203308,
      "logps/rejected": -0.85277259349823,
      "loss": 4.2355,
      "rewards/accuracies": 0.637499988079071,
      "rewards/chosen": -6.633737087249756,
      "rewards/margins": 1.8939902782440186,
      "rewards/rejected": -8.527726173400879,
      "step": 95
    },
    {
      "epoch": 0.21810250817884405,
      "grad_norm": 75.08429330907302,
      "learning_rate": 7.665667969113885e-07,
      "logits/chosen": -9.845739364624023,
      "logits/rejected": -9.587841033935547,
      "logps/chosen": -0.6887632608413696,
      "logps/rejected": -0.872734546661377,
      "loss": 4.2625,
      "rewards/accuracies": 0.668749988079071,
      "rewards/chosen": -6.887633323669434,
      "rewards/margins": 1.839712381362915,
      "rewards/rejected": -8.72734546661377,
      "step": 100
    },
    {
      "epoch": 0.22900763358778625,
      "grad_norm": 69.02863249400193,
      "learning_rate": 7.601982784588541e-07,
      "logits/chosen": -10.62452220916748,
      "logits/rejected": -10.43299674987793,
      "logps/chosen": -0.7019127011299133,
      "logps/rejected": -0.7768973112106323,
      "loss": 4.2785,
      "rewards/accuracies": 0.6187499761581421,
      "rewards/chosen": -7.019126892089844,
      "rewards/margins": 0.7498467564582825,
      "rewards/rejected": -7.768974304199219,
      "step": 105
    },
    {
      "epoch": 0.23991275899672845,
      "grad_norm": 71.91219250814,
      "learning_rate": 7.533062387177842e-07,
      "logits/chosen": -10.65683650970459,
      "logits/rejected": -10.636886596679688,
      "logps/chosen": -0.7213658094406128,
      "logps/rejected": -0.8907195329666138,
      "loss": 4.1753,
      "rewards/accuracies": 0.6875,
      "rewards/chosen": -7.213658332824707,
      "rewards/margins": 1.6935373544692993,
      "rewards/rejected": -8.907196044921875,
      "step": 110
    },
    {
      "epoch": 0.25081788440567065,
      "grad_norm": 73.69796183664997,
      "learning_rate": 7.459006947530617e-07,
      "logits/chosen": -11.56556224822998,
      "logits/rejected": -11.49560260772705,
      "logps/chosen": -0.838672935962677,
      "logps/rejected": -1.0182416439056396,
      "loss": 4.0384,
      "rewards/accuracies": 0.668749988079071,
      "rewards/chosen": -8.38672924041748,
      "rewards/margins": 1.7956879138946533,
      "rewards/rejected": -10.182416915893555,
      "step": 115
    },
    {
      "epoch": 0.2617230098146129,
      "grad_norm": 94.92610924652415,
      "learning_rate": 7.379924099695959e-07,
      "logits/chosen": -11.925847053527832,
      "logits/rejected": -11.710810661315918,
      "logps/chosen": -0.8436716198921204,
      "logps/rejected": -1.0012091398239136,
      "loss": 4.113,
      "rewards/accuracies": 0.675000011920929,
      "rewards/chosen": -8.436716079711914,
      "rewards/margins": 1.5753746032714844,
      "rewards/rejected": -10.012091636657715,
      "step": 120
    },
    {
      "epoch": 0.27262813522355506,
      "grad_norm": 113.62807962706827,
      "learning_rate": 7.29592878468518e-07,
      "logits/chosen": -13.345464706420898,
      "logits/rejected": -12.9557523727417,
      "logps/chosen": -0.8897609710693359,
      "logps/rejected": -1.0490643978118896,
      "loss": 3.9421,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -8.897610664367676,
      "rewards/margins": 1.5930334329605103,
      "rewards/rejected": -10.490643501281738,
      "step": 125
    },
    {
      "epoch": 0.2835332606324973,
      "grad_norm": 118.6845458542213,
      "learning_rate": 7.207143083413642e-07,
      "logits/chosen": -12.795938491821289,
      "logits/rejected": -12.771161079406738,
      "logps/chosen": -0.9606706500053406,
      "logps/rejected": -1.1983877420425415,
      "loss": 3.7793,
      "rewards/accuracies": 0.706250011920929,
      "rewards/chosen": -9.606706619262695,
      "rewards/margins": 2.3771698474884033,
      "rewards/rejected": -11.983877182006836,
      "step": 130
    },
    {
      "epoch": 0.29443838604143946,
      "grad_norm": 112.00779108354124,
      "learning_rate": 7.11369603926526e-07,
      "logits/chosen": -13.919995307922363,
      "logits/rejected": -14.025039672851562,
      "logps/chosen": -1.1236939430236816,
      "logps/rejected": -1.3567845821380615,
      "loss": 3.8057,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -11.236939430236816,
      "rewards/margins": 2.3309078216552734,
      "rewards/rejected": -13.567845344543457,
      "step": 135
    },
    {
      "epoch": 0.3053435114503817,
      "grad_norm": 121.41466704799392,
      "learning_rate": 7.015723470537589e-07,
      "logits/chosen": -14.653108596801758,
      "logits/rejected": -14.910409927368164,
      "logps/chosen": -1.2697908878326416,
      "logps/rejected": -1.5090426206588745,
      "loss": 3.7783,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -12.69791030883789,
      "rewards/margins": 2.392517566680908,
      "rewards/rejected": -15.090426445007324,
      "step": 140
    },
    {
      "epoch": 0.31624863685932386,
      "grad_norm": 118.0007773777601,
      "learning_rate": 6.913367773040073e-07,
      "logits/chosen": -15.646504402160645,
      "logits/rejected": -15.359659194946289,
      "logps/chosen": -1.2195041179656982,
      "logps/rejected": -1.5088005065917969,
      "loss": 3.626,
      "rewards/accuracies": 0.71875,
      "rewards/chosen": -12.195042610168457,
      "rewards/margins": 2.892961263656616,
      "rewards/rejected": -15.088003158569336,
      "step": 145
    },
    {
      "epoch": 0.3271537622682661,
      "grad_norm": 123.35286255999112,
      "learning_rate": 6.806777713132373e-07,
      "logits/chosen": -15.674619674682617,
      "logits/rejected": -15.5663423538208,
      "logps/chosen": -1.3831182718276978,
      "logps/rejected": -1.60293710231781,
      "loss": 3.76,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": -13.831182479858398,
      "rewards/margins": 2.1981873512268066,
      "rewards/rejected": -16.029369354248047,
      "step": 150
    },
    {
      "epoch": 0.33805888767720826,
      "grad_norm": 131.11270501363964,
      "learning_rate": 6.69610821150358e-07,
      "logits/chosen": -15.65868854522705,
      "logits/rejected": -15.700967788696289,
      "logps/chosen": -1.3463469743728638,
      "logps/rejected": -1.6862995624542236,
      "loss": 3.4489,
      "rewards/accuracies": 0.699999988079071,
      "rewards/chosen": -13.463468551635742,
      "rewards/margins": 3.3995232582092285,
      "rewards/rejected": -16.862991333007812,
      "step": 155
    },
    {
      "epoch": 0.3489640130861505,
      "grad_norm": 135.1222468933701,
      "learning_rate": 6.581520118006564e-07,
      "logits/chosen": -15.956563949584961,
      "logits/rejected": -16.274681091308594,
      "logps/chosen": -1.4811228513717651,
      "logps/rejected": -1.8506397008895874,
      "loss": 3.4319,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -14.81122875213623,
      "rewards/margins": 3.695168972015381,
      "rewards/rejected": -18.506397247314453,
      "step": 160
    },
    {
      "epoch": 0.35986913849509267,
      "grad_norm": 117.31734546028652,
      "learning_rate": 6.46317997787473e-07,
      "logits/chosen": -16.915264129638672,
      "logits/rejected": -17.207149505615234,
      "logps/chosen": -1.4308239221572876,
      "logps/rejected": -1.7504551410675049,
      "loss": 3.7315,
      "rewards/accuracies": 0.699999988079071,
      "rewards/chosen": -14.30823802947998,
      "rewards/margins": 3.1963133811950684,
      "rewards/rejected": -17.504552841186523,
      "step": 165
    },
    {
      "epoch": 0.3707742639040349,
      "grad_norm": 135.9602274797816,
      "learning_rate": 6.341259789660969e-07,
      "logits/chosen": -16.72784996032715,
      "logits/rejected": -16.691822052001953,
      "logps/chosen": -1.5182801485061646,
      "logps/rejected": -1.922546148300171,
      "loss": 3.3263,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -15.182802200317383,
      "rewards/margins": 4.042659759521484,
      "rewards/rejected": -19.225460052490234,
      "step": 170
    },
    {
      "epoch": 0.3816793893129771,
      "grad_norm": 149.4264363196396,
      "learning_rate": 6.215936755250596e-07,
      "logits/chosen": -17.34011459350586,
      "logits/rejected": -17.421663284301758,
      "logps/chosen": -1.496767282485962,
      "logps/rejected": -1.85422682762146,
      "loss": 3.5029,
      "rewards/accuracies": 0.7562500238418579,
      "rewards/chosen": -14.967672348022461,
      "rewards/margins": 3.574597120285034,
      "rewards/rejected": -18.54227066040039,
      "step": 175
    },
    {
      "epoch": 0.3925845147219193,
      "grad_norm": 116.14395125879456,
      "learning_rate": 6.087393022311666e-07,
      "logits/chosen": -16.09074592590332,
      "logits/rejected": -15.661008834838867,
      "logps/chosen": -1.6729342937469482,
      "logps/rejected": -2.0850119590759277,
      "loss": 3.3119,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -16.729345321655273,
      "rewards/margins": 4.1207756996154785,
      "rewards/rejected": -20.85011863708496,
      "step": 180
    },
    {
      "epoch": 0.4034896401308615,
      "grad_norm": 302.66398320568743,
      "learning_rate": 5.955815419556934e-07,
      "logits/chosen": -17.55354118347168,
      "logits/rejected": -17.935550689697266,
      "logps/chosen": -1.8943233489990234,
      "logps/rejected": -2.240093231201172,
      "loss": 3.3613,
      "rewards/accuracies": 0.6812499761581421,
      "rewards/chosen": -18.943233489990234,
      "rewards/margins": 3.4576992988586426,
      "rewards/rejected": -22.40093231201172,
      "step": 185
    },
    {
      "epoch": 0.4143947655398037,
      "grad_norm": 125.3361496668939,
      "learning_rate": 5.821395185202285e-07,
      "logits/chosen": -15.341293334960938,
      "logits/rejected": -15.44407844543457,
      "logps/chosen": -1.8814849853515625,
      "logps/rejected": -2.2990918159484863,
      "loss": 3.1174,
      "rewards/accuracies": 0.7124999761581421,
      "rewards/chosen": -18.814849853515625,
      "rewards/margins": 4.17606782913208,
      "rewards/rejected": -22.990917205810547,
      "step": 190
    },
    {
      "epoch": 0.42529989094874593,
      "grad_norm": 123.17711415812771,
      "learning_rate": 5.684327689016264e-07,
      "logits/chosen": -15.80346393585205,
      "logits/rejected": -15.893457412719727,
      "logps/chosen": -1.6922333240509033,
      "logps/rejected": -2.1358566284179688,
      "loss": 3.3943,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -16.922334671020508,
      "rewards/margins": 4.436235427856445,
      "rewards/rejected": -21.358566284179688,
      "step": 195
    },
    {
      "epoch": 0.4362050163576881,
      "grad_norm": 121.59808879397914,
      "learning_rate": 5.544812148364731e-07,
      "logits/chosen": -16.505403518676758,
      "logits/rejected": -16.2810001373291,
      "logps/chosen": -1.7912696599960327,
      "logps/rejected": -2.283963680267334,
      "loss": 3.3937,
      "rewards/accuracies": 0.768750011920929,
      "rewards/chosen": -17.912700653076172,
      "rewards/margins": 4.926938533782959,
      "rewards/rejected": -22.839635848999023,
      "step": 200
    },
    {
      "epoch": 0.44711014176663033,
      "grad_norm": 130.03760301081726,
      "learning_rate": 5.40305133866328e-07,
      "logits/chosen": -16.51849365234375,
      "logits/rejected": -16.17244529724121,
      "logps/chosen": -1.8261501789093018,
      "logps/rejected": -2.216474771499634,
      "loss": 3.3439,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -18.26150131225586,
      "rewards/margins": 3.903247117996216,
      "rewards/rejected": -22.164745330810547,
      "step": 205
    },
    {
      "epoch": 0.4580152671755725,
      "grad_norm": 107.25457872042979,
      "learning_rate": 5.259251298658339e-07,
      "logits/chosen": -14.851339340209961,
      "logits/rejected": -14.814165115356445,
      "logps/chosen": -1.7905569076538086,
      "logps/rejected": -2.320906162261963,
      "loss": 3.1591,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -17.905569076538086,
      "rewards/margins": 5.303492546081543,
      "rewards/rejected": -23.209064483642578,
      "step": 210
    },
    {
      "epoch": 0.46892039258451473,
      "grad_norm": 129.6812695651334,
      "learning_rate": 5.113621030965238e-07,
      "logits/chosen": -16.215179443359375,
      "logits/rejected": -16.335643768310547,
      "logps/chosen": -1.8237041234970093,
      "logps/rejected": -2.207712411880493,
      "loss": 3.2475,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -18.23703956604004,
      "rewards/margins": 3.840085506439209,
      "rewards/rejected": -22.07712745666504,
      "step": 215
    },
    {
      "epoch": 0.4798255179934569,
      "grad_norm": 103.25783765301819,
      "learning_rate": 4.96637219829852e-07,
      "logits/chosen": -14.638501167297363,
      "logits/rejected": -14.695428848266602,
      "logps/chosen": -1.8283414840698242,
      "logps/rejected": -2.376028537750244,
      "loss": 3.3144,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -18.283416748046875,
      "rewards/margins": 5.47686767578125,
      "rewards/rejected": -23.760284423828125,
      "step": 220
    },
    {
      "epoch": 0.49073064340239914,
      "grad_norm": 132.036942227645,
      "learning_rate": 4.817718815835998e-07,
      "logits/chosen": -15.56849479675293,
      "logits/rejected": -15.743542671203613,
      "logps/chosen": -1.9234107732772827,
      "logps/rejected": -2.3087854385375977,
      "loss": 3.2906,
      "rewards/accuracies": 0.6812499761581421,
      "rewards/chosen": -19.234111785888672,
      "rewards/margins": 3.8537425994873047,
      "rewards/rejected": -23.087854385375977,
      "step": 225
    },
    {
      "epoch": 0.5016357688113413,
      "grad_norm": 132.28323726857968,
      "learning_rate": 4.6678769401636887e-07,
      "logits/chosen": -15.513163566589355,
      "logits/rejected": -15.330879211425781,
      "logps/chosen": -1.868847131729126,
      "logps/rejected": -2.3147101402282715,
      "loss": 3.2416,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -18.688472747802734,
      "rewards/margins": 4.4586310386657715,
      "rewards/rejected": -23.14710235595703,
      "step": 230
    },
    {
      "epoch": 0.5125408942202835,
      "grad_norm": 116.14658388209503,
      "learning_rate": 4.517064355253696e-07,
      "logits/chosen": -14.959882736206055,
      "logits/rejected": -14.972663879394531,
      "logps/chosen": -1.9306243658065796,
      "logps/rejected": -2.457444667816162,
      "loss": 2.9094,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -19.306243896484375,
      "rewards/margins": 5.268204689025879,
      "rewards/rejected": -24.574447631835938,
      "step": 235
    },
    {
      "epoch": 0.5234460196292258,
      "grad_norm": 125.188125846376,
      "learning_rate": 4.365500255931484e-07,
      "logits/chosen": -15.104351997375488,
      "logits/rejected": -15.062860488891602,
      "logps/chosen": -1.9797313213348389,
      "logps/rejected": -2.4472575187683105,
      "loss": 3.0593,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -19.797313690185547,
      "rewards/margins": 4.675261497497559,
      "rewards/rejected": -24.472576141357422,
      "step": 240
    },
    {
      "epoch": 0.5343511450381679,
      "grad_norm": 135.15384009467002,
      "learning_rate": 4.213404929292575e-07,
      "logits/chosen": -15.561657905578613,
      "logits/rejected": -15.077102661132812,
      "logps/chosen": -2.040621280670166,
      "logps/rejected": -2.5414981842041016,
      "loss": 2.9158,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -20.40620994567871,
      "rewards/margins": 5.008772373199463,
      "rewards/rejected": -25.41498374938965,
      "step": 245
    },
    {
      "epoch": 0.5452562704471101,
      "grad_norm": 124.60869295390538,
      "learning_rate": 4.060999434531704e-07,
      "logits/chosen": -15.397348403930664,
      "logits/rejected": -15.1492280960083,
      "logps/chosen": -1.8068050146102905,
      "logps/rejected": -2.3928465843200684,
      "loss": 2.8634,
      "rewards/accuracies": 0.793749988079071,
      "rewards/chosen": -18.068050384521484,
      "rewards/margins": 5.860415935516357,
      "rewards/rejected": -23.928464889526367,
      "step": 250
    },
    {
      "epoch": 0.5561613958560524,
      "grad_norm": 129.14805119503163,
      "learning_rate": 3.908505281649805e-07,
      "logits/chosen": -14.593152046203613,
      "logits/rejected": -14.538106918334961,
      "logps/chosen": -1.871093988418579,
      "logps/rejected": -2.36173152923584,
      "loss": 2.9231,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -18.710941314697266,
      "rewards/margins": 4.906374454498291,
      "rewards/rejected": -23.617313385009766,
      "step": 255
    },
    {
      "epoch": 0.5670665212649946,
      "grad_norm": 119.36679390601662,
      "learning_rate": 3.756144109505764e-07,
      "logits/chosen": -15.056838989257812,
      "logits/rejected": -15.148704528808594,
      "logps/chosen": -1.8173519372940063,
      "logps/rejected": -2.4098479747772217,
      "loss": 2.8089,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": -18.173521041870117,
      "rewards/margins": 5.924956321716309,
      "rewards/rejected": -24.098478317260742,
      "step": 260
    },
    {
      "epoch": 0.5779716466739367,
      "grad_norm": 119.03491048259538,
      "learning_rate": 3.6041373636809017e-07,
      "logits/chosen": -15.791028022766113,
      "logits/rejected": -15.623140335083008,
      "logps/chosen": -2.1847808361053467,
      "logps/rejected": -2.698356866836548,
      "loss": 3.1097,
      "rewards/accuracies": 0.7124999761581421,
      "rewards/chosen": -21.847808837890625,
      "rewards/margins": 5.135758399963379,
      "rewards/rejected": -26.983570098876953,
      "step": 265
    },
    {
      "epoch": 0.5888767720828789,
      "grad_norm": 119.59861773726368,
      "learning_rate": 3.4527059746243596e-07,
      "logits/chosen": -15.602317810058594,
      "logits/rejected": -15.679903030395508,
      "logps/chosen": -2.001981735229492,
      "logps/rejected": -2.4404501914978027,
      "loss": 3.1726,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -20.019817352294922,
      "rewards/margins": 4.384683132171631,
      "rewards/rejected": -24.404499053955078,
      "step": 270
    },
    {
      "epoch": 0.5997818974918212,
      "grad_norm": 143.27392642046235,
      "learning_rate": 3.302070036547201e-07,
      "logits/chosen": -15.435728073120117,
      "logits/rejected": -15.314961433410645,
      "logps/chosen": -2.168989896774292,
      "logps/rejected": -2.737366199493408,
      "loss": 3.0792,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -21.689899444580078,
      "rewards/margins": 5.683760643005371,
      "rewards/rejected": -27.3736572265625,
      "step": 275
    },
    {
      "epoch": 0.6106870229007634,
      "grad_norm": 114.06559627918872,
      "learning_rate": 3.152448487531915e-07,
      "logits/chosen": -15.740002632141113,
      "logits/rejected": -15.62346076965332,
      "logps/chosen": -1.9302055835723877,
      "logps/rejected": -2.405974864959717,
      "loss": 2.9702,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -19.302053451538086,
      "rewards/margins": 4.757693767547607,
      "rewards/rejected": -24.059749603271484,
      "step": 280
    },
    {
      "epoch": 0.6215921483097055,
      "grad_norm": 107.47439430691443,
      "learning_rate": 3.0040587913222725e-07,
      "logits/chosen": -14.441022872924805,
      "logits/rejected": -14.633172988891602,
      "logps/chosen": -1.8620984554290771,
      "logps/rejected": -2.345489263534546,
      "loss": 3.0044,
      "rewards/accuracies": 0.824999988079071,
      "rewards/chosen": -18.620983123779297,
      "rewards/margins": 4.833909034729004,
      "rewards/rejected": -23.45489501953125,
      "step": 285
    },
    {
      "epoch": 0.6324972737186477,
      "grad_norm": 154.10880275569383,
      "learning_rate": 2.857116621256018e-07,
      "logits/chosen": -15.512109756469727,
      "logits/rejected": -15.243085861206055,
      "logps/chosen": -1.9362213611602783,
      "logps/rejected": -2.402963399887085,
      "loss": 3.2181,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -19.36220932006836,
      "rewards/margins": 4.667425155639648,
      "rewards/rejected": -24.02963638305664,
      "step": 290
    },
    {
      "epoch": 0.64340239912759,
      "grad_norm": 112.54594135095479,
      "learning_rate": 2.7118355467997835e-07,
      "logits/chosen": -15.574625968933105,
      "logits/rejected": -15.6203031539917,
      "logps/chosen": -1.9695098400115967,
      "logps/rejected": -2.388803482055664,
      "loss": 2.9531,
      "rewards/accuracies": 0.6875,
      "rewards/chosen": -19.695100784301758,
      "rewards/margins": 4.192935943603516,
      "rewards/rejected": -23.888032913208008,
      "step": 295
    },
    {
      "epoch": 0.6543075245365322,
      "grad_norm": 131.96864010640354,
      "learning_rate": 2.5684267231418333e-07,
      "logits/chosen": -16.529542922973633,
      "logits/rejected": -16.21666145324707,
      "logps/chosen": -2.111536741256714,
      "logps/rejected": -2.6558353900909424,
      "loss": 2.8683,
      "rewards/accuracies": 0.7562500238418579,
      "rewards/chosen": -21.11536979675293,
      "rewards/margins": 5.442984580993652,
      "rewards/rejected": -26.5583553314209,
      "step": 300
    },
    {
      "epoch": 0.6652126499454744,
      "grad_norm": 131.182887672942,
      "learning_rate": 2.427098584293759e-07,
      "logits/chosen": -15.954015731811523,
      "logits/rejected": -15.863182067871094,
      "logps/chosen": -2.109534978866577,
      "logps/rejected": -2.553351640701294,
      "loss": 2.9835,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -21.09535026550293,
      "rewards/margins": 4.438165187835693,
      "rewards/rejected": -25.533517837524414,
      "step": 305
    },
    {
      "epoch": 0.6761177753544165,
      "grad_norm": 117.78663856490914,
      "learning_rate": 2.2880565401472287e-07,
      "logits/chosen": -15.363965034484863,
      "logits/rejected": -15.453814506530762,
      "logps/chosen": -2.0754928588867188,
      "logps/rejected": -2.7302043437957764,
      "loss": 2.9351,
      "rewards/accuracies": 0.793749988079071,
      "rewards/chosen": -20.754932403564453,
      "rewards/margins": 6.547112464904785,
      "rewards/rejected": -27.302043914794922,
      "step": 310
    },
    {
      "epoch": 0.6870229007633588,
      "grad_norm": 102.64755949094341,
      "learning_rate": 2.151502677926042e-07,
      "logits/chosen": -15.314285278320312,
      "logits/rejected": -15.252153396606445,
      "logps/chosen": -2.0483040809631348,
      "logps/rejected": -2.753251552581787,
      "loss": 2.7283,
      "rewards/accuracies": 0.7875000238418579,
      "rewards/chosen": -20.483041763305664,
      "rewards/margins": 7.049475193023682,
      "rewards/rejected": -27.532512664794922,
      "step": 315
    },
    {
      "epoch": 0.697928026172301,
      "grad_norm": 101.22013067633767,
      "learning_rate": 2.0176354684674575e-07,
      "logits/chosen": -16.12703514099121,
      "logits/rejected": -15.803072929382324,
      "logps/chosen": -2.151562213897705,
      "logps/rejected": -2.721626043319702,
      "loss": 2.9477,
      "rewards/accuracies": 0.793749988079071,
      "rewards/chosen": -21.515621185302734,
      "rewards/margins": 5.700636863708496,
      "rewards/rejected": -27.216259002685547,
      "step": 320
    },
    {
      "epoch": 0.7088331515812432,
      "grad_norm": 125.44812959020139,
      "learning_rate": 1.8866494777596534e-07,
      "logits/chosen": -15.385473251342773,
      "logits/rejected": -15.099942207336426,
      "logps/chosen": -2.194483518600464,
      "logps/rejected": -2.6979141235351562,
      "loss": 3.0708,
      "rewards/accuracies": 0.7562500238418579,
      "rewards/chosen": -21.944835662841797,
      "rewards/margins": 5.034304141998291,
      "rewards/rejected": -26.979141235351562,
      "step": 325
    },
    {
      "epoch": 0.7197382769901853,
      "grad_norm": 95.2552912244021,
      "learning_rate": 1.7587350841546007e-07,
      "logits/chosen": -15.96777629852295,
      "logits/rejected": -16.115123748779297,
      "logps/chosen": -1.9400516748428345,
      "logps/rejected": -2.51167368888855,
      "loss": 2.8539,
      "rewards/accuracies": 0.8062499761581421,
      "rewards/chosen": -19.400516510009766,
      "rewards/margins": 5.716220855712891,
      "rewards/rejected": -25.11673927307129,
      "step": 330
    },
    {
      "epoch": 0.7306434023991276,
      "grad_norm": 128.737726612481,
      "learning_rate": 1.634078201667347e-07,
      "logits/chosen": -15.98449420928955,
      "logits/rejected": -16.243240356445312,
      "logps/chosen": -2.1536316871643066,
      "logps/rejected": -2.7533793449401855,
      "loss": 2.9433,
      "rewards/accuracies": 0.8125,
      "rewards/chosen": -21.53631591796875,
      "rewards/margins": 5.9974775314331055,
      "rewards/rejected": -27.53379249572754,
      "step": 335
    },
    {
      "epoch": 0.7415485278080698,
      "grad_norm": 150.4063701643456,
      "learning_rate": 1.512860009763891e-07,
      "logits/chosen": -15.605964660644531,
      "logits/rejected": -16.06555938720703,
      "logps/chosen": -1.8955068588256836,
      "logps/rejected": -2.390796422958374,
      "loss": 2.9364,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -18.955068588256836,
      "rewards/margins": 4.952897071838379,
      "rewards/rejected": -23.9079647064209,
      "step": 340
    },
    {
      "epoch": 0.752453653217012,
      "grad_norm": 135.6732671225431,
      "learning_rate": 1.3952566900303602e-07,
      "logits/chosen": -16.359243392944336,
      "logits/rejected": -16.223447799682617,
      "logps/chosen": -2.149503469467163,
      "logps/rejected": -2.7064743041992188,
      "loss": 2.929,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -21.49503517150879,
      "rewards/margins": 5.569708347320557,
      "rewards/rejected": -27.064743041992188,
      "step": 345
    },
    {
      "epoch": 0.7633587786259542,
      "grad_norm": 125.758255019621,
      "learning_rate": 1.2814391701062391e-07,
      "logits/chosen": -14.762056350708008,
      "logits/rejected": -14.951733589172363,
      "logps/chosen": -2.00016450881958,
      "logps/rejected": -2.528573989868164,
      "loss": 3.0167,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -20.00164794921875,
      "rewards/margins": 5.284093379974365,
      "rewards/rejected": -25.285737991333008,
      "step": 350
    },
    {
      "epoch": 0.7742639040348964,
      "grad_norm": 123.77860608266465,
      "learning_rate": 1.17157287525381e-07,
      "logits/chosen": -14.972272872924805,
      "logits/rejected": -15.062446594238281,
      "logps/chosen": -2.0419890880584717,
      "logps/rejected": -2.7083749771118164,
      "loss": 3.1472,
      "rewards/accuracies": 0.793749988079071,
      "rewards/chosen": -20.419891357421875,
      "rewards/margins": 6.663858890533447,
      "rewards/rejected": -27.083751678466797,
      "step": 355
    },
    {
      "epoch": 0.7851690294438386,
      "grad_norm": 128.49716646084636,
      "learning_rate": 1.0658174879249e-07,
      "logits/chosen": -15.742474555969238,
      "logits/rejected": -15.778717041015625,
      "logps/chosen": -2.0881805419921875,
      "logps/rejected": -2.6481823921203613,
      "loss": 2.9629,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -20.881805419921875,
      "rewards/margins": 5.6000165939331055,
      "rewards/rejected": -26.481822967529297,
      "step": 360
    },
    {
      "epoch": 0.7960741548527808,
      "grad_norm": 105.81239577262055,
      "learning_rate": 9.643267156743626e-08,
      "logits/chosen": -15.465214729309082,
      "logits/rejected": -15.414011001586914,
      "logps/chosen": -2.0494282245635986,
      "logps/rejected": -2.6396915912628174,
      "loss": 2.9365,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -20.49428367614746,
      "rewards/margins": 5.9026336669921875,
      "rewards/rejected": -26.39691734313965,
      "step": 365
    },
    {
      "epoch": 0.806979280261723,
      "grad_norm": 106.05628455304455,
      "learning_rate": 8.672480677576266e-08,
      "logits/chosen": -15.015769958496094,
      "logits/rejected": -14.94062614440918,
      "logps/chosen": -2.0296902656555176,
      "logps/rejected": -2.648134231567383,
      "loss": 2.8337,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -20.29690170288086,
      "rewards/margins": 6.184438705444336,
      "rewards/rejected": -26.481342315673828,
      "step": 370
    },
    {
      "epoch": 0.8178844056706652,
      "grad_norm": 139.7590362680712,
      "learning_rate": 7.7472264073702e-08,
      "logits/chosen": -15.844355583190918,
      "logits/rejected": -15.914880752563477,
      "logps/chosen": -2.140533208847046,
      "logps/rejected": -2.731921434402466,
      "loss": 2.7732,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -21.405330657958984,
      "rewards/margins": 5.913882255554199,
      "rewards/rejected": -27.3192138671875,
      "step": 375
    },
    {
      "epoch": 0.8287895310796074,
      "grad_norm": 123.27756798269391,
      "learning_rate": 6.868849134084534e-08,
      "logits/chosen": -14.827451705932617,
      "logits/rejected": -14.896784782409668,
      "logps/chosen": -2.1212401390075684,
      "logps/rejected": -2.6644959449768066,
      "loss": 2.7211,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -21.21240234375,
      "rewards/margins": 5.432553291320801,
      "rewards/rejected": -26.644954681396484,
      "step": 380
    },
    {
      "epoch": 0.8396946564885496,
      "grad_norm": 102.3035020753843,
      "learning_rate": 6.038625513465372e-08,
      "logits/chosen": -15.892682075500488,
      "logits/rejected": -15.998703002929688,
      "logps/chosen": -2.0303637981414795,
      "logps/rejected": -2.4150891304016113,
      "loss": 2.9474,
      "rewards/accuracies": 0.6875,
      "rewards/chosen": -20.303638458251953,
      "rewards/margins": 3.8472514152526855,
      "rewards/rejected": -24.15088653564453,
      "step": 385
    },
    {
      "epoch": 0.8505997818974919,
      "grad_norm": 110.56211402431573,
      "learning_rate": 5.2577622135220546e-08,
      "logits/chosen": -15.678448677062988,
      "logits/rejected": -15.579752922058105,
      "logps/chosen": -2.146989345550537,
      "logps/rejected": -2.7087996006011963,
      "loss": 2.8812,
      "rewards/accuracies": 0.8062499761581421,
      "rewards/chosen": -21.469894409179688,
      "rewards/margins": 5.618099689483643,
      "rewards/rejected": -27.087993621826172,
      "step": 390
    },
    {
      "epoch": 0.861504907306434,
      "grad_norm": 136.5376747108113,
      "learning_rate": 4.527394160725375e-08,
      "logits/chosen": -15.390423774719238,
      "logits/rejected": -15.318705558776855,
      "logps/chosen": -2.0882272720336914,
      "logps/rejected": -2.5445287227630615,
      "loss": 2.9208,
      "rewards/accuracies": 0.7562500238418579,
      "rewards/chosen": -20.882274627685547,
      "rewards/margins": 4.563012599945068,
      "rewards/rejected": -25.44528579711914,
      "step": 395
    },
    {
      "epoch": 0.8724100327153762,
      "grad_norm": 110.11074526975227,
      "learning_rate": 3.848582890476728e-08,
      "logits/chosen": -13.988882064819336,
      "logits/rejected": -14.337841987609863,
      "logps/chosen": -1.9440772533416748,
      "logps/rejected": -2.5067763328552246,
      "loss": 2.7931,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -19.440773010253906,
      "rewards/margins": 5.626988410949707,
      "rewards/rejected": -25.067760467529297,
      "step": 400
    },
    {
      "epoch": 0.8724100327153762,
      "eval_logits/chosen": -13.542856216430664,
      "eval_logits/rejected": -13.604004859924316,
      "eval_logps/chosen": -1.9483120441436768,
      "eval_logps/rejected": -2.4730513095855713,
      "eval_loss": 2.762831449508667,
      "eval_rewards/accuracies": 0.7807376980781555,
      "eval_rewards/chosen": -19.48311996459961,
      "eval_rewards/margins": 5.247389316558838,
      "eval_rewards/rejected": -24.73050880432129,
      "eval_runtime": 66.0141,
      "eval_samples_per_second": 29.403,
      "eval_steps_per_second": 1.848,
      "step": 400
    },
    {
      "epoch": 0.8833151581243184,
      "grad_norm": 106.81510695706733,
      "learning_rate": 3.222315004245715e-08,
      "logits/chosen": -15.698140144348145,
      "logits/rejected": -15.643528938293457,
      "logps/chosen": -2.1120400428771973,
      "logps/rejected": -2.7991552352905273,
      "loss": 2.7377,
      "rewards/accuracies": 0.800000011920929,
      "rewards/chosen": -21.120403289794922,
      "rewards/margins": 6.871151924133301,
      "rewards/rejected": -27.991552352905273,
      "step": 405
    },
    {
      "epoch": 0.8942202835332607,
      "grad_norm": 127.02527681895255,
      "learning_rate": 2.6495007356185817e-08,
      "logits/chosen": -14.854022026062012,
      "logits/rejected": -15.155965805053711,
      "logps/chosen": -2.028907299041748,
      "logps/rejected": -2.589961528778076,
      "loss": 2.7722,
      "rewards/accuracies": 0.793749988079071,
      "rewards/chosen": -20.289072036743164,
      "rewards/margins": 5.610544204711914,
      "rewards/rejected": -25.899616241455078,
      "step": 410
    },
    {
      "epoch": 0.9051254089422028,
      "grad_norm": 118.59066087169252,
      "learning_rate": 2.1309726273417605e-08,
      "logits/chosen": -15.23388385772705,
      "logits/rejected": -15.077451705932617,
      "logps/chosen": -2.019240379333496,
      "logps/rejected": -2.622464895248413,
      "loss": 2.7385,
      "rewards/accuracies": 0.8062499761581421,
      "rewards/chosen": -20.192401885986328,
      "rewards/margins": 6.0322442054748535,
      "rewards/rejected": -26.224645614624023,
      "step": 415
    },
    {
      "epoch": 0.916030534351145,
      "grad_norm": 125.5340583771102,
      "learning_rate": 1.667484321283146e-08,
      "logits/chosen": -16.14217758178711,
      "logits/rejected": -15.975908279418945,
      "logps/chosen": -2.1546969413757324,
      "logps/rejected": -2.7095894813537598,
      "loss": 2.8351,
      "rewards/accuracies": 0.7875000238418579,
      "rewards/chosen": -21.54697036743164,
      "rewards/margins": 5.548920631408691,
      "rewards/rejected": -27.09589195251465,
      "step": 420
    },
    {
      "epoch": 0.9269356597600873,
      "grad_norm": 126.63955014497344,
      "learning_rate": 1.2597094630699156e-08,
      "logits/chosen": -16.528745651245117,
      "logits/rejected": -16.543407440185547,
      "logps/chosen": -1.9707733392715454,
      "logps/rejected": -2.48225474357605,
      "loss": 2.7718,
      "rewards/accuracies": 0.8125,
      "rewards/chosen": -19.707733154296875,
      "rewards/margins": 5.1148176193237305,
      "rewards/rejected": -24.822551727294922,
      "step": 425
    },
    {
      "epoch": 0.9378407851690295,
      "grad_norm": 141.60001641865918,
      "learning_rate": 9.082407229950017e-09,
      "logits/chosen": -16.1947021484375,
      "logits/rejected": -16.0598087310791,
      "logps/chosen": -2.160557985305786,
      "logps/rejected": -2.6326889991760254,
      "loss": 3.0444,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -21.605579376220703,
      "rewards/margins": 4.721311092376709,
      "rewards/rejected": -26.326889038085938,
      "step": 430
    },
    {
      "epoch": 0.9487459105779716,
      "grad_norm": 114.72393277887394,
      "learning_rate": 6.1358893461505916e-09,
      "logits/chosen": -15.125299453735352,
      "logits/rejected": -15.178013801574707,
      "logps/chosen": -2.1593708992004395,
      "logps/rejected": -2.785780429840088,
      "loss": 2.7684,
      "rewards/accuracies": 0.78125,
      "rewards/chosen": -21.593708038330078,
      "rewards/margins": 6.264092445373535,
      "rewards/rejected": -27.857803344726562,
      "step": 435
    },
    {
      "epoch": 0.9596510359869138,
      "grad_norm": 120.19496730357073,
      "learning_rate": 3.761823522920426e-09,
      "logits/chosen": -15.59558391571045,
      "logits/rejected": -15.761117935180664,
      "logps/chosen": -2.194023847579956,
      "logps/rejected": -2.795337677001953,
      "loss": 2.8552,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -21.940235137939453,
      "rewards/margins": 6.013139247894287,
      "rewards/rejected": -27.953380584716797,
      "step": 440
    },
    {
      "epoch": 0.9705561613958561,
      "grad_norm": 139.41969950436,
      "learning_rate": 1.963660287575086e-09,
      "logits/chosen": -14.188136100769043,
      "logits/rejected": -14.617622375488281,
      "logps/chosen": -2.016371011734009,
      "logps/rejected": -2.6547458171844482,
      "loss": 2.746,
      "rewards/accuracies": 0.8187500238418579,
      "rewards/chosen": -20.16370964050293,
      "rewards/margins": 6.383749961853027,
      "rewards/rejected": -26.547454833984375,
      "step": 445
    },
    {
      "epoch": 0.9814612868047983,
      "grad_norm": 121.93850210912117,
      "learning_rate": 7.44013136042465e-10,
      "logits/chosen": -15.886639595031738,
      "logits/rejected": -16.0024356842041,
      "logps/chosen": -2.115778684616089,
      "logps/rejected": -2.6602773666381836,
      "loss": 2.9859,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -21.157787322998047,
      "rewards/margins": 5.444985389709473,
      "rewards/rejected": -26.602771759033203,
      "step": 450
    },
    {
      "epoch": 0.9923664122137404,
      "grad_norm": 122.9917872545958,
      "learning_rate": 1.0465473434155824e-10,
      "logits/chosen": -13.859865188598633,
      "logits/rejected": -13.9098539352417,
      "logps/chosen": -2.006869077682495,
      "logps/rejected": -2.4212822914123535,
      "loss": 3.0127,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -20.06869125366211,
      "rewards/margins": 4.144131660461426,
      "rewards/rejected": -24.21282386779785,
      "step": 455
    },
    {
      "epoch": 0.9989094874591058,
      "step": 458,
      "total_flos": 0.0,
      "train_loss": 3.523275049492782,
      "train_runtime": 6692.3835,
      "train_samples_per_second": 8.769,
      "train_steps_per_second": 0.068
    }
  ],
  "logging_steps": 5,
  "max_steps": 458,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 1000000,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}