File size: 42,367 Bytes
a174eeb
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
{
  "best_global_step": null,
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 3.0,
  "eval_steps": 500,
  "global_step": 3750,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.04,
      "grad_norm": 24.803234100341797,
      "learning_rate": 1.3066666666666667e-06,
      "logits/chosen": 2.3097052574157715,
      "logits/rejected": 2.3757848739624023,
      "logps/chosen": -67.16219329833984,
      "logps/rejected": -80.8543930053711,
      "loss": 1.0777,
      "nll_loss": 0.41877487301826477,
      "rewards/accuracies": 0.7237499952316284,
      "rewards/chosen": -6.716219425201416,
      "rewards/margins": 1.3692210912704468,
      "rewards/rejected": -8.085439682006836,
      "step": 50
    },
    {
      "epoch": 0.08,
      "grad_norm": 12.049943923950195,
      "learning_rate": 2.64e-06,
      "logits/chosen": 2.2844557762145996,
      "logits/rejected": 2.364481210708618,
      "logps/chosen": -61.06942367553711,
      "logps/rejected": -75.95172119140625,
      "loss": 1.0254,
      "nll_loss": 0.38137227296829224,
      "rewards/accuracies": 0.7174999713897705,
      "rewards/chosen": -6.106942176818848,
      "rewards/margins": 1.4882298707962036,
      "rewards/rejected": -7.595171928405762,
      "step": 100
    },
    {
      "epoch": 0.12,
      "grad_norm": 19.429025650024414,
      "learning_rate": 3.973333333333333e-06,
      "logits/chosen": 2.2412264347076416,
      "logits/rejected": 2.2957522869110107,
      "logps/chosen": -56.071502685546875,
      "logps/rejected": -67.70003509521484,
      "loss": 1.0618,
      "nll_loss": 0.3372814655303955,
      "rewards/accuracies": 0.6650000214576721,
      "rewards/chosen": -5.607150077819824,
      "rewards/margins": 1.1628532409667969,
      "rewards/rejected": -6.770003795623779,
      "step": 150
    },
    {
      "epoch": 0.16,
      "grad_norm": 14.493091583251953,
      "learning_rate": 5.306666666666667e-06,
      "logits/chosen": 2.1775848865509033,
      "logits/rejected": 2.25844144821167,
      "logps/chosen": -50.50484848022461,
      "logps/rejected": -64.37080383300781,
      "loss": 0.9298,
      "nll_loss": 0.3056701719760895,
      "rewards/accuracies": 0.7237499952316284,
      "rewards/chosen": -5.050484657287598,
      "rewards/margins": 1.3865959644317627,
      "rewards/rejected": -6.437079906463623,
      "step": 200
    },
    {
      "epoch": 0.2,
      "grad_norm": 17.336822509765625,
      "learning_rate": 6.640000000000001e-06,
      "logits/chosen": 2.203662872314453,
      "logits/rejected": 2.28377103805542,
      "logps/chosen": -51.80202102661133,
      "logps/rejected": -65.30438232421875,
      "loss": 0.9291,
      "nll_loss": 0.2998887896537781,
      "rewards/accuracies": 0.7149999737739563,
      "rewards/chosen": -5.180202007293701,
      "rewards/margins": 1.3502373695373535,
      "rewards/rejected": -6.530439376831055,
      "step": 250
    },
    {
      "epoch": 0.24,
      "grad_norm": 25.542375564575195,
      "learning_rate": 7.973333333333334e-06,
      "logits/chosen": 2.2555084228515625,
      "logits/rejected": 2.3327550888061523,
      "logps/chosen": -49.52461242675781,
      "logps/rejected": -64.92198181152344,
      "loss": 0.8288,
      "nll_loss": 0.29204535484313965,
      "rewards/accuracies": 0.7674999833106995,
      "rewards/chosen": -4.9524617195129395,
      "rewards/margins": 1.5397369861602783,
      "rewards/rejected": -6.492198467254639,
      "step": 300
    },
    {
      "epoch": 0.28,
      "grad_norm": 10.780914306640625,
      "learning_rate": 9.306666666666667e-06,
      "logits/chosen": 2.2995264530181885,
      "logits/rejected": 2.3702523708343506,
      "logps/chosen": -49.0943489074707,
      "logps/rejected": -62.10791015625,
      "loss": 0.9434,
      "nll_loss": 0.29649874567985535,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -4.909435272216797,
      "rewards/margins": 1.3013556003570557,
      "rewards/rejected": -6.210791110992432,
      "step": 350
    },
    {
      "epoch": 0.32,
      "grad_norm": 11.368802070617676,
      "learning_rate": 9.998752338940612e-06,
      "logits/chosen": 2.4777705669403076,
      "logits/rejected": 2.5543289184570312,
      "logps/chosen": -46.3427848815918,
      "logps/rejected": -58.78908157348633,
      "loss": 0.8661,
      "nll_loss": 0.27519530057907104,
      "rewards/accuracies": 0.7137500047683716,
      "rewards/chosen": -4.634278297424316,
      "rewards/margins": 1.2446298599243164,
      "rewards/rejected": -5.878908634185791,
      "step": 400
    },
    {
      "epoch": 0.36,
      "grad_norm": 13.878002166748047,
      "learning_rate": 9.988142751731797e-06,
      "logits/chosen": 2.4681615829467773,
      "logits/rejected": 2.540001153945923,
      "logps/chosen": -46.92790985107422,
      "logps/rejected": -63.343467712402344,
      "loss": 0.8081,
      "nll_loss": 0.2872486710548401,
      "rewards/accuracies": 0.7412499785423279,
      "rewards/chosen": -4.692790985107422,
      "rewards/margins": 1.6415550708770752,
      "rewards/rejected": -6.334346294403076,
      "step": 450
    },
    {
      "epoch": 0.4,
      "grad_norm": 14.774677276611328,
      "learning_rate": 9.966729958067638e-06,
      "logits/chosen": 2.509334087371826,
      "logits/rejected": 2.5635790824890137,
      "logps/chosen": -48.37749481201172,
      "logps/rejected": -64.00433349609375,
      "loss": 0.873,
      "nll_loss": 0.2907007336616516,
      "rewards/accuracies": 0.7200000286102295,
      "rewards/chosen": -4.837749481201172,
      "rewards/margins": 1.5626837015151978,
      "rewards/rejected": -6.400432586669922,
      "step": 500
    },
    {
      "epoch": 0.44,
      "grad_norm": 18.610092163085938,
      "learning_rate": 9.934560333291077e-06,
      "logits/chosen": 2.5441677570343018,
      "logits/rejected": 2.613149404525757,
      "logps/chosen": -48.265201568603516,
      "logps/rejected": -63.49806213378906,
      "loss": 0.8581,
      "nll_loss": 0.2854183316230774,
      "rewards/accuracies": 0.7287499904632568,
      "rewards/chosen": -4.826520919799805,
      "rewards/margins": 1.5232864618301392,
      "rewards/rejected": -6.349807262420654,
      "step": 550
    },
    {
      "epoch": 0.48,
      "grad_norm": 13.354905128479004,
      "learning_rate": 9.89170354964594e-06,
      "logits/chosen": 2.646564245223999,
      "logits/rejected": 2.700756549835205,
      "logps/chosen": -45.61009216308594,
      "logps/rejected": -60.2086067199707,
      "loss": 0.8363,
      "nll_loss": 0.27716541290283203,
      "rewards/accuracies": 0.7300000190734863,
      "rewards/chosen": -4.561009883880615,
      "rewards/margins": 1.4598517417907715,
      "rewards/rejected": -6.0208611488342285,
      "step": 600
    },
    {
      "epoch": 0.52,
      "grad_norm": 5.81559419631958,
      "learning_rate": 9.83825242538238e-06,
      "logits/chosen": 2.519411087036133,
      "logits/rejected": 2.5865354537963867,
      "logps/chosen": -46.41543960571289,
      "logps/rejected": -63.667991638183594,
      "loss": 0.7762,
      "nll_loss": 0.27636775374412537,
      "rewards/accuracies": 0.7587500214576721,
      "rewards/chosen": -4.641544342041016,
      "rewards/margins": 1.7252554893493652,
      "rewards/rejected": -6.366799831390381,
      "step": 650
    },
    {
      "epoch": 0.56,
      "grad_norm": 9.440147399902344,
      "learning_rate": 9.774322723733216e-06,
      "logits/chosen": 2.4503791332244873,
      "logits/rejected": 2.5173845291137695,
      "logps/chosen": -44.43014144897461,
      "logps/rejected": -62.06157302856445,
      "loss": 0.7447,
      "nll_loss": 0.2715470492839813,
      "rewards/accuracies": 0.7724999785423279,
      "rewards/chosen": -4.443014621734619,
      "rewards/margins": 1.7631423473358154,
      "rewards/rejected": -6.206157207489014,
      "step": 700
    },
    {
      "epoch": 0.6,
      "grad_norm": 9.054909706115723,
      "learning_rate": 9.700052902196541e-06,
      "logits/chosen": 2.3488733768463135,
      "logits/rejected": 2.4224016666412354,
      "logps/chosen": -45.02783203125,
      "logps/rejected": -63.32554626464844,
      "loss": 0.7256,
      "nll_loss": 0.27446305751800537,
      "rewards/accuracies": 0.8075000047683716,
      "rewards/chosen": -4.502783298492432,
      "rewards/margins": 1.8297713994979858,
      "rewards/rejected": -6.332554817199707,
      "step": 750
    },
    {
      "epoch": 0.64,
      "grad_norm": 16.773876190185547,
      "learning_rate": 9.615603812667618e-06,
      "logits/chosen": 2.4701356887817383,
      "logits/rejected": 2.5356545448303223,
      "logps/chosen": -44.635948181152344,
      "logps/rejected": -62.55714416503906,
      "loss": 0.7836,
      "nll_loss": 0.2690128684043884,
      "rewards/accuracies": 0.7524999976158142,
      "rewards/chosen": -4.463594913482666,
      "rewards/margins": 1.7921198606491089,
      "rewards/rejected": -6.2557148933410645,
      "step": 800
    },
    {
      "epoch": 0.68,
      "grad_norm": 12.325571060180664,
      "learning_rate": 9.521158353069494e-06,
      "logits/chosen": 2.450385332107544,
      "logits/rejected": 2.5140254497528076,
      "logps/chosen": -47.131587982177734,
      "logps/rejected": -65.35118865966797,
      "loss": 0.7507,
      "nll_loss": 0.27970045804977417,
      "rewards/accuracies": 0.7762500047683716,
      "rewards/chosen": -4.713159561157227,
      "rewards/margins": 1.8219600915908813,
      "rewards/rejected": -6.535119533538818,
      "step": 850
    },
    {
      "epoch": 0.72,
      "grad_norm": 18.231292724609375,
      "learning_rate": 9.416921071236821e-06,
      "logits/chosen": 2.4923791885375977,
      "logits/rejected": 2.5595548152923584,
      "logps/chosen": -46.485965728759766,
      "logps/rejected": -63.454063415527344,
      "loss": 0.8048,
      "nll_loss": 0.27803030610084534,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -4.648596286773682,
      "rewards/margins": 1.6968092918395996,
      "rewards/rejected": -6.345406532287598,
      "step": 900
    },
    {
      "epoch": 0.76,
      "grad_norm": 18.762359619140625,
      "learning_rate": 9.303117721910801e-06,
      "logits/chosen": 2.571411371231079,
      "logits/rejected": 2.6245839595794678,
      "logps/chosen": -44.32810974121094,
      "logps/rejected": -61.370662689208984,
      "loss": 0.7592,
      "nll_loss": 0.2649703919887543,
      "rewards/accuracies": 0.7662500143051147,
      "rewards/chosen": -4.4328107833862305,
      "rewards/margins": 1.7042557001113892,
      "rewards/rejected": -6.137066841125488,
      "step": 950
    },
    {
      "epoch": 0.8,
      "grad_norm": 13.24950122833252,
      "learning_rate": 9.179994777804677e-06,
      "logits/chosen": 2.5039403438568115,
      "logits/rejected": 2.5774261951446533,
      "logps/chosen": -45.20160675048828,
      "logps/rejected": -62.881683349609375,
      "loss": 0.7717,
      "nll_loss": 0.2698582708835602,
      "rewards/accuracies": 0.7787500023841858,
      "rewards/chosen": -4.520160675048828,
      "rewards/margins": 1.7680076360702515,
      "rewards/rejected": -6.288168430328369,
      "step": 1000
    },
    {
      "epoch": 0.84,
      "grad_norm": 12.429143905639648,
      "learning_rate": 9.04781889579873e-06,
      "logits/chosen": 2.5070714950561523,
      "logits/rejected": 2.5675339698791504,
      "logps/chosen": -45.09945297241211,
      "logps/rejected": -64.37616729736328,
      "loss": 0.7095,
      "nll_loss": 0.2686724364757538,
      "rewards/accuracies": 0.7912499904632568,
      "rewards/chosen": -4.509944915771484,
      "rewards/margins": 1.9276715517044067,
      "rewards/rejected": -6.437616348266602,
      "step": 1050
    },
    {
      "epoch": 0.88,
      "grad_norm": 13.30483341217041,
      "learning_rate": 8.906876339420851e-06,
      "logits/chosen": 2.388920783996582,
      "logits/rejected": 2.4621076583862305,
      "logps/chosen": -42.8404541015625,
      "logps/rejected": -62.490779876708984,
      "loss": 0.6913,
      "nll_loss": 0.25781476497650146,
      "rewards/accuracies": 0.8112499713897705,
      "rewards/chosen": -4.284045219421387,
      "rewards/margins": 1.9650330543518066,
      "rewards/rejected": -6.249078750610352,
      "step": 1100
    },
    {
      "epoch": 0.92,
      "grad_norm": 13.460766792297363,
      "learning_rate": 8.757472358863481e-06,
      "logits/chosen": 2.4224061965942383,
      "logits/rejected": 2.5014472007751465,
      "logps/chosen": -45.438575744628906,
      "logps/rejected": -63.852115631103516,
      "loss": 0.7312,
      "nll_loss": 0.26547786593437195,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -4.543857574462891,
      "rewards/margins": 1.841354250907898,
      "rewards/rejected": -6.385211944580078,
      "step": 1150
    },
    {
      "epoch": 0.96,
      "grad_norm": 9.995412826538086,
      "learning_rate": 8.599930529879669e-06,
      "logits/chosen": 2.289586067199707,
      "logits/rejected": 2.3660268783569336,
      "logps/chosen": -43.745357513427734,
      "logps/rejected": -62.561038970947266,
      "loss": 0.6825,
      "nll_loss": 0.2639794647693634,
      "rewards/accuracies": 0.8100000023841858,
      "rewards/chosen": -4.37453556060791,
      "rewards/margins": 1.8815685510635376,
      "rewards/rejected": -6.256104946136475,
      "step": 1200
    },
    {
      "epoch": 1.0,
      "grad_norm": 9.7256498336792,
      "learning_rate": 8.434592052990044e-06,
      "logits/chosen": 2.1411895751953125,
      "logits/rejected": 2.216475248336792,
      "logps/chosen": -41.174503326416016,
      "logps/rejected": -59.75578308105469,
      "loss": 0.687,
      "nll_loss": 0.25248458981513977,
      "rewards/accuracies": 0.7912499904632568,
      "rewards/chosen": -4.11745023727417,
      "rewards/margins": 1.8581281900405884,
      "rewards/rejected": -5.9755778312683105,
      "step": 1250
    },
    {
      "epoch": 1.04,
      "grad_norm": 5.8602166175842285,
      "learning_rate": 8.261815014518465e-06,
      "logits/chosen": 2.094317674636841,
      "logits/rejected": 2.1707794666290283,
      "logps/chosen": -42.47966003417969,
      "logps/rejected": -67.64102172851562,
      "loss": 0.5271,
      "nll_loss": 0.25495222210884094,
      "rewards/accuracies": 0.8837500214576721,
      "rewards/chosen": -4.247966289520264,
      "rewards/margins": 2.516136884689331,
      "rewards/rejected": -6.764101982116699,
      "step": 1300
    },
    {
      "epoch": 1.08,
      "grad_norm": 7.209846019744873,
      "learning_rate": 8.081973611056784e-06,
      "logits/chosen": 1.899792194366455,
      "logits/rejected": 1.9656882286071777,
      "logps/chosen": -41.67923355102539,
      "logps/rejected": -67.45558166503906,
      "loss": 0.5335,
      "nll_loss": 0.24939106404781342,
      "rewards/accuracies": 0.8774999976158142,
      "rewards/chosen": -4.167923450469971,
      "rewards/margins": 2.577634811401367,
      "rewards/rejected": -6.745559215545654,
      "step": 1350
    },
    {
      "epoch": 1.12,
      "grad_norm": 12.841861724853516,
      "learning_rate": 7.89545733903834e-06,
      "logits/chosen": 1.8675593137741089,
      "logits/rejected": 1.9421709775924683,
      "logps/chosen": -43.389644622802734,
      "logps/rejected": -68.71646881103516,
      "loss": 0.5398,
      "nll_loss": 0.2613371014595032,
      "rewards/accuracies": 0.8974999785423279,
      "rewards/chosen": -4.338963985443115,
      "rewards/margins": 2.5326828956604004,
      "rewards/rejected": -6.871647834777832,
      "step": 1400
    },
    {
      "epoch": 1.16,
      "grad_norm": 13.226041793823242,
      "learning_rate": 7.702670151175435e-06,
      "logits/chosen": 1.8003557920455933,
      "logits/rejected": 1.877779245376587,
      "logps/chosen": -42.74818801879883,
      "logps/rejected": -67.07624053955078,
      "loss": 0.5569,
      "nll_loss": 0.25505152344703674,
      "rewards/accuracies": 0.8899999856948853,
      "rewards/chosen": -4.274819374084473,
      "rewards/margins": 2.4328057765960693,
      "rewards/rejected": -6.707624435424805,
      "step": 1450
    },
    {
      "epoch": 1.2,
      "grad_norm": 6.528555393218994,
      "learning_rate": 7.5040295815877e-06,
      "logits/chosen": 1.8589718341827393,
      "logits/rejected": 1.9315310716629028,
      "logps/chosen": -42.12541198730469,
      "logps/rejected": -68.34070587158203,
      "loss": 0.5316,
      "nll_loss": 0.25116249918937683,
      "rewards/accuracies": 0.8849999904632568,
      "rewards/chosen": -4.212540626525879,
      "rewards/margins": 2.621530055999756,
      "rewards/rejected": -6.834071636199951,
      "step": 1500
    },
    {
      "epoch": 1.24,
      "grad_norm": 8.806782722473145,
      "learning_rate": 7.299965841516164e-06,
      "logits/chosen": 1.896395206451416,
      "logits/rejected": 1.9738116264343262,
      "logps/chosen": -42.76396560668945,
      "logps/rejected": -68.06053161621094,
      "loss": 0.5404,
      "nll_loss": 0.2575341463088989,
      "rewards/accuracies": 0.8799999952316284,
      "rewards/chosen": -4.27639627456665,
      "rewards/margins": 2.5296568870544434,
      "rewards/rejected": -6.8060526847839355,
      "step": 1550
    },
    {
      "epoch": 1.28,
      "grad_norm": 13.923650741577148,
      "learning_rate": 7.090920887581507e-06,
      "logits/chosen": 1.881974220275879,
      "logits/rejected": 1.9411964416503906,
      "logps/chosen": -41.719696044921875,
      "logps/rejected": -66.24463653564453,
      "loss": 0.5483,
      "nll_loss": 0.24970975518226624,
      "rewards/accuracies": 0.8762500286102295,
      "rewards/chosen": -4.171969413757324,
      "rewards/margins": 2.452493667602539,
      "rewards/rejected": -6.62446403503418,
      "step": 1600
    },
    {
      "epoch": 1.32,
      "grad_norm": 11.483983993530273,
      "learning_rate": 6.877347464604446e-06,
      "logits/chosen": 1.8503919839859009,
      "logits/rejected": 1.9176833629608154,
      "logps/chosen": -42.25181579589844,
      "logps/rejected": -67.97564697265625,
      "loss": 0.541,
      "nll_loss": 0.25359034538269043,
      "rewards/accuracies": 0.8812500238418579,
      "rewards/chosen": -4.225182056427002,
      "rewards/margins": 2.572382926940918,
      "rewards/rejected": -6.797564506530762,
      "step": 1650
    },
    {
      "epoch": 1.3599999999999999,
      "grad_norm": 9.980610847473145,
      "learning_rate": 6.659708125061242e-06,
      "logits/chosen": 1.7879056930541992,
      "logits/rejected": 1.8512917757034302,
      "logps/chosen": -41.724449157714844,
      "logps/rejected": -68.18392181396484,
      "loss": 0.5127,
      "nll_loss": 0.2496194839477539,
      "rewards/accuracies": 0.9075000286102295,
      "rewards/chosen": -4.172445297241211,
      "rewards/margins": 2.645946741104126,
      "rewards/rejected": -6.818392276763916,
      "step": 1700
    },
    {
      "epoch": 1.4,
      "grad_norm": 8.7877779006958,
      "learning_rate": 6.438474227298065e-06,
      "logits/chosen": 1.813609004020691,
      "logits/rejected": 1.8792080879211426,
      "logps/chosen": -42.19902420043945,
      "logps/rejected": -67.10396575927734,
      "loss": 0.5475,
      "nll_loss": 0.2563365399837494,
      "rewards/accuracies": 0.8887500166893005,
      "rewards/chosen": -4.219902038574219,
      "rewards/margins": 2.4904944896698,
      "rewards/rejected": -6.710396766662598,
      "step": 1750
    },
    {
      "epoch": 1.44,
      "grad_norm": 9.004395484924316,
      "learning_rate": 6.2141249146737545e-06,
      "logits/chosen": 1.8060498237609863,
      "logits/rejected": 1.8766037225723267,
      "logps/chosen": -40.13544845581055,
      "logps/rejected": -64.01081085205078,
      "loss": 0.5417,
      "nll_loss": 0.23975929617881775,
      "rewards/accuracies": 0.8712499737739563,
      "rewards/chosen": -4.013545036315918,
      "rewards/margins": 2.3875370025634766,
      "rewards/rejected": -6.401081562042236,
      "step": 1800
    },
    {
      "epoch": 1.48,
      "grad_norm": 14.659512519836426,
      "learning_rate": 5.987146077842015e-06,
      "logits/chosen": 1.7913141250610352,
      "logits/rejected": 1.8706274032592773,
      "logps/chosen": -41.8923454284668,
      "logps/rejected": -68.07263946533203,
      "loss": 0.5291,
      "nll_loss": 0.24965395033359528,
      "rewards/accuracies": 0.8812500238418579,
      "rewards/chosen": -4.189234733581543,
      "rewards/margins": 2.6180291175842285,
      "rewards/rejected": -6.8072638511657715,
      "step": 1850
    },
    {
      "epoch": 1.52,
      "grad_norm": 6.758883953094482,
      "learning_rate": 5.7580293024204455e-06,
      "logits/chosen": 1.7456320524215698,
      "logits/rejected": 1.8188166618347168,
      "logps/chosen": -40.59235763549805,
      "logps/rejected": -67.43293762207031,
      "loss": 0.4978,
      "nll_loss": 0.2416204810142517,
      "rewards/accuracies": 0.8924999833106995,
      "rewards/chosen": -4.0592360496521,
      "rewards/margins": 2.6840572357177734,
      "rewards/rejected": -6.743292808532715,
      "step": 1900
    },
    {
      "epoch": 1.56,
      "grad_norm": 11.839475631713867,
      "learning_rate": 5.5272708043255605e-06,
      "logits/chosen": 1.771327257156372,
      "logits/rejected": 1.8334521055221558,
      "logps/chosen": -40.4869499206543,
      "logps/rejected": -64.32723999023438,
      "loss": 0.5586,
      "nll_loss": 0.24806198477745056,
      "rewards/accuracies": 0.8675000071525574,
      "rewards/chosen": -4.048694610595703,
      "rewards/margins": 2.3840293884277344,
      "rewards/rejected": -6.4327239990234375,
      "step": 1950
    },
    {
      "epoch": 1.6,
      "grad_norm": 12.160818099975586,
      "learning_rate": 5.295370355079615e-06,
      "logits/chosen": 1.7191228866577148,
      "logits/rejected": 1.7949659824371338,
      "logps/chosen": -40.17674255371094,
      "logps/rejected": -67.11283874511719,
      "loss": 0.5123,
      "nll_loss": 0.24404479563236237,
      "rewards/accuracies": 0.8774999976158142,
      "rewards/chosen": -4.017674446105957,
      "rewards/margins": 2.6936092376708984,
      "rewards/rejected": -6.711284160614014,
      "step": 2000
    },
    {
      "epoch": 1.6400000000000001,
      "grad_norm": 12.399029731750488,
      "learning_rate": 5.062830199416764e-06,
      "logits/chosen": 1.7645864486694336,
      "logits/rejected": 1.8290934562683105,
      "logps/chosen": -41.4688720703125,
      "logps/rejected": -67.27618408203125,
      "loss": 0.5257,
      "nll_loss": 0.24901245534420013,
      "rewards/accuracies": 0.8949999809265137,
      "rewards/chosen": -4.146886825561523,
      "rewards/margins": 2.580731153488159,
      "rewards/rejected": -6.727618217468262,
      "step": 2050
    },
    {
      "epoch": 1.6800000000000002,
      "grad_norm": 13.637982368469238,
      "learning_rate": 4.8301539675328205e-06,
      "logits/chosen": 1.6826503276824951,
      "logits/rejected": 1.7500274181365967,
      "logps/chosen": -40.96113586425781,
      "logps/rejected": -67.78974914550781,
      "loss": 0.513,
      "nll_loss": 0.2408064305782318,
      "rewards/accuracies": 0.8987500071525574,
      "rewards/chosen": -4.096114158630371,
      "rewards/margins": 2.6828606128692627,
      "rewards/rejected": -6.778974533081055,
      "step": 2100
    },
    {
      "epoch": 1.72,
      "grad_norm": 5.442174911499023,
      "learning_rate": 4.597845584334387e-06,
      "logits/chosen": 1.6728637218475342,
      "logits/rejected": 1.7252342700958252,
      "logps/chosen": -42.251564025878906,
      "logps/rejected": -69.24467468261719,
      "loss": 0.5304,
      "nll_loss": 0.25177592039108276,
      "rewards/accuracies": 0.8899999856948853,
      "rewards/chosen": -4.225156307220459,
      "rewards/margins": 2.699312210083008,
      "rewards/rejected": -6.92446756362915,
      "step": 2150
    },
    {
      "epoch": 1.76,
      "grad_norm": 10.488734245300293,
      "learning_rate": 4.366408178049728e-06,
      "logits/chosen": 1.6670646667480469,
      "logits/rejected": 1.7333457469940186,
      "logps/chosen": -39.98529052734375,
      "logps/rejected": -65.87380981445312,
      "loss": 0.5198,
      "nll_loss": 0.2439391314983368,
      "rewards/accuracies": 0.8762500286102295,
      "rewards/chosen": -3.9985289573669434,
      "rewards/margins": 2.588852643966675,
      "rewards/rejected": -6.587381362915039,
      "step": 2200
    },
    {
      "epoch": 1.8,
      "grad_norm": 11.035846710205078,
      "learning_rate": 4.136342990565055e-06,
      "logits/chosen": 1.6644293069839478,
      "logits/rejected": 1.7320383787155151,
      "logps/chosen": -40.47782897949219,
      "logps/rejected": -69.36231994628906,
      "loss": 0.5059,
      "nll_loss": 0.2478472888469696,
      "rewards/accuracies": 0.8949999809265137,
      "rewards/chosen": -4.047782897949219,
      "rewards/margins": 2.888448476791382,
      "rewards/rejected": -6.93623161315918,
      "step": 2250
    },
    {
      "epoch": 1.8399999999999999,
      "grad_norm": 14.5468168258667,
      "learning_rate": 3.908148291846225e-06,
      "logits/chosen": 1.6440070867538452,
      "logits/rejected": 1.6980115175247192,
      "logps/chosen": -37.658931732177734,
      "logps/rejected": -64.18938446044922,
      "loss": 0.4975,
      "nll_loss": 0.23423399031162262,
      "rewards/accuracies": 0.8899999856948853,
      "rewards/chosen": -3.7658934593200684,
      "rewards/margins": 2.6530449390411377,
      "rewards/rejected": -6.418938636779785,
      "step": 2300
    },
    {
      "epoch": 1.88,
      "grad_norm": 15.06531810760498,
      "learning_rate": 3.6823183007969375e-06,
      "logits/chosen": 1.633602261543274,
      "logits/rejected": 1.69327974319458,
      "logps/chosen": -38.722007751464844,
      "logps/rejected": -65.98956298828125,
      "loss": 0.512,
      "nll_loss": 0.23453904688358307,
      "rewards/accuracies": 0.8837500214576721,
      "rewards/chosen": -3.8722009658813477,
      "rewards/margins": 2.726755380630493,
      "rewards/rejected": -6.598956108093262,
      "step": 2350
    },
    {
      "epoch": 1.92,
      "grad_norm": 9.41374397277832,
      "learning_rate": 3.4593421148906523e-06,
      "logits/chosen": 1.6280794143676758,
      "logits/rejected": 1.694667935371399,
      "logps/chosen": -41.9329833984375,
      "logps/rejected": -68.24996948242188,
      "loss": 0.5314,
      "nll_loss": 0.24245284497737885,
      "rewards/accuracies": 0.8887500166893005,
      "rewards/chosen": -4.19329833984375,
      "rewards/margins": 2.6316983699798584,
      "rewards/rejected": -6.8249969482421875,
      "step": 2400
    },
    {
      "epoch": 1.96,
      "grad_norm": 10.686447143554688,
      "learning_rate": 3.239702650894364e-06,
      "logits/chosen": 1.648103952407837,
      "logits/rejected": 1.7120572328567505,
      "logps/chosen": -39.89656066894531,
      "logps/rejected": -67.84294128417969,
      "loss": 0.5132,
      "nll_loss": 0.23985882103443146,
      "rewards/accuracies": 0.8799999952316284,
      "rewards/chosen": -3.989656448364258,
      "rewards/margins": 2.794638156890869,
      "rewards/rejected": -6.784294605255127,
      "step": 2450
    },
    {
      "epoch": 2.0,
      "grad_norm": 6.338383197784424,
      "learning_rate": 3.023875598978419e-06,
      "logits/chosen": 1.6715092658996582,
      "logits/rejected": 1.7344988584518433,
      "logps/chosen": -39.51519012451172,
      "logps/rejected": -68.46276092529297,
      "loss": 0.4834,
      "nll_loss": 0.24019798636436462,
      "rewards/accuracies": 0.9112499952316284,
      "rewards/chosen": -3.951519250869751,
      "rewards/margins": 2.8947577476501465,
      "rewards/rejected": -6.846276760101318,
      "step": 2500
    },
    {
      "epoch": 2.04,
      "grad_norm": 5.144768238067627,
      "learning_rate": 2.812328392477536e-06,
      "logits/chosen": 1.5945950746536255,
      "logits/rejected": 1.6544932126998901,
      "logps/chosen": -37.963443756103516,
      "logps/rejected": -69.92556762695312,
      "loss": 0.3913,
      "nll_loss": 0.22887782752513885,
      "rewards/accuracies": 0.9712499976158142,
      "rewards/chosen": -3.796344041824341,
      "rewards/margins": 3.196213722229004,
      "rewards/rejected": -6.992558002471924,
      "step": 2550
    },
    {
      "epoch": 2.08,
      "grad_norm": 5.904083251953125,
      "learning_rate": 2.6055191955342886e-06,
      "logits/chosen": 1.5594576597213745,
      "logits/rejected": 1.6052813529968262,
      "logps/chosen": -39.600093841552734,
      "logps/rejected": -73.07125854492188,
      "loss": 0.398,
      "nll_loss": 0.23794788122177124,
      "rewards/accuracies": 0.9662500023841858,
      "rewards/chosen": -3.960009813308716,
      "rewards/margins": 3.3471157550811768,
      "rewards/rejected": -7.307126045227051,
      "step": 2600
    },
    {
      "epoch": 2.12,
      "grad_norm": 5.723136901855469,
      "learning_rate": 2.403895910817593e-06,
      "logits/chosen": 1.532933235168457,
      "logits/rejected": 1.5803064107894897,
      "logps/chosen": -38.91565704345703,
      "logps/rejected": -70.781982421875,
      "loss": 0.4101,
      "nll_loss": 0.23784016072750092,
      "rewards/accuracies": 0.9574999809265137,
      "rewards/chosen": -3.8915653228759766,
      "rewards/margins": 3.186633348464966,
      "rewards/rejected": -7.07819938659668,
      "step": 2650
    },
    {
      "epoch": 2.16,
      "grad_norm": 7.6983561515808105,
      "learning_rate": 2.2078952094652705e-06,
      "logits/chosen": 1.4740697145462036,
      "logits/rejected": 1.5456255674362183,
      "logps/chosen": -39.75840377807617,
      "logps/rejected": -75.84871673583984,
      "loss": 0.3908,
      "nll_loss": 0.23151804506778717,
      "rewards/accuracies": 0.9612500071525574,
      "rewards/chosen": -3.9758400917053223,
      "rewards/margins": 3.6090316772460938,
      "rewards/rejected": -7.584871292114258,
      "step": 2700
    },
    {
      "epoch": 2.2,
      "grad_norm": 7.529278755187988,
      "learning_rate": 2.017941585351591e-06,
      "logits/chosen": 1.495976209640503,
      "logits/rejected": 1.5389071702957153,
      "logps/chosen": -39.2768440246582,
      "logps/rejected": -71.7361831665039,
      "loss": 0.4138,
      "nll_loss": 0.24059143662452698,
      "rewards/accuracies": 0.9574999809265137,
      "rewards/chosen": -3.9276845455169678,
      "rewards/margins": 3.245933771133423,
      "rewards/rejected": -7.173618793487549,
      "step": 2750
    },
    {
      "epoch": 2.24,
      "grad_norm": 6.076731204986572,
      "learning_rate": 1.8344464357280722e-06,
      "logits/chosen": 1.4751191139221191,
      "logits/rejected": 1.5268608331680298,
      "logps/chosen": -38.4973258972168,
      "logps/rejected": -71.8733901977539,
      "loss": 0.3943,
      "nll_loss": 0.22699041664600372,
      "rewards/accuracies": 0.9700000286102295,
      "rewards/chosen": -3.8497328758239746,
      "rewards/margins": 3.3376071453094482,
      "rewards/rejected": -7.187338829040527,
      "step": 2800
    },
    {
      "epoch": 2.2800000000000002,
      "grad_norm": 5.921169757843018,
      "learning_rate": 1.6578071702286396e-06,
      "logits/chosen": 1.4693129062652588,
      "logits/rejected": 1.5106064081192017,
      "logps/chosen": -38.8640022277832,
      "logps/rejected": -72.8025131225586,
      "loss": 0.4103,
      "nll_loss": 0.2377665936946869,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": -3.8864006996154785,
      "rewards/margins": 3.393850088119507,
      "rewards/rejected": -7.280250072479248,
      "step": 2850
    },
    {
      "epoch": 2.32,
      "grad_norm": 5.996314525604248,
      "learning_rate": 1.4884063501688539e-06,
      "logits/chosen": 1.4711755514144897,
      "logits/rejected": 1.537732720375061,
      "logps/chosen": -38.81364440917969,
      "logps/rejected": -73.80380249023438,
      "loss": 0.3908,
      "nll_loss": 0.22893303632736206,
      "rewards/accuracies": 0.9737499952316284,
      "rewards/chosen": -3.881364345550537,
      "rewards/margins": 3.499016523361206,
      "rewards/rejected": -7.380380153656006,
      "step": 2900
    },
    {
      "epoch": 2.36,
      "grad_norm": 10.235599517822266,
      "learning_rate": 1.3266108600032928e-06,
      "logits/chosen": 1.469245195388794,
      "logits/rejected": 1.5233672857284546,
      "logps/chosen": -39.08128356933594,
      "logps/rejected": -70.8739242553711,
      "loss": 0.4124,
      "nll_loss": 0.2312445044517517,
      "rewards/accuracies": 0.9512500166893005,
      "rewards/chosen": -3.9081289768218994,
      "rewards/margins": 3.1792635917663574,
      "rewards/rejected": -7.087392807006836,
      "step": 2950
    },
    {
      "epoch": 2.4,
      "grad_norm": 8.700451850891113,
      "learning_rate": 1.1727711127355118e-06,
      "logits/chosen": 1.445876955986023,
      "logits/rejected": 1.494391679763794,
      "logps/chosen": -38.116580963134766,
      "logps/rejected": -70.0914535522461,
      "loss": 0.3982,
      "nll_loss": 0.22633756697177887,
      "rewards/accuracies": 0.9574999809265137,
      "rewards/chosen": -3.8116586208343506,
      "rewards/margins": 3.1974875926971436,
      "rewards/rejected": -7.009146213531494,
      "step": 3000
    },
    {
      "epoch": 2.44,
      "grad_norm": 9.24150562286377,
      "learning_rate": 1.0272202910015083e-06,
      "logits/chosen": 1.4509037733078003,
      "logits/rejected": 1.5117911100387573,
      "logps/chosen": -39.70808410644531,
      "logps/rejected": -74.1922836303711,
      "loss": 0.4038,
      "nll_loss": 0.23606754839420319,
      "rewards/accuracies": 0.9662500023841858,
      "rewards/chosen": -3.970808744430542,
      "rewards/margins": 3.4484190940856934,
      "rewards/rejected": -7.419227123260498,
      "step": 3050
    },
    {
      "epoch": 2.48,
      "grad_norm": 10.763381958007812,
      "learning_rate": 8.902736254703347e-07,
      "logits/chosen": 1.4675711393356323,
      "logits/rejected": 1.5065741539001465,
      "logps/chosen": -36.786865234375,
      "logps/rejected": -67.48006439208984,
      "loss": 0.405,
      "nll_loss": 0.22343981266021729,
      "rewards/accuracies": 0.9624999761581421,
      "rewards/chosen": -3.6786866188049316,
      "rewards/margins": 3.0693204402923584,
      "rewards/rejected": -6.748007297515869,
      "step": 3100
    },
    {
      "epoch": 2.52,
      "grad_norm": 5.758388996124268,
      "learning_rate": 7.622277121246513e-07,
      "logits/chosen": 1.483445405960083,
      "logits/rejected": 1.5389055013656616,
      "logps/chosen": -38.62077331542969,
      "logps/rejected": -71.79267883300781,
      "loss": 0.4121,
      "nll_loss": 0.23168590664863586,
      "rewards/accuracies": 0.9587500095367432,
      "rewards/chosen": -3.8620777130126953,
      "rewards/margins": 3.3171918392181396,
      "rewards/rejected": -7.179268836975098,
      "step": 3150
    },
    {
      "epoch": 2.56,
      "grad_norm": 7.307446002960205,
      "learning_rate": 6.433598698998766e-07,
      "logits/chosen": 1.4688174724578857,
      "logits/rejected": 1.5306994915008545,
      "logps/chosen": -39.013912200927734,
      "logps/rejected": -71.06527709960938,
      "loss": 0.416,
      "nll_loss": 0.23500658571720123,
      "rewards/accuracies": 0.9424999952316284,
      "rewards/chosen": -3.9013912677764893,
      "rewards/margins": 3.2051360607147217,
      "rewards/rejected": -7.106527328491211,
      "step": 3200
    },
    {
      "epoch": 2.6,
      "grad_norm": 6.810940742492676,
      "learning_rate": 5.339275400731331e-07,
      "logits/chosen": 1.4731531143188477,
      "logits/rejected": 1.5175625085830688,
      "logps/chosen": -39.05414581298828,
      "logps/rejected": -69.38569641113281,
      "loss": 0.4207,
      "nll_loss": 0.2375660389661789,
      "rewards/accuracies": 0.949999988079071,
      "rewards/chosen": -3.9054150581359863,
      "rewards/margins": 3.0331552028656006,
      "rewards/rejected": -6.93856954574585,
      "step": 3250
    },
    {
      "epoch": 2.64,
      "grad_norm": 5.935739517211914,
      "learning_rate": 4.3416772870275295e-07,
      "logits/chosen": 1.447837233543396,
      "logits/rejected": 1.5085737705230713,
      "logps/chosen": -38.072574615478516,
      "logps/rejected": -71.93220520019531,
      "loss": 0.3981,
      "nll_loss": 0.22540073096752167,
      "rewards/accuracies": 0.9537500143051147,
      "rewards/chosen": -3.807257652282715,
      "rewards/margins": 3.385963201522827,
      "rewards/rejected": -7.193220138549805,
      "step": 3300
    },
    {
      "epoch": 2.68,
      "grad_norm": 6.746799468994141,
      "learning_rate": 3.442964933259474e-07,
      "logits/chosen": 1.4588629007339478,
      "logits/rejected": 1.517850637435913,
      "logps/chosen": -38.37602615356445,
      "logps/rejected": -72.57811737060547,
      "loss": 0.3998,
      "nll_loss": 0.23044270277023315,
      "rewards/accuracies": 0.9587500095367432,
      "rewards/chosen": -3.8376026153564453,
      "rewards/margins": 3.4202094078063965,
      "rewards/rejected": -7.257812023162842,
      "step": 3350
    },
    {
      "epoch": 2.7199999999999998,
      "grad_norm": 7.958431243896484,
      "learning_rate": 2.6450847502627883e-07,
      "logits/chosen": 1.4416861534118652,
      "logits/rejected": 1.5023283958435059,
      "logps/chosen": -39.2740364074707,
      "logps/rejected": -73.73572540283203,
      "loss": 0.4036,
      "nll_loss": 0.22917690873146057,
      "rewards/accuracies": 0.9662500023841858,
      "rewards/chosen": -3.927403450012207,
      "rewards/margins": 3.446169376373291,
      "rewards/rejected": -7.373574256896973,
      "step": 3400
    },
    {
      "epoch": 2.76,
      "grad_norm": 8.157621383666992,
      "learning_rate": 1.9497647688442478e-07,
      "logits/chosen": 1.4702584743499756,
      "logits/rejected": 1.5254310369491577,
      "logps/chosen": -38.0818977355957,
      "logps/rejected": -70.6051254272461,
      "loss": 0.4055,
      "nll_loss": 0.23146916925907135,
      "rewards/accuracies": 0.9549999833106995,
      "rewards/chosen": -3.808190107345581,
      "rewards/margins": 3.252322196960449,
      "rewards/rejected": -7.060512065887451,
      "step": 3450
    },
    {
      "epoch": 2.8,
      "grad_norm": 6.671606540679932,
      "learning_rate": 1.358510897251808e-07,
      "logits/chosen": 1.4628223180770874,
      "logits/rejected": 1.525000810623169,
      "logps/chosen": -37.097007751464844,
      "logps/rejected": -72.08160400390625,
      "loss": 0.3906,
      "nll_loss": 0.22892680764198303,
      "rewards/accuracies": 0.9700000286102295,
      "rewards/chosen": -3.7097012996673584,
      "rewards/margins": 3.498459577560425,
      "rewards/rejected": -7.208160400390625,
      "step": 3500
    },
    {
      "epoch": 2.84,
      "grad_norm": 8.076554298400879,
      "learning_rate": 8.726036597126619e-08,
      "logits/chosen": 1.470629096031189,
      "logits/rejected": 1.535922646522522,
      "logps/chosen": -36.950157165527344,
      "logps/rejected": -68.89533996582031,
      "loss": 0.395,
      "nll_loss": 0.22603529691696167,
      "rewards/accuracies": 0.9549999833106995,
      "rewards/chosen": -3.6950161457061768,
      "rewards/margins": 3.1945183277130127,
      "rewards/rejected": -6.8895344734191895,
      "step": 3550
    },
    {
      "epoch": 2.88,
      "grad_norm": 6.550802707672119,
      "learning_rate": 4.93095423102935e-08,
      "logits/chosen": 1.4719560146331787,
      "logits/rejected": 1.5248959064483643,
      "logps/chosen": -40.710941314697266,
      "logps/rejected": -74.42230987548828,
      "loss": 0.4132,
      "nll_loss": 0.24109821021556854,
      "rewards/accuracies": 0.9524999856948853,
      "rewards/chosen": -4.071094512939453,
      "rewards/margins": 3.371136426925659,
      "rewards/rejected": -7.442230701446533,
      "step": 3600
    },
    {
      "epoch": 2.92,
      "grad_norm": 6.743546485900879,
      "learning_rate": 2.2080811775535006e-08,
      "logits/chosen": 1.4744410514831543,
      "logits/rejected": 1.5339540243148804,
      "logps/chosen": -39.543033599853516,
      "logps/rejected": -74.6586685180664,
      "loss": 0.3964,
      "nll_loss": 0.23712339997291565,
      "rewards/accuracies": 0.9674999713897705,
      "rewards/chosen": -3.954303026199341,
      "rewards/margins": 3.511564254760742,
      "rewards/rejected": -7.46586799621582,
      "step": 3650
    },
    {
      "epoch": 2.96,
      "grad_norm": 11.845396041870117,
      "learning_rate": 5.633145734114665e-09,
      "logits/chosen": 1.4899775981903076,
      "logits/rejected": 1.533370018005371,
      "logps/chosen": -38.53838348388672,
      "logps/rejected": -71.89134979248047,
      "loss": 0.4058,
      "nll_loss": 0.23659667372703552,
      "rewards/accuracies": 0.9674999713897705,
      "rewards/chosen": -3.853839159011841,
      "rewards/margins": 3.3352959156036377,
      "rewards/rejected": -7.18913459777832,
      "step": 3700
    },
    {
      "epoch": 3.0,
      "grad_norm": 9.659139633178711,
      "learning_rate": 2.1661681620654963e-12,
      "logits/chosen": 1.4696215391159058,
      "logits/rejected": 1.526814579963684,
      "logps/chosen": -38.22508239746094,
      "logps/rejected": -72.54460144042969,
      "loss": 0.3938,
      "nll_loss": 0.22828102111816406,
      "rewards/accuracies": 0.9637500047683716,
      "rewards/chosen": -3.8225083351135254,
      "rewards/margins": 3.4319519996643066,
      "rewards/rejected": -7.254461288452148,
      "step": 3750
    }
  ],
  "logging_steps": 50,
  "max_steps": 3750,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 3,
  "save_steps": 200,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}