| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 20.0, | |
| "eval_steps": 500, | |
| "global_step": 580, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.3448275862068966, | |
| "grad_norm": 3.9107234477996826, | |
| "learning_rate": 8.620689655172415e-07, | |
| "logits/chosen": -2.363854169845581, | |
| "logits/rejected": -2.338671922683716, | |
| "logps/chosen": -76.8819351196289, | |
| "logps/rejected": -75.86869049072266, | |
| "loss": 0.6942, | |
| "rewards/accuracies": 0.375, | |
| "rewards/chosen": -0.003996999468654394, | |
| "rewards/margins": -0.004145228303968906, | |
| "rewards/rejected": 0.00014822949015069753, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.6896551724137931, | |
| "grad_norm": 4.328212738037109, | |
| "learning_rate": 1.724137931034483e-06, | |
| "logits/chosen": -2.3617663383483887, | |
| "logits/rejected": -2.3477540016174316, | |
| "logps/chosen": -93.76543426513672, | |
| "logps/rejected": -75.16656494140625, | |
| "loss": 0.6931, | |
| "rewards/accuracies": 0.48750001192092896, | |
| "rewards/chosen": 0.002653368515893817, | |
| "rewards/margins": 0.006352785974740982, | |
| "rewards/rejected": -0.003699416993185878, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 1.0344827586206897, | |
| "grad_norm": 3.473827838897705, | |
| "learning_rate": 2.5862068965517246e-06, | |
| "logits/chosen": -2.3354058265686035, | |
| "logits/rejected": -2.3330740928649902, | |
| "logps/chosen": -138.9960479736328, | |
| "logps/rejected": -67.00438690185547, | |
| "loss": 0.6927, | |
| "rewards/accuracies": 0.44999998807907104, | |
| "rewards/chosen": 0.01745142787694931, | |
| "rewards/margins": 0.013760591857135296, | |
| "rewards/rejected": 0.0036908381152898073, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 1.3793103448275863, | |
| "grad_norm": 3.9458820819854736, | |
| "learning_rate": 3.448275862068966e-06, | |
| "logits/chosen": -2.3478751182556152, | |
| "logits/rejected": -2.3700289726257324, | |
| "logps/chosen": -75.59983825683594, | |
| "logps/rejected": -67.20357513427734, | |
| "loss": 0.69, | |
| "rewards/accuracies": 0.5, | |
| "rewards/chosen": 0.00024204826331697404, | |
| "rewards/margins": 0.005866119172424078, | |
| "rewards/rejected": -0.005624071694910526, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.7241379310344827, | |
| "grad_norm": 4.299890041351318, | |
| "learning_rate": 4.310344827586207e-06, | |
| "logits/chosen": -2.338815212249756, | |
| "logits/rejected": -2.3532769680023193, | |
| "logps/chosen": -72.74342346191406, | |
| "logps/rejected": -87.63409423828125, | |
| "loss": 0.6899, | |
| "rewards/accuracies": 0.574999988079071, | |
| "rewards/chosen": -0.003249581903219223, | |
| "rewards/margins": 0.006888723932206631, | |
| "rewards/rejected": -0.010138307698071003, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 2.0689655172413794, | |
| "grad_norm": 4.957555294036865, | |
| "learning_rate": 4.999818897894192e-06, | |
| "logits/chosen": -2.344572067260742, | |
| "logits/rejected": -2.345888614654541, | |
| "logps/chosen": -80.16046142578125, | |
| "logps/rejected": -73.05986022949219, | |
| "loss": 0.6802, | |
| "rewards/accuracies": 0.637499988079071, | |
| "rewards/chosen": -0.013288321904838085, | |
| "rewards/margins": 0.020968889817595482, | |
| "rewards/rejected": -0.03425721079111099, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 2.413793103448276, | |
| "grad_norm": 4.212078094482422, | |
| "learning_rate": 4.9934830787948756e-06, | |
| "logits/chosen": -2.3330090045928955, | |
| "logits/rejected": -2.328575372695923, | |
| "logps/chosen": -73.63365173339844, | |
| "logps/rejected": -74.45356750488281, | |
| "loss": 0.6693, | |
| "rewards/accuracies": 0.7124999761581421, | |
| "rewards/chosen": -0.009874681942164898, | |
| "rewards/margins": 0.04785541445016861, | |
| "rewards/rejected": -0.05773010104894638, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 2.7586206896551726, | |
| "grad_norm": 4.446742534637451, | |
| "learning_rate": 4.978118375700895e-06, | |
| "logits/chosen": -2.3472495079040527, | |
| "logits/rejected": -2.3697409629821777, | |
| "logps/chosen": -73.44490814208984, | |
| "logps/rejected": -89.69837951660156, | |
| "loss": 0.6553, | |
| "rewards/accuracies": 0.8500000238418579, | |
| "rewards/chosen": -0.022817375138401985, | |
| "rewards/margins": 0.09045850485563278, | |
| "rewards/rejected": -0.11327588558197021, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 3.103448275862069, | |
| "grad_norm": 4.435715198516846, | |
| "learning_rate": 4.953780424089803e-06, | |
| "logits/chosen": -2.346717596054077, | |
| "logits/rejected": -2.3626608848571777, | |
| "logps/chosen": -85.64796447753906, | |
| "logps/rejected": -80.06268310546875, | |
| "loss": 0.6432, | |
| "rewards/accuracies": 0.737500011920929, | |
| "rewards/chosen": -0.0543200746178627, | |
| "rewards/margins": 0.1085677295923233, | |
| "rewards/rejected": -0.1628878116607666, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 3.4482758620689653, | |
| "grad_norm": 4.5200910568237305, | |
| "learning_rate": 4.920557351506409e-06, | |
| "logits/chosen": -2.358900547027588, | |
| "logits/rejected": -2.3591980934143066, | |
| "logps/chosen": -82.83473205566406, | |
| "logps/rejected": -80.66204071044922, | |
| "loss": 0.5996, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -0.046993009746074677, | |
| "rewards/margins": 0.20401433110237122, | |
| "rewards/rejected": -0.2510073184967041, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 3.793103448275862, | |
| "grad_norm": 4.106562614440918, | |
| "learning_rate": 4.878569458453592e-06, | |
| "logits/chosen": -2.347762107849121, | |
| "logits/rejected": -2.33540415763855, | |
| "logps/chosen": -72.23506164550781, | |
| "logps/rejected": -85.77088928222656, | |
| "loss": 0.6003, | |
| "rewards/accuracies": 0.824999988079071, | |
| "rewards/chosen": -0.051100969314575195, | |
| "rewards/margins": 0.24443332850933075, | |
| "rewards/rejected": -0.29553431272506714, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 4.137931034482759, | |
| "grad_norm": 4.67632532119751, | |
| "learning_rate": 4.827968782785062e-06, | |
| "logits/chosen": -2.353370428085327, | |
| "logits/rejected": -2.378964900970459, | |
| "logps/chosen": -67.91919708251953, | |
| "logps/rejected": -86.26390838623047, | |
| "loss": 0.5716, | |
| "rewards/accuracies": 0.7875000238418579, | |
| "rewards/chosen": -0.11952020972967148, | |
| "rewards/margins": 0.30119410157203674, | |
| "rewards/rejected": -0.42071428894996643, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 4.482758620689655, | |
| "grad_norm": 4.5131025314331055, | |
| "learning_rate": 4.7689385491773934e-06, | |
| "logits/chosen": -2.3820648193359375, | |
| "logits/rejected": -2.403965473175049, | |
| "logps/chosen": -71.66841125488281, | |
| "logps/rejected": -77.90869140625, | |
| "loss": 0.5362, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -0.0694696456193924, | |
| "rewards/margins": 0.47986412048339844, | |
| "rewards/rejected": -0.5493337512016296, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 4.827586206896552, | |
| "grad_norm": 4.446605205535889, | |
| "learning_rate": 4.70169250567482e-06, | |
| "logits/chosen": -2.372788906097412, | |
| "logits/rejected": -2.379612684249878, | |
| "logps/chosen": -69.44779968261719, | |
| "logps/rejected": -80.1120376586914, | |
| "loss": 0.5186, | |
| "rewards/accuracies": 0.862500011920929, | |
| "rewards/chosen": -0.12896183133125305, | |
| "rewards/margins": 0.5137210488319397, | |
| "rewards/rejected": -0.6426829099655151, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 5.172413793103448, | |
| "grad_norm": 4.605374813079834, | |
| "learning_rate": 4.626474149709127e-06, | |
| "logits/chosen": -2.3140041828155518, | |
| "logits/rejected": -2.336975574493408, | |
| "logps/chosen": -87.05641174316406, | |
| "logps/rejected": -94.53668212890625, | |
| "loss": 0.4794, | |
| "rewards/accuracies": 0.8374999761581421, | |
| "rewards/chosen": -0.15513856709003448, | |
| "rewards/margins": 0.6417296528816223, | |
| "rewards/rejected": -0.7968682050704956, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 5.517241379310345, | |
| "grad_norm": 6.027809143066406, | |
| "learning_rate": 4.54355584639723e-06, | |
| "logits/chosen": -2.365192413330078, | |
| "logits/rejected": -2.3872745037078857, | |
| "logps/chosen": -73.7781753540039, | |
| "logps/rejected": -82.1873550415039, | |
| "loss": 0.4599, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -0.25834327936172485, | |
| "rewards/margins": 0.5800828337669373, | |
| "rewards/rejected": -0.8384261131286621, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 5.862068965517241, | |
| "grad_norm": 5.196141242980957, | |
| "learning_rate": 4.45323784230908e-06, | |
| "logits/chosen": -2.391653299331665, | |
| "logits/rejected": -2.394854784011841, | |
| "logps/chosen": -72.33963775634766, | |
| "logps/rejected": -86.77699279785156, | |
| "loss": 0.4467, | |
| "rewards/accuracies": 0.8999999761581421, | |
| "rewards/chosen": -0.2543284595012665, | |
| "rewards/margins": 0.7806032299995422, | |
| "rewards/rejected": -1.0349315404891968, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 6.206896551724138, | |
| "grad_norm": 5.285089015960693, | |
| "learning_rate": 4.355847178277025e-06, | |
| "logits/chosen": -2.3918795585632324, | |
| "logits/rejected": -2.4050662517547607, | |
| "logps/chosen": -77.20832061767578, | |
| "logps/rejected": -98.5592041015625, | |
| "loss": 0.4052, | |
| "rewards/accuracies": 0.887499988079071, | |
| "rewards/chosen": -0.35020238161087036, | |
| "rewards/margins": 1.0169063806533813, | |
| "rewards/rejected": -1.367108702659607, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 6.551724137931035, | |
| "grad_norm": 5.308840751647949, | |
| "learning_rate": 4.2517365051833564e-06, | |
| "logits/chosen": -2.3778843879699707, | |
| "logits/rejected": -2.3865675926208496, | |
| "logps/chosen": -69.90904235839844, | |
| "logps/rejected": -86.66047668457031, | |
| "loss": 0.3809, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -0.42647725343704224, | |
| "rewards/margins": 0.9949015378952026, | |
| "rewards/rejected": -1.4213788509368896, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 6.896551724137931, | |
| "grad_norm": 6.095545768737793, | |
| "learning_rate": 4.141282807014034e-06, | |
| "logits/chosen": -2.407437801361084, | |
| "logits/rejected": -2.4296727180480957, | |
| "logps/chosen": -82.95701599121094, | |
| "logps/rejected": -83.4615478515625, | |
| "loss": 0.3742, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -0.6070786714553833, | |
| "rewards/margins": 0.968428909778595, | |
| "rewards/rejected": -1.5755075216293335, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 7.241379310344827, | |
| "grad_norm": 5.683840751647949, | |
| "learning_rate": 4.024886035802432e-06, | |
| "logits/chosen": -2.411830186843872, | |
| "logits/rejected": -2.4354729652404785, | |
| "logps/chosen": -61.66484451293945, | |
| "logps/rejected": -90.95805358886719, | |
| "loss": 0.3421, | |
| "rewards/accuracies": 0.8999999761581421, | |
| "rewards/chosen": -0.5454065203666687, | |
| "rewards/margins": 1.2011836767196655, | |
| "rewards/rejected": -1.7465900182724, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 7.586206896551724, | |
| "grad_norm": 5.584890842437744, | |
| "learning_rate": 3.9029676634059565e-06, | |
| "logits/chosen": -2.3816287517547607, | |
| "logits/rejected": -2.37958025932312, | |
| "logps/chosen": -87.12813568115234, | |
| "logps/rejected": -92.26463317871094, | |
| "loss": 0.3138, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -0.5636588335037231, | |
| "rewards/margins": 1.3944618701934814, | |
| "rewards/rejected": -1.9581207036972046, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 7.931034482758621, | |
| "grad_norm": 7.526435375213623, | |
| "learning_rate": 3.7759691553595214e-06, | |
| "logits/chosen": -2.399634599685669, | |
| "logits/rejected": -2.407938241958618, | |
| "logps/chosen": -82.55473327636719, | |
| "logps/rejected": -89.76951599121094, | |
| "loss": 0.3046, | |
| "rewards/accuracies": 0.9750000238418579, | |
| "rewards/chosen": -0.9372364282608032, | |
| "rewards/margins": 1.309622049331665, | |
| "rewards/rejected": -2.246858596801758, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 8.275862068965518, | |
| "grad_norm": 5.150575637817383, | |
| "learning_rate": 3.6443503723320837e-06, | |
| "logits/chosen": -2.36824893951416, | |
| "logits/rejected": -2.3758621215820312, | |
| "logps/chosen": -85.68852996826172, | |
| "logps/rejected": -102.09485626220703, | |
| "loss": 0.2638, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -0.9749407768249512, | |
| "rewards/margins": 1.6947702169418335, | |
| "rewards/rejected": -2.669710874557495, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 8.620689655172415, | |
| "grad_norm": 6.282366752624512, | |
| "learning_rate": 3.508587904974522e-06, | |
| "logits/chosen": -2.3659470081329346, | |
| "logits/rejected": -2.367462635040283, | |
| "logps/chosen": -87.00023651123047, | |
| "logps/rejected": -123.0126953125, | |
| "loss": 0.2405, | |
| "rewards/accuracies": 0.9624999761581421, | |
| "rewards/chosen": -1.026289939880371, | |
| "rewards/margins": 1.9990599155426025, | |
| "rewards/rejected": -3.0253496170043945, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 8.96551724137931, | |
| "grad_norm": 6.396641731262207, | |
| "learning_rate": 3.3691733481883693e-06, | |
| "logits/chosen": -2.3716306686401367, | |
| "logits/rejected": -2.3546042442321777, | |
| "logps/chosen": -87.29267883300781, | |
| "logps/rejected": -98.01856994628906, | |
| "loss": 0.2546, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -1.2485218048095703, | |
| "rewards/margins": 1.980831503868103, | |
| "rewards/rejected": -3.229353427886963, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 9.310344827586206, | |
| "grad_norm": 5.644590377807617, | |
| "learning_rate": 3.226611521064278e-06, | |
| "logits/chosen": -2.2916407585144043, | |
| "logits/rejected": -2.3004841804504395, | |
| "logps/chosen": -79.30825805664062, | |
| "logps/rejected": -113.4272689819336, | |
| "loss": 0.1999, | |
| "rewards/accuracies": 0.9750000238418579, | |
| "rewards/chosen": -1.1944876909255981, | |
| "rewards/margins": 2.365144729614258, | |
| "rewards/rejected": -3.5596320629119873, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 9.655172413793103, | |
| "grad_norm": 6.177416801452637, | |
| "learning_rate": 3.0814186389357765e-06, | |
| "logits/chosen": -2.346484661102295, | |
| "logits/rejected": -2.346285581588745, | |
| "logps/chosen": -77.01927947998047, | |
| "logps/rejected": -105.2061538696289, | |
| "loss": 0.1875, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -1.260752558708191, | |
| "rewards/margins": 2.249293565750122, | |
| "rewards/rejected": -3.5100464820861816, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 6.281085014343262, | |
| "learning_rate": 2.9341204441673267e-06, | |
| "logits/chosen": -2.3005917072296143, | |
| "logits/rejected": -2.3319716453552246, | |
| "logps/chosen": -92.97987365722656, | |
| "logps/rejected": -112.0716552734375, | |
| "loss": 0.2036, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -1.6851022243499756, | |
| "rewards/margins": 2.1508073806762695, | |
| "rewards/rejected": -3.835909605026245, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 10.344827586206897, | |
| "grad_norm": 5.364060401916504, | |
| "learning_rate": 2.785250302445062e-06, | |
| "logits/chosen": -2.3305928707122803, | |
| "logits/rejected": -2.3451201915740967, | |
| "logps/chosen": -85.11759185791016, | |
| "logps/rejected": -115.65193176269531, | |
| "loss": 0.1664, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -1.5494135618209839, | |
| "rewards/margins": 2.6376235485076904, | |
| "rewards/rejected": -4.187037467956543, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 10.689655172413794, | |
| "grad_norm": 5.540137767791748, | |
| "learning_rate": 2.6353472714635443e-06, | |
| "logits/chosen": -2.3217194080352783, | |
| "logits/rejected": -2.345512628555298, | |
| "logps/chosen": -83.56015014648438, | |
| "logps/rejected": -111.0335922241211, | |
| "loss": 0.1446, | |
| "rewards/accuracies": 0.9624999761581421, | |
| "rewards/chosen": -1.6171716451644897, | |
| "rewards/margins": 2.553410053253174, | |
| "rewards/rejected": -4.170581340789795, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 11.03448275862069, | |
| "grad_norm": 5.321618556976318, | |
| "learning_rate": 2.4849541490017868e-06, | |
| "logits/chosen": -2.2711265087127686, | |
| "logits/rejected": -2.274831533432007, | |
| "logps/chosen": -93.46397399902344, | |
| "logps/rejected": -138.03286743164062, | |
| "loss": 0.1416, | |
| "rewards/accuracies": 0.9624999761581421, | |
| "rewards/chosen": -2.0557093620300293, | |
| "rewards/margins": 2.809981346130371, | |
| "rewards/rejected": -4.8656907081604, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 11.379310344827585, | |
| "grad_norm": 6.168019771575928, | |
| "learning_rate": 2.3346155074564712e-06, | |
| "logits/chosen": -2.2561862468719482, | |
| "logits/rejected": -2.260565996170044, | |
| "logps/chosen": -89.53536224365234, | |
| "logps/rejected": -119.193603515625, | |
| "loss": 0.1174, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.2866904735565186, | |
| "rewards/margins": 2.866985321044922, | |
| "rewards/rejected": -5.1536760330200195, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 11.724137931034482, | |
| "grad_norm": 4.970952033996582, | |
| "learning_rate": 2.184875721949277e-06, | |
| "logits/chosen": -2.204287052154541, | |
| "logits/rejected": -2.2466628551483154, | |
| "logps/chosen": -171.76095581054688, | |
| "logps/rejected": -147.26907348632812, | |
| "loss": 0.1152, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.4709744453430176, | |
| "rewards/margins": 3.782508134841919, | |
| "rewards/rejected": -6.253481864929199, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 12.068965517241379, | |
| "grad_norm": 4.031501770019531, | |
| "learning_rate": 2.0362769991485514e-06, | |
| "logits/chosen": -2.245434522628784, | |
| "logits/rejected": -2.267470121383667, | |
| "logps/chosen": -98.18807220458984, | |
| "logps/rejected": -143.83383178710938, | |
| "loss": 0.1104, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.5792040824890137, | |
| "rewards/margins": 3.238723039627075, | |
| "rewards/rejected": -5.817927360534668, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 12.413793103448276, | |
| "grad_norm": 4.009598731994629, | |
| "learning_rate": 1.8893574139429226e-06, | |
| "logits/chosen": -2.2339296340942383, | |
| "logits/rejected": -2.254106283187866, | |
| "logps/chosen": -104.34129333496094, | |
| "logps/rejected": -130.74081420898438, | |
| "loss": 0.0859, | |
| "rewards/accuracies": 0.9750000238418579, | |
| "rewards/chosen": -2.8974556922912598, | |
| "rewards/margins": 3.275568723678589, | |
| "rewards/rejected": -6.1730241775512695, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 12.758620689655173, | |
| "grad_norm": 6.944422721862793, | |
| "learning_rate": 1.744648961076068e-06, | |
| "logits/chosen": -2.252500295639038, | |
| "logits/rejected": -2.25302791595459, | |
| "logps/chosen": -97.11503601074219, | |
| "logps/rejected": -135.61911010742188, | |
| "loss": 0.0901, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -2.814434051513672, | |
| "rewards/margins": 3.549976348876953, | |
| "rewards/rejected": -6.364409923553467, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 13.10344827586207, | |
| "grad_norm": 5.268442630767822, | |
| "learning_rate": 1.602675628797636e-06, | |
| "logits/chosen": -2.2403993606567383, | |
| "logits/rejected": -2.240546703338623, | |
| "logps/chosen": -101.62408447265625, | |
| "logps/rejected": -136.84268188476562, | |
| "loss": 0.0902, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.9260125160217285, | |
| "rewards/margins": 3.421342134475708, | |
| "rewards/rejected": -6.347354888916016, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 13.448275862068966, | |
| "grad_norm": 4.1959404945373535, | |
| "learning_rate": 1.4639515015056205e-06, | |
| "logits/chosen": -2.2759053707122803, | |
| "logits/rejected": -2.2657554149627686, | |
| "logps/chosen": -103.3962631225586, | |
| "logps/rejected": -138.56057739257812, | |
| "loss": 0.0729, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -2.688096523284912, | |
| "rewards/margins": 4.030136585235596, | |
| "rewards/rejected": -6.718233585357666, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 13.793103448275861, | |
| "grad_norm": 4.1255574226379395, | |
| "learning_rate": 1.328978898250525e-06, | |
| "logits/chosen": -2.180371046066284, | |
| "logits/rejected": -2.213881015777588, | |
| "logps/chosen": -100.98771667480469, | |
| "logps/rejected": -135.07044982910156, | |
| "loss": 0.0747, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -3.0276432037353516, | |
| "rewards/margins": 3.8850128650665283, | |
| "rewards/rejected": -6.912655830383301, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 14.137931034482758, | |
| "grad_norm": 5.081278324127197, | |
| "learning_rate": 1.198246553841744e-06, | |
| "logits/chosen": -2.237619161605835, | |
| "logits/rejected": -2.219052314758301, | |
| "logps/chosen": -98.90690612792969, | |
| "logps/rejected": -131.66647338867188, | |
| "loss": 0.0576, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -3.216820478439331, | |
| "rewards/margins": 3.5623440742492676, | |
| "rewards/rejected": -6.779164791107178, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 14.482758620689655, | |
| "grad_norm": 4.580889701843262, | |
| "learning_rate": 1.0722278491423998e-06, | |
| "logits/chosen": -2.1594581604003906, | |
| "logits/rejected": -2.183774471282959, | |
| "logps/chosen": -106.17756652832031, | |
| "logps/rejected": -163.53775024414062, | |
| "loss": 0.0545, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -3.6679279804229736, | |
| "rewards/margins": 4.358443260192871, | |
| "rewards/rejected": -8.026371955871582, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 14.827586206896552, | |
| "grad_norm": 7.627413749694824, | |
| "learning_rate": 9.513790969606926e-07, | |
| "logits/chosen": -2.1865813732147217, | |
| "logits/rejected": -2.2028133869171143, | |
| "logps/chosen": -109.03575134277344, | |
| "logps/rejected": -147.6363983154297, | |
| "loss": 0.0552, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -3.6653411388397217, | |
| "rewards/margins": 3.91802978515625, | |
| "rewards/rejected": -7.583371162414551, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 15.172413793103448, | |
| "grad_norm": 4.102783679962158, | |
| "learning_rate": 8.361378897445643e-07, | |
| "logits/chosen": -2.18919038772583, | |
| "logits/rejected": -2.2041501998901367, | |
| "logps/chosen": -110.42228698730469, | |
| "logps/rejected": -157.49429321289062, | |
| "loss": 0.0596, | |
| "rewards/accuracies": 0.9624999761581421, | |
| "rewards/chosen": -3.86995005607605, | |
| "rewards/margins": 4.187633037567139, | |
| "rewards/rejected": -8.057581901550293, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 15.517241379310345, | |
| "grad_norm": 4.329521656036377, | |
| "learning_rate": 7.269215150626391e-07, | |
| "logits/chosen": -2.1742970943450928, | |
| "logits/rejected": -2.1848294734954834, | |
| "logps/chosen": -107.87260437011719, | |
| "logps/rejected": -150.42959594726562, | |
| "loss": 0.047, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -3.9351372718811035, | |
| "rewards/margins": 4.330621242523193, | |
| "rewards/rejected": -8.26575756072998, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 15.862068965517242, | |
| "grad_norm": 4.850955963134766, | |
| "learning_rate": 6.241254446089942e-07, | |
| "logits/chosen": -2.2147865295410156, | |
| "logits/rejected": -2.2335610389709473, | |
| "logps/chosen": -105.22979736328125, | |
| "logps/rejected": -149.566162109375, | |
| "loss": 0.0443, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -3.2662596702575684, | |
| "rewards/margins": 4.226635932922363, | |
| "rewards/rejected": -7.49289608001709, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 16.20689655172414, | |
| "grad_norm": 3.0453879833221436, | |
| "learning_rate": 5.281219022030423e-07, | |
| "logits/chosen": -2.158506155014038, | |
| "logits/rejected": -2.1904215812683105, | |
| "logps/chosen": -119.72633361816406, | |
| "logps/rejected": -159.1144561767578, | |
| "loss": 0.0409, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -3.4267520904541016, | |
| "rewards/margins": 4.899367332458496, | |
| "rewards/rejected": -8.326119422912598, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 16.551724137931036, | |
| "grad_norm": 3.7533819675445557, | |
| "learning_rate": 4.392585159698087e-07, | |
| "logits/chosen": -2.152306318283081, | |
| "logits/rejected": -2.1774165630340576, | |
| "logps/chosen": -112.7719955444336, | |
| "logps/rejected": -165.19509887695312, | |
| "loss": 0.0416, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.251755714416504, | |
| "rewards/margins": 4.445396900177002, | |
| "rewards/rejected": -8.697153091430664, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 16.896551724137932, | |
| "grad_norm": 7.166266918182373, | |
| "learning_rate": 3.578570595810274e-07, | |
| "logits/chosen": -2.1818134784698486, | |
| "logits/rejected": -2.2025341987609863, | |
| "logps/chosen": -96.93048095703125, | |
| "logps/rejected": -171.3131103515625, | |
| "loss": 0.0417, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.020156383514404, | |
| "rewards/margins": 5.221116542816162, | |
| "rewards/rejected": -9.241272926330566, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 17.24137931034483, | |
| "grad_norm": 3.491783618927002, | |
| "learning_rate": 2.8421228711503127e-07, | |
| "logits/chosen": -2.1831130981445312, | |
| "logits/rejected": -2.1822221279144287, | |
| "logps/chosen": -113.05125427246094, | |
| "logps/rejected": -171.09835815429688, | |
| "loss": 0.0369, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -4.272199630737305, | |
| "rewards/margins": 4.791441440582275, | |
| "rewards/rejected": -9.063641548156738, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 17.586206896551722, | |
| "grad_norm": 2.9002482891082764, | |
| "learning_rate": 2.1859086575439225e-07, | |
| "logits/chosen": -2.211790084838867, | |
| "logits/rejected": -2.1946873664855957, | |
| "logps/chosen": -129.7812042236328, | |
| "logps/rejected": -168.2603302001953, | |
| "loss": 0.0359, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -4.544430732727051, | |
| "rewards/margins": 4.710877418518066, | |
| "rewards/rejected": -9.2553071975708, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 17.93103448275862, | |
| "grad_norm": 3.7184576988220215, | |
| "learning_rate": 1.6123041018599766e-07, | |
| "logits/chosen": -2.1540164947509766, | |
| "logits/rejected": -2.147338390350342, | |
| "logps/chosen": -117.89579772949219, | |
| "logps/rejected": -158.124755859375, | |
| "loss": 0.0413, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -4.149048328399658, | |
| "rewards/margins": 4.793323516845703, | |
| "rewards/rejected": -8.94237232208252, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 18.275862068965516, | |
| "grad_norm": 3.723693370819092, | |
| "learning_rate": 1.1233862220001168e-07, | |
| "logits/chosen": -2.169210195541382, | |
| "logits/rejected": -2.184584140777588, | |
| "logps/chosen": -107.45458984375, | |
| "logps/rejected": -161.4453582763672, | |
| "loss": 0.0356, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -3.943101167678833, | |
| "rewards/margins": 5.099206447601318, | |
| "rewards/rejected": -9.04230785369873, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 18.620689655172413, | |
| "grad_norm": 2.663479804992676, | |
| "learning_rate": 7.209253860320897e-08, | |
| "logits/chosen": -2.141265392303467, | |
| "logits/rejected": -2.1849865913391113, | |
| "logps/chosen": -155.58615112304688, | |
| "logps/rejected": -161.23976135253906, | |
| "loss": 0.0346, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -2.961081027984619, | |
| "rewards/margins": 6.033589839935303, | |
| "rewards/rejected": -8.994670867919922, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 18.96551724137931, | |
| "grad_norm": 3.3965961933135986, | |
| "learning_rate": 4.063789016999331e-08, | |
| "logits/chosen": -2.2001423835754395, | |
| "logits/rejected": -2.214458703994751, | |
| "logps/chosen": -128.25711059570312, | |
| "logps/rejected": -166.47296142578125, | |
| "loss": 0.0394, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.388108730316162, | |
| "rewards/margins": 4.831867694854736, | |
| "rewards/rejected": -9.219977378845215, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 19.310344827586206, | |
| "grad_norm": 3.214250326156616, | |
| "learning_rate": 1.808857395232788e-08, | |
| "logits/chosen": -2.147885799407959, | |
| "logits/rejected": -2.1538643836975098, | |
| "logps/chosen": -119.9779281616211, | |
| "logps/rejected": -163.17922973632812, | |
| "loss": 0.0352, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -4.499241828918457, | |
| "rewards/margins": 4.586289882659912, | |
| "rewards/rejected": -9.085531234741211, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 19.655172413793103, | |
| "grad_norm": 3.0123510360717773, | |
| "learning_rate": 4.526240859345499e-09, | |
| "logits/chosen": -2.1862876415252686, | |
| "logits/rejected": -2.1756527423858643, | |
| "logps/chosen": -135.414306640625, | |
| "logps/rejected": -169.79623413085938, | |
| "loss": 0.0344, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.599732398986816, | |
| "rewards/margins": 4.528085231781006, | |
| "rewards/rejected": -9.12781810760498, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 2.647522449493408, | |
| "learning_rate": 0.0, | |
| "logits/chosen": -2.175386428833008, | |
| "logits/rejected": -2.183178424835205, | |
| "logps/chosen": -106.87957763671875, | |
| "logps/rejected": -163.1853790283203, | |
| "loss": 0.036, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -4.439215183258057, | |
| "rewards/margins": 4.813824653625488, | |
| "rewards/rejected": -9.253039360046387, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "step": 580, | |
| "total_flos": 2.1130760508114207e+18, | |
| "train_loss": 0.27193391800954425, | |
| "train_runtime": 4447.2543, | |
| "train_samples_per_second": 8.324, | |
| "train_steps_per_second": 0.13 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 580, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 20, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.1130760508114207e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |