| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 14.479638009049774, | |
| "eval_steps": 500, | |
| "global_step": 400, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.36199095022624433, | |
| "grad_norm": 4.351782321929932, | |
| "learning_rate": 9.259259259259259e-07, | |
| "logits/chosen": -2.3496108055114746, | |
| "logits/rejected": -2.354576349258423, | |
| "logps/chosen": -86.63139343261719, | |
| "logps/rejected": -76.90333557128906, | |
| "loss": 0.6923, | |
| "rewards/accuracies": 0.42500001192092896, | |
| "rewards/chosen": -0.00019822348258458078, | |
| "rewards/margins": -0.002337233629077673, | |
| "rewards/rejected": 0.00213901000097394, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.7239819004524887, | |
| "grad_norm": 3.8085336685180664, | |
| "learning_rate": 1.8518518518518519e-06, | |
| "logits/chosen": -2.3482470512390137, | |
| "logits/rejected": -2.3420329093933105, | |
| "logps/chosen": -73.18113708496094, | |
| "logps/rejected": -72.52565002441406, | |
| "loss": 0.6943, | |
| "rewards/accuracies": 0.4000000059604645, | |
| "rewards/chosen": 0.00011640912998700514, | |
| "rewards/margins": -0.002153881825506687, | |
| "rewards/rejected": 0.002270291093736887, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 1.085972850678733, | |
| "grad_norm": 5.329021453857422, | |
| "learning_rate": 2.7777777777777783e-06, | |
| "logits/chosen": -2.3689351081848145, | |
| "logits/rejected": -2.3885178565979004, | |
| "logps/chosen": -72.15180969238281, | |
| "logps/rejected": -74.70518493652344, | |
| "loss": 0.6924, | |
| "rewards/accuracies": 0.5625, | |
| "rewards/chosen": 0.0020837264601141214, | |
| "rewards/margins": 0.00603193324059248, | |
| "rewards/rejected": -0.0039482079446315765, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 1.4479638009049773, | |
| "grad_norm": 3.671722888946533, | |
| "learning_rate": 3.7037037037037037e-06, | |
| "logits/chosen": -2.3091399669647217, | |
| "logits/rejected": -2.3273236751556396, | |
| "logps/chosen": -69.45216369628906, | |
| "logps/rejected": -260.0699157714844, | |
| "loss": 0.6899, | |
| "rewards/accuracies": 0.550000011920929, | |
| "rewards/chosen": -0.0094170942902565, | |
| "rewards/margins": 0.14604035019874573, | |
| "rewards/rejected": -0.15545745193958282, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.8099547511312217, | |
| "grad_norm": 4.172375679016113, | |
| "learning_rate": 4.62962962962963e-06, | |
| "logits/chosen": -2.31831693649292, | |
| "logits/rejected": -2.335618495941162, | |
| "logps/chosen": -72.82856750488281, | |
| "logps/rejected": -90.61243438720703, | |
| "loss": 0.6865, | |
| "rewards/accuracies": 0.625, | |
| "rewards/chosen": -0.01846747286617756, | |
| "rewards/margins": 0.015672579407691956, | |
| "rewards/rejected": -0.034140050411224365, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 2.171945701357466, | |
| "grad_norm": 4.386168479919434, | |
| "learning_rate": 4.998119881260576e-06, | |
| "logits/chosen": -2.3203041553497314, | |
| "logits/rejected": -2.3315658569335938, | |
| "logps/chosen": -67.68244934082031, | |
| "logps/rejected": -86.2356185913086, | |
| "loss": 0.6775, | |
| "rewards/accuracies": 0.637499988079071, | |
| "rewards/chosen": -0.03993881493806839, | |
| "rewards/margins": 0.027381544932723045, | |
| "rewards/rejected": -0.06732036173343658, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 2.5339366515837103, | |
| "grad_norm": 4.38440465927124, | |
| "learning_rate": 4.9866405060165044e-06, | |
| "logits/chosen": -2.3607311248779297, | |
| "logits/rejected": -2.3799984455108643, | |
| "logps/chosen": -72.8117904663086, | |
| "logps/rejected": -79.02535247802734, | |
| "loss": 0.6636, | |
| "rewards/accuracies": 0.6875, | |
| "rewards/chosen": -0.06544635444879532, | |
| "rewards/margins": 0.069185271859169, | |
| "rewards/rejected": -0.13463163375854492, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 2.8959276018099547, | |
| "grad_norm": 4.333106517791748, | |
| "learning_rate": 4.964774158361991e-06, | |
| "logits/chosen": -2.392198324203491, | |
| "logits/rejected": -2.3854451179504395, | |
| "logps/chosen": -64.03775787353516, | |
| "logps/rejected": -71.01383972167969, | |
| "loss": 0.6473, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.09924488514661789, | |
| "rewards/margins": 0.1070803552865982, | |
| "rewards/rejected": -0.2063252478837967, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 3.257918552036199, | |
| "grad_norm": 4.575820446014404, | |
| "learning_rate": 4.93261217644956e-06, | |
| "logits/chosen": -2.3316967487335205, | |
| "logits/rejected": -2.3399548530578613, | |
| "logps/chosen": -66.9199447631836, | |
| "logps/rejected": -67.97264099121094, | |
| "loss": 0.6292, | |
| "rewards/accuracies": 0.675000011920929, | |
| "rewards/chosen": -0.1342274248600006, | |
| "rewards/margins": 0.1109505444765091, | |
| "rewards/rejected": -0.2451779842376709, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 3.6199095022624435, | |
| "grad_norm": 4.666302680969238, | |
| "learning_rate": 4.8902889044347e-06, | |
| "logits/chosen": -2.3398795127868652, | |
| "logits/rejected": -2.3582355976104736, | |
| "logps/chosen": -76.05484771728516, | |
| "logps/rejected": -78.4997787475586, | |
| "loss": 0.6015, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.15528908371925354, | |
| "rewards/margins": 0.17627069354057312, | |
| "rewards/rejected": -0.33155977725982666, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 3.981900452488688, | |
| "grad_norm": 4.472248554229736, | |
| "learning_rate": 4.837981131305475e-06, | |
| "logits/chosen": -2.3747470378875732, | |
| "logits/rejected": -2.3589282035827637, | |
| "logps/chosen": -75.08245086669922, | |
| "logps/rejected": -88.42532348632812, | |
| "loss": 0.582, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": -0.22706560790538788, | |
| "rewards/margins": 0.2718750834465027, | |
| "rewards/rejected": -0.49894070625305176, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 4.343891402714932, | |
| "grad_norm": 4.420746326446533, | |
| "learning_rate": 4.775907352415367e-06, | |
| "logits/chosen": -2.3004438877105713, | |
| "logits/rejected": -2.33561635017395, | |
| "logps/chosen": -92.03379821777344, | |
| "logps/rejected": -84.3415298461914, | |
| "loss": 0.5537, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -0.2706041932106018, | |
| "rewards/margins": 0.2943420112133026, | |
| "rewards/rejected": -0.564946174621582, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 4.705882352941177, | |
| "grad_norm": 4.54348611831665, | |
| "learning_rate": 4.70432685680402e-06, | |
| "logits/chosen": -2.4000728130340576, | |
| "logits/rejected": -2.3912758827209473, | |
| "logps/chosen": -70.9849624633789, | |
| "logps/rejected": -79.27720642089844, | |
| "loss": 0.5306, | |
| "rewards/accuracies": 0.800000011920929, | |
| "rewards/chosen": -0.2949373126029968, | |
| "rewards/margins": 0.4063098430633545, | |
| "rewards/rejected": -0.7012470960617065, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 5.067873303167421, | |
| "grad_norm": 4.43679141998291, | |
| "learning_rate": 4.623538644118244e-06, | |
| "logits/chosen": -2.3637630939483643, | |
| "logits/rejected": -2.3869171142578125, | |
| "logps/chosen": -63.164039611816406, | |
| "logps/rejected": -78.0440673828125, | |
| "loss": 0.5028, | |
| "rewards/accuracies": 0.824999988079071, | |
| "rewards/chosen": -0.2740551233291626, | |
| "rewards/margins": 0.5061118006706238, | |
| "rewards/rejected": -0.7801669836044312, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 5.429864253393665, | |
| "grad_norm": 4.934199810028076, | |
| "learning_rate": 4.533880175657419e-06, | |
| "logits/chosen": -2.342858076095581, | |
| "logits/rejected": -2.3417344093322754, | |
| "logps/chosen": -80.54903411865234, | |
| "logps/rejected": -79.4267807006836, | |
| "loss": 0.4663, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -0.3702576756477356, | |
| "rewards/margins": 0.5886551141738892, | |
| "rewards/rejected": -0.95891273021698, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 5.791855203619909, | |
| "grad_norm": 5.551158905029297, | |
| "learning_rate": 4.435725964760331e-06, | |
| "logits/chosen": -2.36419415473938, | |
| "logits/rejected": -2.388049364089966, | |
| "logps/chosen": -69.40000915527344, | |
| "logps/rejected": -91.36970520019531, | |
| "loss": 0.4636, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -0.42509594559669495, | |
| "rewards/margins": 0.7187294363975525, | |
| "rewards/rejected": -1.1438252925872803, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 6.153846153846154, | |
| "grad_norm": 4.6863555908203125, | |
| "learning_rate": 4.329486012421531e-06, | |
| "logits/chosen": -2.3798611164093018, | |
| "logits/rejected": -2.388327121734619, | |
| "logps/chosen": -84.27762603759766, | |
| "logps/rejected": -95.46641540527344, | |
| "loss": 0.4245, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -0.4049956202507019, | |
| "rewards/margins": 0.965179443359375, | |
| "rewards/rejected": -1.3701750040054321, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 6.515837104072398, | |
| "grad_norm": 5.001020431518555, | |
| "learning_rate": 4.215604094671835e-06, | |
| "logits/chosen": -2.376713991165161, | |
| "logits/rejected": -2.413689374923706, | |
| "logps/chosen": -75.96709442138672, | |
| "logps/rejected": -88.12495422363281, | |
| "loss": 0.449, | |
| "rewards/accuracies": 0.9125000238418579, | |
| "rewards/chosen": -0.3547631800174713, | |
| "rewards/margins": 0.8548161387443542, | |
| "rewards/rejected": -1.2095792293548584, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 6.877828054298643, | |
| "grad_norm": 5.340312957763672, | |
| "learning_rate": 4.094555908876765e-06, | |
| "logits/chosen": -2.385936737060547, | |
| "logits/rejected": -2.4080803394317627, | |
| "logps/chosen": -66.79461669921875, | |
| "logps/rejected": -85.04891204833984, | |
| "loss": 0.3764, | |
| "rewards/accuracies": 0.925000011920929, | |
| "rewards/chosen": -0.4008978307247162, | |
| "rewards/margins": 0.8593969345092773, | |
| "rewards/rejected": -1.2602946758270264, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 7.239819004524887, | |
| "grad_norm": 4.7585883140563965, | |
| "learning_rate": 3.966847086696045e-06, | |
| "logits/chosen": -2.4034409523010254, | |
| "logits/rejected": -2.425673246383667, | |
| "logps/chosen": -78.95824432373047, | |
| "logps/rejected": -92.42652893066406, | |
| "loss": 0.3362, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -0.5016047358512878, | |
| "rewards/margins": 1.3585138320922852, | |
| "rewards/rejected": -1.8601186275482178, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 7.601809954751131, | |
| "grad_norm": 5.299131870269775, | |
| "learning_rate": 3.833011082004229e-06, | |
| "logits/chosen": -2.405515193939209, | |
| "logits/rejected": -2.4070794582366943, | |
| "logps/chosen": -70.76835632324219, | |
| "logps/rejected": -94.3203353881836, | |
| "loss": 0.317, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -0.7332154512405396, | |
| "rewards/margins": 1.352203130722046, | |
| "rewards/rejected": -2.085418701171875, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 7.963800904977376, | |
| "grad_norm": 6.385330677032471, | |
| "learning_rate": 3.693606942594873e-06, | |
| "logits/chosen": -2.3782029151916504, | |
| "logits/rejected": -2.3890140056610107, | |
| "logps/chosen": -74.07151794433594, | |
| "logps/rejected": -97.96099853515625, | |
| "loss": 0.3338, | |
| "rewards/accuracies": 0.875, | |
| "rewards/chosen": -0.8589820861816406, | |
| "rewards/margins": 1.498375415802002, | |
| "rewards/rejected": -2.3573575019836426, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 8.32579185520362, | |
| "grad_norm": 5.3784027099609375, | |
| "learning_rate": 3.549216974976073e-06, | |
| "logits/chosen": -2.411106586456299, | |
| "logits/rejected": -2.4102063179016113, | |
| "logps/chosen": -78.97357940673828, | |
| "logps/rejected": -104.22176361083984, | |
| "loss": 0.2894, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": -0.850872814655304, | |
| "rewards/margins": 1.5754810571670532, | |
| "rewards/rejected": -2.426353931427002, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 8.687782805429864, | |
| "grad_norm": 5.306555271148682, | |
| "learning_rate": 3.400444312011776e-06, | |
| "logits/chosen": -2.3758366107940674, | |
| "logits/rejected": -2.393923282623291, | |
| "logps/chosen": -75.22766876220703, | |
| "logps/rejected": -105.24629211425781, | |
| "loss": 0.2666, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -0.6813958287239075, | |
| "rewards/margins": 1.697392225265503, | |
| "rewards/rejected": -2.3787879943847656, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 9.049773755656108, | |
| "grad_norm": 5.835656642913818, | |
| "learning_rate": 3.2479103935691047e-06, | |
| "logits/chosen": -2.345181465148926, | |
| "logits/rejected": -2.3525192737579346, | |
| "logps/chosen": -99.5466537475586, | |
| "logps/rejected": -115.3326644897461, | |
| "loss": 0.2556, | |
| "rewards/accuracies": 0.9125000238418579, | |
| "rewards/chosen": -1.1117511987686157, | |
| "rewards/margins": 1.774484634399414, | |
| "rewards/rejected": -2.8862357139587402, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 9.411764705882353, | |
| "grad_norm": 5.279304504394531, | |
| "learning_rate": 3.092252370695298e-06, | |
| "logits/chosen": -2.3707776069641113, | |
| "logits/rejected": -2.3579325675964355, | |
| "logps/chosen": -107.48176574707031, | |
| "logps/rejected": -119.63212585449219, | |
| "loss": 0.223, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -1.34574556350708, | |
| "rewards/margins": 2.105149745941162, | |
| "rewards/rejected": -3.450894832611084, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 9.773755656108598, | |
| "grad_norm": 6.106879711151123, | |
| "learning_rate": 2.9341204441673267e-06, | |
| "logits/chosen": -2.367492198944092, | |
| "logits/rejected": -2.3800530433654785, | |
| "logps/chosen": -94.10942077636719, | |
| "logps/rejected": -108.04286193847656, | |
| "loss": 0.2108, | |
| "rewards/accuracies": 0.9624999761581421, | |
| "rewards/chosen": -1.319504976272583, | |
| "rewards/margins": 2.0621278285980225, | |
| "rewards/rejected": -3.3816330432891846, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 10.135746606334841, | |
| "grad_norm": 4.824020862579346, | |
| "learning_rate": 2.7741751485313295e-06, | |
| "logits/chosen": -2.29042649269104, | |
| "logits/rejected": -2.2942726612091064, | |
| "logps/chosen": -88.15667724609375, | |
| "logps/rejected": -116.263671875, | |
| "loss": 0.2067, | |
| "rewards/accuracies": 0.9624999761581421, | |
| "rewards/chosen": -1.352482795715332, | |
| "rewards/margins": 2.1809113025665283, | |
| "rewards/rejected": -3.5333938598632812, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 10.497737556561086, | |
| "grad_norm": 5.073860168457031, | |
| "learning_rate": 2.6130845929767662e-06, | |
| "logits/chosen": -2.319164276123047, | |
| "logits/rejected": -2.327695846557617, | |
| "logps/chosen": -89.06367492675781, | |
| "logps/rejected": -113.19393157958984, | |
| "loss": 0.1791, | |
| "rewards/accuracies": 0.9750000238418579, | |
| "rewards/chosen": -1.4301341772079468, | |
| "rewards/margins": 2.2849090099334717, | |
| "rewards/rejected": -3.715043306350708, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 10.85972850678733, | |
| "grad_norm": 6.990862846374512, | |
| "learning_rate": 2.4515216705704396e-06, | |
| "logits/chosen": -2.322035551071167, | |
| "logits/rejected": -2.343055248260498, | |
| "logps/chosen": -79.6242904663086, | |
| "logps/rejected": -117.89810943603516, | |
| "loss": 0.173, | |
| "rewards/accuracies": 0.949999988079071, | |
| "rewards/chosen": -1.5006747245788574, | |
| "rewards/margins": 2.6087098121643066, | |
| "rewards/rejected": -4.109384059906006, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 11.221719457013574, | |
| "grad_norm": 4.418057918548584, | |
| "learning_rate": 2.290161247507733e-06, | |
| "logits/chosen": -2.313563585281372, | |
| "logits/rejected": -2.3379547595977783, | |
| "logps/chosen": -96.51193237304688, | |
| "logps/rejected": -128.72109985351562, | |
| "loss": 0.1525, | |
| "rewards/accuracies": 0.9750000238418579, | |
| "rewards/chosen": -1.7700340747833252, | |
| "rewards/margins": 2.5669429302215576, | |
| "rewards/rejected": -4.336977005004883, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 11.583710407239819, | |
| "grad_norm": 5.5379638671875, | |
| "learning_rate": 2.129677344121879e-06, | |
| "logits/chosen": -2.3077690601348877, | |
| "logits/rejected": -2.3094325065612793, | |
| "logps/chosen": -97.48485565185547, | |
| "logps/rejected": -120.9418716430664, | |
| "loss": 0.1367, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -1.6908330917358398, | |
| "rewards/margins": 2.76543927192688, | |
| "rewards/rejected": -4.456272602081299, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 11.945701357466064, | |
| "grad_norm": 8.315099716186523, | |
| "learning_rate": 1.970740319426474e-06, | |
| "logits/chosen": -2.2925028800964355, | |
| "logits/rejected": -2.31669545173645, | |
| "logps/chosen": -88.35062408447266, | |
| "logps/rejected": -125.6523208618164, | |
| "loss": 0.1436, | |
| "rewards/accuracies": 0.9750000238418579, | |
| "rewards/chosen": -2.12321400642395, | |
| "rewards/margins": 2.9592180252075195, | |
| "rewards/rejected": -5.082431316375732, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 12.307692307692308, | |
| "grad_norm": 5.883168697357178, | |
| "learning_rate": 1.8140140709517467e-06, | |
| "logits/chosen": -2.2664735317230225, | |
| "logits/rejected": -2.3047263622283936, | |
| "logps/chosen": -87.54987335205078, | |
| "logps/rejected": -120.87919616699219, | |
| "loss": 0.1185, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -1.9975258111953735, | |
| "rewards/margins": 3.0641627311706543, | |
| "rewards/rejected": -5.061688423156738, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 12.669683257918551, | |
| "grad_norm": 4.8013458251953125, | |
| "learning_rate": 1.6601532615711452e-06, | |
| "logits/chosen": -2.2390546798706055, | |
| "logits/rejected": -2.233534097671509, | |
| "logps/chosen": -102.70186614990234, | |
| "logps/rejected": -136.1205291748047, | |
| "loss": 0.111, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.10892915725708, | |
| "rewards/margins": 3.2926182746887207, | |
| "rewards/rejected": -5.401547431945801, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 13.031674208144796, | |
| "grad_norm": 4.820348739624023, | |
| "learning_rate": 1.509800584902108e-06, | |
| "logits/chosen": -2.2781853675842285, | |
| "logits/rejected": -2.3037526607513428, | |
| "logps/chosen": -94.27687072753906, | |
| "logps/rejected": -143.52162170410156, | |
| "loss": 0.1133, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.185966730117798, | |
| "rewards/margins": 3.149972438812256, | |
| "rewards/rejected": -5.335939884185791, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 13.393665158371041, | |
| "grad_norm": 4.292584419250488, | |
| "learning_rate": 1.3635840807037487e-06, | |
| "logits/chosen": -2.260101079940796, | |
| "logits/rejected": -2.270853042602539, | |
| "logps/chosen": -112.0027847290039, | |
| "logps/rejected": -155.92738342285156, | |
| "loss": 0.0866, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": -2.6447501182556152, | |
| "rewards/margins": 3.6627235412597656, | |
| "rewards/rejected": -6.307473659515381, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 13.755656108597286, | |
| "grad_norm": 4.569724082946777, | |
| "learning_rate": 1.2221145114853172e-06, | |
| "logits/chosen": -2.2924914360046387, | |
| "logits/rejected": -2.3221309185028076, | |
| "logps/chosen": -104.69401550292969, | |
| "logps/rejected": -138.05264282226562, | |
| "loss": 0.0953, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.8196561336517334, | |
| "rewards/margins": 3.5180962085723877, | |
| "rewards/rejected": -6.337752342224121, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 14.117647058823529, | |
| "grad_norm": 4.538058280944824, | |
| "learning_rate": 1.085982811283654e-06, | |
| "logits/chosen": -2.2493948936462402, | |
| "logits/rejected": -2.2872321605682373, | |
| "logps/chosen": -102.8060531616211, | |
| "logps/rejected": -135.58599853515625, | |
| "loss": 0.0879, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.9318606853485107, | |
| "rewards/margins": 3.4618797302246094, | |
| "rewards/rejected": -6.393740177154541, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 14.479638009049774, | |
| "grad_norm": 4.173523426055908, | |
| "learning_rate": 9.557576172663577e-07, | |
| "logits/chosen": -2.246553659439087, | |
| "logits/rejected": -2.2457058429718018, | |
| "logps/chosen": -109.64227294921875, | |
| "logps/rejected": -147.6108856201172, | |
| "loss": 0.0783, | |
| "rewards/accuracies": 0.987500011920929, | |
| "rewards/chosen": -2.916919708251953, | |
| "rewards/margins": 3.4423089027404785, | |
| "rewards/rejected": -6.359227657318115, | |
| "step": 400 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 540, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 20, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.4780471294474322e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |