| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 19.547511312217196, |
| "eval_steps": 500, |
| "global_step": 540, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.36199095022624433, |
| "grad_norm": 0.5034113526344299, |
| "learning_rate": 9.259259259259259e-07, |
| "logits/chosen": 1.8096959590911865, |
| "logits/rejected": 1.7158682346343994, |
| "logps/chosen": -106.03636169433594, |
| "logps/rejected": -84.16211700439453, |
| "loss": 0.695, |
| "rewards/accuracies": 0.3499999940395355, |
| "rewards/chosen": -0.00430720392614603, |
| "rewards/margins": 0.00014101080887485296, |
| "rewards/rejected": -0.004448213614523411, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.7239819004524887, |
| "grad_norm": 0.4689481854438782, |
| "learning_rate": 1.8518518518518519e-06, |
| "logits/chosen": 1.6904691457748413, |
| "logits/rejected": 1.6791242361068726, |
| "logps/chosen": -82.1568603515625, |
| "logps/rejected": -81.06039428710938, |
| "loss": 0.6942, |
| "rewards/accuracies": 0.5375000238418579, |
| "rewards/chosen": -0.002815738320350647, |
| "rewards/margins": 0.005291810259222984, |
| "rewards/rejected": -0.008107547648251057, |
| "step": 20 |
| }, |
| { |
| "epoch": 1.085972850678733, |
| "grad_norm": 0.5041429996490479, |
| "learning_rate": 2.7777777777777783e-06, |
| "logits/chosen": 1.877396583557129, |
| "logits/rejected": 1.9337213039398193, |
| "logps/chosen": -76.86979675292969, |
| "logps/rejected": -86.98431396484375, |
| "loss": 0.6937, |
| "rewards/accuracies": 0.48750001192092896, |
| "rewards/chosen": -0.002632059855386615, |
| "rewards/margins": -0.00260960147716105, |
| "rewards/rejected": -2.245856376248412e-05, |
| "step": 30 |
| }, |
| { |
| "epoch": 1.4479638009049773, |
| "grad_norm": 0.49786925315856934, |
| "learning_rate": 3.7037037037037037e-06, |
| "logits/chosen": 1.622865915298462, |
| "logits/rejected": 1.649457573890686, |
| "logps/chosen": -77.10165405273438, |
| "logps/rejected": -75.56939697265625, |
| "loss": 0.6927, |
| "rewards/accuracies": 0.5249999761581421, |
| "rewards/chosen": -0.005956621374934912, |
| "rewards/margins": -0.0038306519854813814, |
| "rewards/rejected": -0.0021259686909615993, |
| "step": 40 |
| }, |
| { |
| "epoch": 1.8099547511312217, |
| "grad_norm": 0.6152529120445251, |
| "learning_rate": 4.62962962962963e-06, |
| "logits/chosen": 1.7763454914093018, |
| "logits/rejected": 1.8556989431381226, |
| "logps/chosen": -84.42935943603516, |
| "logps/rejected": -97.08312225341797, |
| "loss": 0.6937, |
| "rewards/accuracies": 0.5625, |
| "rewards/chosen": 0.008006314747035503, |
| "rewards/margins": 0.005445700138807297, |
| "rewards/rejected": 0.0025606155395507812, |
| "step": 50 |
| }, |
| { |
| "epoch": 2.171945701357466, |
| "grad_norm": 0.539294958114624, |
| "learning_rate": 4.998119881260576e-06, |
| "logits/chosen": 1.7218296527862549, |
| "logits/rejected": 1.759447455406189, |
| "logps/chosen": -71.67170715332031, |
| "logps/rejected": -86.7840805053711, |
| "loss": 0.6935, |
| "rewards/accuracies": 0.4625000059604645, |
| "rewards/chosen": 0.0036853079218417406, |
| "rewards/margins": -0.002186889760196209, |
| "rewards/rejected": 0.005872196517884731, |
| "step": 60 |
| }, |
| { |
| "epoch": 2.5339366515837103, |
| "grad_norm": 0.5369167327880859, |
| "learning_rate": 4.9866405060165044e-06, |
| "logits/chosen": 1.7708947658538818, |
| "logits/rejected": 1.8614221811294556, |
| "logps/chosen": -86.73070526123047, |
| "logps/rejected": -83.09151458740234, |
| "loss": 0.6909, |
| "rewards/accuracies": 0.5, |
| "rewards/chosen": 0.008487144485116005, |
| "rewards/margins": 0.004935932345688343, |
| "rewards/rejected": 0.0035512119065970182, |
| "step": 70 |
| }, |
| { |
| "epoch": 2.8959276018099547, |
| "grad_norm": 0.5544038414955139, |
| "learning_rate": 4.964774158361991e-06, |
| "logits/chosen": 1.772223711013794, |
| "logits/rejected": 1.834602952003479, |
| "logps/chosen": -69.32654571533203, |
| "logps/rejected": -77.55828094482422, |
| "loss": 0.6943, |
| "rewards/accuracies": 0.5249999761581421, |
| "rewards/chosen": -0.012612676247954369, |
| "rewards/margins": -0.011703734286129475, |
| "rewards/rejected": -0.0009089424274861813, |
| "step": 80 |
| }, |
| { |
| "epoch": 3.257918552036199, |
| "grad_norm": 0.5990457534790039, |
| "learning_rate": 4.93261217644956e-06, |
| "logits/chosen": 1.8591846227645874, |
| "logits/rejected": 1.8804810047149658, |
| "logps/chosen": -79.4209213256836, |
| "logps/rejected": -76.4483642578125, |
| "loss": 0.6927, |
| "rewards/accuracies": 0.4749999940395355, |
| "rewards/chosen": -0.0016643519047647715, |
| "rewards/margins": 0.006901158951222897, |
| "rewards/rejected": -0.008565512485802174, |
| "step": 90 |
| }, |
| { |
| "epoch": 3.6199095022624435, |
| "grad_norm": 0.6173596978187561, |
| "learning_rate": 4.8902889044347e-06, |
| "logits/chosen": 1.777044653892517, |
| "logits/rejected": 1.7735064029693604, |
| "logps/chosen": -85.6143569946289, |
| "logps/rejected": -79.8615493774414, |
| "loss": 0.6894, |
| "rewards/accuracies": 0.5375000238418579, |
| "rewards/chosen": -0.0019477702444419265, |
| "rewards/margins": 0.002171102212741971, |
| "rewards/rejected": -0.004118871875107288, |
| "step": 100 |
| }, |
| { |
| "epoch": 3.981900452488688, |
| "grad_norm": 0.6671907901763916, |
| "learning_rate": 4.837981131305475e-06, |
| "logits/chosen": 1.6935539245605469, |
| "logits/rejected": 1.7035915851593018, |
| "logps/chosen": -83.69059753417969, |
| "logps/rejected": -89.67396545410156, |
| "loss": 0.6907, |
| "rewards/accuracies": 0.48750001192092896, |
| "rewards/chosen": -0.005695090629160404, |
| "rewards/margins": 0.0027313486207276583, |
| "rewards/rejected": -0.00842643715441227, |
| "step": 110 |
| }, |
| { |
| "epoch": 4.343891402714932, |
| "grad_norm": 0.9499974250793457, |
| "learning_rate": 4.775907352415367e-06, |
| "logits/chosen": 1.801300287246704, |
| "logits/rejected": 1.8080482482910156, |
| "logps/chosen": -92.2596206665039, |
| "logps/rejected": -85.3951416015625, |
| "loss": 0.6863, |
| "rewards/accuracies": 0.637499988079071, |
| "rewards/chosen": 0.000912027433514595, |
| "rewards/margins": 0.017921410501003265, |
| "rewards/rejected": -0.01700938306748867, |
| "step": 120 |
| }, |
| { |
| "epoch": 4.705882352941177, |
| "grad_norm": 0.6524670124053955, |
| "learning_rate": 4.70432685680402e-06, |
| "logits/chosen": 1.8659807443618774, |
| "logits/rejected": 1.8858706951141357, |
| "logps/chosen": -78.38592529296875, |
| "logps/rejected": -81.4170913696289, |
| "loss": 0.6873, |
| "rewards/accuracies": 0.5375000238418579, |
| "rewards/chosen": 0.002340498147532344, |
| "rewards/margins": 0.014832529239356518, |
| "rewards/rejected": -0.01249203272163868, |
| "step": 130 |
| }, |
| { |
| "epoch": 5.067873303167421, |
| "grad_norm": 0.7461792826652527, |
| "learning_rate": 4.623538644118244e-06, |
| "logits/chosen": 1.7999374866485596, |
| "logits/rejected": 1.869183897972107, |
| "logps/chosen": -71.37133026123047, |
| "logps/rejected": -83.36344909667969, |
| "loss": 0.6829, |
| "rewards/accuracies": 0.612500011920929, |
| "rewards/chosen": -0.013370734639465809, |
| "rewards/margins": 0.013819178566336632, |
| "rewards/rejected": -0.027189914137125015, |
| "step": 140 |
| }, |
| { |
| "epoch": 5.429864253393665, |
| "grad_norm": 0.6743089556694031, |
| "learning_rate": 4.533880175657419e-06, |
| "logits/chosen": 1.8166606426239014, |
| "logits/rejected": 1.7869322299957275, |
| "logps/chosen": -87.78181457519531, |
| "logps/rejected": -71.32334899902344, |
| "loss": 0.6758, |
| "rewards/accuracies": 0.6499999761581421, |
| "rewards/chosen": -0.0059204003773629665, |
| "rewards/margins": 0.03691873326897621, |
| "rewards/rejected": -0.04283912852406502, |
| "step": 150 |
| }, |
| { |
| "epoch": 5.791855203619909, |
| "grad_norm": 0.7719553112983704, |
| "learning_rate": 4.435725964760331e-06, |
| "logits/chosen": 1.7498859167099, |
| "logits/rejected": 1.7752033472061157, |
| "logps/chosen": -72.18727111816406, |
| "logps/rejected": -84.50582122802734, |
| "loss": 0.6786, |
| "rewards/accuracies": 0.675000011920929, |
| "rewards/chosen": -0.023137167096138, |
| "rewards/margins": 0.042744893580675125, |
| "rewards/rejected": -0.06588206440210342, |
| "step": 160 |
| }, |
| { |
| "epoch": 6.153846153846154, |
| "grad_norm": 0.8010989427566528, |
| "learning_rate": 4.329486012421531e-06, |
| "logits/chosen": 1.7785524129867554, |
| "logits/rejected": 1.770937204360962, |
| "logps/chosen": -87.91519165039062, |
| "logps/rejected": -86.76083374023438, |
| "loss": 0.6689, |
| "rewards/accuracies": 0.6499999761581421, |
| "rewards/chosen": -0.019223090261220932, |
| "rewards/margins": 0.03557150438427925, |
| "rewards/rejected": -0.054794590920209885, |
| "step": 170 |
| }, |
| { |
| "epoch": 6.515837104072398, |
| "grad_norm": 0.9732075929641724, |
| "learning_rate": 4.215604094671835e-06, |
| "logits/chosen": 1.76633620262146, |
| "logits/rejected": 1.782987356185913, |
| "logps/chosen": -85.80155944824219, |
| "logps/rejected": -85.45477294921875, |
| "loss": 0.6648, |
| "rewards/accuracies": 0.675000011920929, |
| "rewards/chosen": -0.009860477410256863, |
| "rewards/margins": 0.0789995938539505, |
| "rewards/rejected": -0.08886007964611053, |
| "step": 180 |
| }, |
| { |
| "epoch": 6.877828054298643, |
| "grad_norm": 0.8777466416358948, |
| "learning_rate": 4.094555908876765e-06, |
| "logits/chosen": 1.8455078601837158, |
| "logits/rejected": 1.871913194656372, |
| "logps/chosen": -71.23600006103516, |
| "logps/rejected": -71.44808197021484, |
| "loss": 0.6606, |
| "rewards/accuracies": 0.612500011920929, |
| "rewards/chosen": -0.04135243222117424, |
| "rewards/margins": 0.06673365831375122, |
| "rewards/rejected": -0.10808608680963516, |
| "step": 190 |
| }, |
| { |
| "epoch": 7.239819004524887, |
| "grad_norm": 0.8959658741950989, |
| "learning_rate": 3.966847086696045e-06, |
| "logits/chosen": 1.8061046600341797, |
| "logits/rejected": 1.8204586505889893, |
| "logps/chosen": -80.91191864013672, |
| "logps/rejected": -73.68770599365234, |
| "loss": 0.6511, |
| "rewards/accuracies": 0.7875000238418579, |
| "rewards/chosen": -0.037331439554691315, |
| "rewards/margins": 0.11111048609018326, |
| "rewards/rejected": -0.14844192564487457, |
| "step": 200 |
| }, |
| { |
| "epoch": 7.601809954751131, |
| "grad_norm": 0.9700360298156738, |
| "learning_rate": 3.833011082004229e-06, |
| "logits/chosen": 1.7680559158325195, |
| "logits/rejected": 1.8058135509490967, |
| "logps/chosen": -77.65849304199219, |
| "logps/rejected": -78.11155700683594, |
| "loss": 0.645, |
| "rewards/accuracies": 0.7250000238418579, |
| "rewards/chosen": -0.0448901504278183, |
| "rewards/margins": 0.12249104678630829, |
| "rewards/rejected": -0.1673811972141266, |
| "step": 210 |
| }, |
| { |
| "epoch": 7.963800904977376, |
| "grad_norm": 1.070133924484253, |
| "learning_rate": 3.693606942594873e-06, |
| "logits/chosen": 1.7242870330810547, |
| "logits/rejected": 1.7760801315307617, |
| "logps/chosen": -79.4949722290039, |
| "logps/rejected": -84.17713928222656, |
| "loss": 0.6516, |
| "rewards/accuracies": 0.699999988079071, |
| "rewards/chosen": -0.06297416985034943, |
| "rewards/margins": 0.1389354169368744, |
| "rewards/rejected": -0.20190958678722382, |
| "step": 220 |
| }, |
| { |
| "epoch": 8.32579185520362, |
| "grad_norm": 0.880223274230957, |
| "learning_rate": 3.549216974976073e-06, |
| "logits/chosen": 1.8455660343170166, |
| "logits/rejected": 1.8497081995010376, |
| "logps/chosen": -82.68794250488281, |
| "logps/rejected": -87.30531311035156, |
| "loss": 0.6396, |
| "rewards/accuracies": 0.6875, |
| "rewards/chosen": -0.09025909751653671, |
| "rewards/margins": 0.1426868587732315, |
| "rewards/rejected": -0.23294594883918762, |
| "step": 230 |
| }, |
| { |
| "epoch": 8.687782805429864, |
| "grad_norm": 0.9315020442008972, |
| "learning_rate": 3.400444312011776e-06, |
| "logits/chosen": 1.7252687215805054, |
| "logits/rejected": 1.7745378017425537, |
| "logps/chosen": -76.34024047851562, |
| "logps/rejected": -89.45391845703125, |
| "loss": 0.6325, |
| "rewards/accuracies": 0.737500011920929, |
| "rewards/chosen": -0.07824709266424179, |
| "rewards/margins": 0.14033600687980652, |
| "rewards/rejected": -0.21858307719230652, |
| "step": 240 |
| }, |
| { |
| "epoch": 9.049773755656108, |
| "grad_norm": 0.9806249141693115, |
| "learning_rate": 3.2479103935691047e-06, |
| "logits/chosen": 1.7673581838607788, |
| "logits/rejected": 1.8124672174453735, |
| "logps/chosen": -91.80342102050781, |
| "logps/rejected": -91.12406921386719, |
| "loss": 0.6314, |
| "rewards/accuracies": 0.75, |
| "rewards/chosen": -0.10384390503168106, |
| "rewards/margins": 0.189823716878891, |
| "rewards/rejected": -0.29366764426231384, |
| "step": 250 |
| }, |
| { |
| "epoch": 9.411764705882353, |
| "grad_norm": 1.0357781648635864, |
| "learning_rate": 3.092252370695298e-06, |
| "logits/chosen": 1.7498366832733154, |
| "logits/rejected": 1.6984941959381104, |
| "logps/chosen": -95.309326171875, |
| "logps/rejected": -80.66096496582031, |
| "loss": 0.6213, |
| "rewards/accuracies": 0.6875, |
| "rewards/chosen": -0.14446459710597992, |
| "rewards/margins": 0.1758357733488083, |
| "rewards/rejected": -0.3203004002571106, |
| "step": 260 |
| }, |
| { |
| "epoch": 9.773755656108598, |
| "grad_norm": 0.9333145022392273, |
| "learning_rate": 2.9341204441673267e-06, |
| "logits/chosen": 1.8672996759414673, |
| "logits/rejected": 1.7658369541168213, |
| "logps/chosen": -104.98448181152344, |
| "logps/rejected": -81.6786880493164, |
| "loss": 0.6131, |
| "rewards/accuracies": 0.7875000238418579, |
| "rewards/chosen": -0.05903264880180359, |
| "rewards/margins": 0.22136883437633514, |
| "rewards/rejected": -0.28040146827697754, |
| "step": 270 |
| }, |
| { |
| "epoch": 10.135746606334841, |
| "grad_norm": 0.9343512654304504, |
| "learning_rate": 2.7741751485313295e-06, |
| "logits/chosen": 1.7365798950195312, |
| "logits/rejected": 1.748958945274353, |
| "logps/chosen": -84.08332824707031, |
| "logps/rejected": -87.89321899414062, |
| "loss": 0.6121, |
| "rewards/accuracies": 0.6875, |
| "rewards/chosen": -0.13087712228298187, |
| "rewards/margins": 0.16090884804725647, |
| "rewards/rejected": -0.29178598523139954, |
| "step": 280 |
| }, |
| { |
| "epoch": 10.497737556561086, |
| "grad_norm": 0.9784368872642517, |
| "learning_rate": 2.6130845929767662e-06, |
| "logits/chosen": 1.7028934955596924, |
| "logits/rejected": 1.6997534036636353, |
| "logps/chosen": -86.77079010009766, |
| "logps/rejected": -85.89974975585938, |
| "loss": 0.612, |
| "rewards/accuracies": 0.7250000238418579, |
| "rewards/chosen": -0.10491044819355011, |
| "rewards/margins": 0.19683709740638733, |
| "rewards/rejected": -0.30174753069877625, |
| "step": 290 |
| }, |
| { |
| "epoch": 10.85972850678733, |
| "grad_norm": 1.030173659324646, |
| "learning_rate": 2.4515216705704396e-06, |
| "logits/chosen": 1.8057117462158203, |
| "logits/rejected": 1.8746700286865234, |
| "logps/chosen": -74.07008361816406, |
| "logps/rejected": -89.6046142578125, |
| "loss": 0.5997, |
| "rewards/accuracies": 0.8125, |
| "rewards/chosen": -0.058038175106048584, |
| "rewards/margins": 0.2527470886707306, |
| "rewards/rejected": -0.3107852637767792, |
| "step": 300 |
| }, |
| { |
| "epoch": 11.221719457013574, |
| "grad_norm": 1.0045686960220337, |
| "learning_rate": 2.290161247507733e-06, |
| "logits/chosen": 1.7842658758163452, |
| "logits/rejected": 1.83527410030365, |
| "logps/chosen": -85.32085418701172, |
| "logps/rejected": -91.45645904541016, |
| "loss": 0.5967, |
| "rewards/accuracies": 0.800000011920929, |
| "rewards/chosen": -0.1375085413455963, |
| "rewards/margins": 0.2444867342710495, |
| "rewards/rejected": -0.3819952607154846, |
| "step": 310 |
| }, |
| { |
| "epoch": 11.583710407239819, |
| "grad_norm": 0.967028796672821, |
| "learning_rate": 2.129677344121879e-06, |
| "logits/chosen": 1.8002763986587524, |
| "logits/rejected": 1.7979824542999268, |
| "logps/chosen": -97.67089080810547, |
| "logps/rejected": -79.14571380615234, |
| "loss": 0.5852, |
| "rewards/accuracies": 0.762499988079071, |
| "rewards/chosen": -0.1535075604915619, |
| "rewards/margins": 0.18517914414405823, |
| "rewards/rejected": -0.3386867046356201, |
| "step": 320 |
| }, |
| { |
| "epoch": 11.945701357466064, |
| "grad_norm": 1.017149806022644, |
| "learning_rate": 1.970740319426474e-06, |
| "logits/chosen": 1.8270378112792969, |
| "logits/rejected": 1.8545118570327759, |
| "logps/chosen": -74.19200897216797, |
| "logps/rejected": -86.13984680175781, |
| "loss": 0.6015, |
| "rewards/accuracies": 0.7749999761581421, |
| "rewards/chosen": -0.1458810567855835, |
| "rewards/margins": 0.20589685440063477, |
| "rewards/rejected": -0.35177791118621826, |
| "step": 330 |
| }, |
| { |
| "epoch": 12.307692307692308, |
| "grad_norm": 0.9695628881454468, |
| "learning_rate": 1.8140140709517467e-06, |
| "logits/chosen": 1.745535135269165, |
| "logits/rejected": 1.7830098867416382, |
| "logps/chosen": -84.69477081298828, |
| "logps/rejected": -82.7412109375, |
| "loss": 0.5882, |
| "rewards/accuracies": 0.762499988079071, |
| "rewards/chosen": -0.12116970866918564, |
| "rewards/margins": 0.25774163007736206, |
| "rewards/rejected": -0.3789113759994507, |
| "step": 340 |
| }, |
| { |
| "epoch": 12.669683257918551, |
| "grad_norm": 1.103960394859314, |
| "learning_rate": 1.6601532615711452e-06, |
| "logits/chosen": 1.777008295059204, |
| "logits/rejected": 1.6892492771148682, |
| "logps/chosen": -100.1700210571289, |
| "logps/rejected": -90.1813735961914, |
| "loss": 0.5802, |
| "rewards/accuracies": 0.7749999761581421, |
| "rewards/chosen": -0.10893051326274872, |
| "rewards/margins": 0.29678279161453247, |
| "rewards/rejected": -0.4057133197784424, |
| "step": 350 |
| }, |
| { |
| "epoch": 13.031674208144796, |
| "grad_norm": 1.0589226484298706, |
| "learning_rate": 1.509800584902108e-06, |
| "logits/chosen": 1.7005256414413452, |
| "logits/rejected": 1.7957611083984375, |
| "logps/chosen": -79.087158203125, |
| "logps/rejected": -102.69563293457031, |
| "loss": 0.5814, |
| "rewards/accuracies": 0.862500011920929, |
| "rewards/chosen": -0.13767623901367188, |
| "rewards/margins": 0.3307909071445465, |
| "rewards/rejected": -0.4684671461582184, |
| "step": 360 |
| }, |
| { |
| "epoch": 13.393665158371041, |
| "grad_norm": 1.056600570678711, |
| "learning_rate": 1.3635840807037487e-06, |
| "logits/chosen": 1.6859004497528076, |
| "logits/rejected": 1.7192614078521729, |
| "logps/chosen": -96.27987670898438, |
| "logps/rejected": -96.15126037597656, |
| "loss": 0.5703, |
| "rewards/accuracies": 0.8500000238418579, |
| "rewards/chosen": -0.16748157143592834, |
| "rewards/margins": 0.38242021203041077, |
| "rewards/rejected": -0.5499017834663391, |
| "step": 370 |
| }, |
| { |
| "epoch": 13.755656108597286, |
| "grad_norm": 2.783137321472168, |
| "learning_rate": 1.2221145114853172e-06, |
| "logits/chosen": 1.8700748682022095, |
| "logits/rejected": 1.886313796043396, |
| "logps/chosen": -82.22146606445312, |
| "logps/rejected": -79.1581802368164, |
| "loss": 0.5791, |
| "rewards/accuracies": 0.8125, |
| "rewards/chosen": -0.16777488589286804, |
| "rewards/margins": 0.2746582329273224, |
| "rewards/rejected": -0.44243311882019043, |
| "step": 380 |
| }, |
| { |
| "epoch": 14.117647058823529, |
| "grad_norm": 1.137703776359558, |
| "learning_rate": 1.085982811283654e-06, |
| "logits/chosen": 1.7121073007583618, |
| "logits/rejected": 1.74045729637146, |
| "logps/chosen": -81.7417221069336, |
| "logps/rejected": -82.1438980102539, |
| "loss": 0.5792, |
| "rewards/accuracies": 0.762499988079071, |
| "rewards/chosen": -0.21657264232635498, |
| "rewards/margins": 0.23677067458629608, |
| "rewards/rejected": -0.45334330201148987, |
| "step": 390 |
| }, |
| { |
| "epoch": 14.479638009049774, |
| "grad_norm": 1.115824580192566, |
| "learning_rate": 9.557576172663577e-07, |
| "logits/chosen": 1.782957673072815, |
| "logits/rejected": 1.740311861038208, |
| "logps/chosen": -95.57044982910156, |
| "logps/rejected": -87.47151184082031, |
| "loss": 0.5648, |
| "rewards/accuracies": 0.800000011920929, |
| "rewards/chosen": -0.13624358177185059, |
| "rewards/margins": 0.3296685218811035, |
| "rewards/rejected": -0.4659121632575989, |
| "step": 400 |
| }, |
| { |
| "epoch": 14.841628959276019, |
| "grad_norm": 1.0241665840148926, |
| "learning_rate": 8.319828944714508e-07, |
| "logits/chosen": 1.8231852054595947, |
| "logits/rejected": 1.7309335470199585, |
| "logps/chosen": -90.45216369628906, |
| "logps/rejected": -84.18362426757812, |
| "loss": 0.5692, |
| "rewards/accuracies": 0.800000011920929, |
| "rewards/chosen": -0.21606135368347168, |
| "rewards/margins": 0.3016027808189392, |
| "rewards/rejected": -0.5176641345024109, |
| "step": 410 |
| }, |
| { |
| "epoch": 15.203619909502262, |
| "grad_norm": 1.0937505960464478, |
| "learning_rate": 7.151756636052529e-07, |
| "logits/chosen": 1.6469671726226807, |
| "logits/rejected": 1.755052924156189, |
| "logps/chosen": -69.20260620117188, |
| "logps/rejected": -91.29935455322266, |
| "loss": 0.5813, |
| "rewards/accuracies": 0.762499988079071, |
| "rewards/chosen": -0.16654367744922638, |
| "rewards/margins": 0.29352378845214844, |
| "rewards/rejected": -0.4600674510002136, |
| "step": 420 |
| }, |
| { |
| "epoch": 15.565610859728507, |
| "grad_norm": 1.189382553100586, |
| "learning_rate": 6.058238413897052e-07, |
| "logits/chosen": 1.7920663356781006, |
| "logits/rejected": 1.767615556716919, |
| "logps/chosen": -83.10585021972656, |
| "logps/rejected": -87.55345916748047, |
| "loss": 0.5717, |
| "rewards/accuracies": 0.8500000238418579, |
| "rewards/chosen": -0.18599426746368408, |
| "rewards/margins": 0.30218419432640076, |
| "rewards/rejected": -0.48817843198776245, |
| "step": 430 |
| }, |
| { |
| "epoch": 15.927601809954751, |
| "grad_norm": 1.0084723234176636, |
| "learning_rate": 5.043842024802675e-07, |
| "logits/chosen": 1.6690113544464111, |
| "logits/rejected": 1.7104164361953735, |
| "logps/chosen": -77.92027282714844, |
| "logps/rejected": -88.28632354736328, |
| "loss": 0.5616, |
| "rewards/accuracies": 0.824999988079071, |
| "rewards/chosen": -0.1644773930311203, |
| "rewards/margins": 0.31829944252967834, |
| "rewards/rejected": -0.48277679085731506, |
| "step": 440 |
| }, |
| { |
| "epoch": 16.289592760180994, |
| "grad_norm": 1.099560022354126, |
| "learning_rate": 4.1128047146765936e-07, |
| "logits/chosen": 1.777606725692749, |
| "logits/rejected": 1.7520103454589844, |
| "logps/chosen": -92.19166564941406, |
| "logps/rejected": -74.17485809326172, |
| "loss": 0.5653, |
| "rewards/accuracies": 0.800000011920929, |
| "rewards/chosen": -0.1612701714038849, |
| "rewards/margins": 0.32633036375045776, |
| "rewards/rejected": -0.4876004755496979, |
| "step": 450 |
| }, |
| { |
| "epoch": 16.65158371040724, |
| "grad_norm": 1.0390541553497314, |
| "learning_rate": 3.269015529333805e-07, |
| "logits/chosen": 1.6575822830200195, |
| "logits/rejected": 1.785485863685608, |
| "logps/chosen": -81.9493179321289, |
| "logps/rejected": -101.3016128540039, |
| "loss": 0.5579, |
| "rewards/accuracies": 0.8125, |
| "rewards/chosen": -0.20868897438049316, |
| "rewards/margins": 0.3946114182472229, |
| "rewards/rejected": -0.6033004522323608, |
| "step": 460 |
| }, |
| { |
| "epoch": 17.013574660633484, |
| "grad_norm": 1.1404516696929932, |
| "learning_rate": 2.515999069522676e-07, |
| "logits/chosen": 1.6690752506256104, |
| "logits/rejected": 1.6541496515274048, |
| "logps/chosen": -88.44367980957031, |
| "logps/rejected": -71.12745666503906, |
| "loss": 0.5675, |
| "rewards/accuracies": 0.824999988079071, |
| "rewards/chosen": -0.19705724716186523, |
| "rewards/margins": 0.26251187920570374, |
| "rewards/rejected": -0.45956912636756897, |
| "step": 470 |
| }, |
| { |
| "epoch": 17.375565610859727, |
| "grad_norm": 1.175963282585144, |
| "learning_rate": 1.8569007682777417e-07, |
| "logits/chosen": 1.6609585285186768, |
| "logits/rejected": 1.6719191074371338, |
| "logps/chosen": -94.07273864746094, |
| "logps/rejected": -76.59893035888672, |
| "loss": 0.5678, |
| "rewards/accuracies": 0.7749999761581421, |
| "rewards/chosen": -0.1740456521511078, |
| "rewards/margins": 0.3117365837097168, |
| "rewards/rejected": -0.4857822358608246, |
| "step": 480 |
| }, |
| { |
| "epoch": 17.737556561085974, |
| "grad_norm": 1.2133598327636719, |
| "learning_rate": 1.2944737520980883e-07, |
| "logits/chosen": 1.7220537662506104, |
| "logits/rejected": 1.7643390893936157, |
| "logps/chosen": -83.6407699584961, |
| "logps/rejected": -81.10734558105469, |
| "loss": 0.5613, |
| "rewards/accuracies": 0.862500011920929, |
| "rewards/chosen": -0.13465282320976257, |
| "rewards/margins": 0.4027688503265381, |
| "rewards/rejected": -0.5374216437339783, |
| "step": 490 |
| }, |
| { |
| "epoch": 18.099547511312217, |
| "grad_norm": 1.1120809316635132, |
| "learning_rate": 8.310673408334496e-08, |
| "logits/chosen": 1.5700793266296387, |
| "logits/rejected": 1.5635223388671875, |
| "logps/chosen": -86.44955444335938, |
| "logps/rejected": -86.97091674804688, |
| "loss": 0.5626, |
| "rewards/accuracies": 0.8374999761581421, |
| "rewards/chosen": -0.17045333981513977, |
| "rewards/margins": 0.29403942823410034, |
| "rewards/rejected": -0.46449270844459534, |
| "step": 500 |
| }, |
| { |
| "epoch": 18.46153846153846, |
| "grad_norm": 1.177876353263855, |
| "learning_rate": 4.6861723431538273e-08, |
| "logits/chosen": 1.5987281799316406, |
| "logits/rejected": 1.7516323328018188, |
| "logps/chosen": -79.38053131103516, |
| "logps/rejected": -97.10954284667969, |
| "loss": 0.5634, |
| "rewards/accuracies": 0.887499988079071, |
| "rewards/chosen": -0.1326012909412384, |
| "rewards/margins": 0.40437060594558716, |
| "rewards/rejected": -0.5369717478752136, |
| "step": 510 |
| }, |
| { |
| "epoch": 18.823529411764707, |
| "grad_norm": 1.2317314147949219, |
| "learning_rate": 2.0863742672497244e-08, |
| "logits/chosen": 1.71554434299469, |
| "logits/rejected": 1.740006446838379, |
| "logps/chosen": -83.19001007080078, |
| "logps/rejected": -78.78974151611328, |
| "loss": 0.573, |
| "rewards/accuracies": 0.75, |
| "rewards/chosen": -0.23421506583690643, |
| "rewards/margins": 0.25640812516212463, |
| "rewards/rejected": -0.4906231760978699, |
| "step": 520 |
| }, |
| { |
| "epoch": 19.18552036199095, |
| "grad_norm": 1.118938684463501, |
| "learning_rate": 5.221388247169945e-09, |
| "logits/chosen": 1.7030197381973267, |
| "logits/rejected": 1.7396942377090454, |
| "logps/chosen": -86.79795837402344, |
| "logps/rejected": -97.96685791015625, |
| "loss": 0.5571, |
| "rewards/accuracies": 0.75, |
| "rewards/chosen": -0.17865614593029022, |
| "rewards/margins": 0.3226766586303711, |
| "rewards/rejected": -0.5013328790664673, |
| "step": 530 |
| }, |
| { |
| "epoch": 19.547511312217196, |
| "grad_norm": 1.1492843627929688, |
| "learning_rate": 0.0, |
| "logits/chosen": 1.6407601833343506, |
| "logits/rejected": 1.645991325378418, |
| "logps/chosen": -74.029296875, |
| "logps/rejected": -82.86942291259766, |
| "loss": 0.5595, |
| "rewards/accuracies": 0.762499988079071, |
| "rewards/chosen": -0.228031724691391, |
| "rewards/margins": 0.28408557176589966, |
| "rewards/rejected": -0.5121172666549683, |
| "step": 540 |
| }, |
| { |
| "epoch": 19.547511312217196, |
| "step": 540, |
| "total_flos": 1.848581334679683e+18, |
| "train_loss": 0.6237324387938888, |
| "train_runtime": 3964.2678, |
| "train_samples_per_second": 8.915, |
| "train_steps_per_second": 0.136 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 540, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 20, |
| "save_steps": 50, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.848581334679683e+18, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|