| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.998691442030882, | |
| "eval_steps": 100, | |
| "global_step": 477, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002093692750588851, | |
| "grad_norm": 5.344821351720907, | |
| "learning_rate": 1.0416666666666666e-08, | |
| "logits/chosen": -0.9863196015357971, | |
| "logits/rejected": -1.0256754159927368, | |
| "logps/chosen": -474.4989013671875, | |
| "logps/rejected": -524.1282958984375, | |
| "loss": 0.6931, | |
| "rewards/accuracies": 0.0, | |
| "rewards/chosen": 0.0, | |
| "rewards/margins": 0.0, | |
| "rewards/rejected": 0.0, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.02093692750588851, | |
| "grad_norm": 6.122277210971997, | |
| "learning_rate": 1.0416666666666667e-07, | |
| "logits/chosen": -1.199143886566162, | |
| "logits/rejected": -1.211472749710083, | |
| "logps/chosen": -504.70758056640625, | |
| "logps/rejected": -517.0758666992188, | |
| "loss": 0.6927, | |
| "rewards/accuracies": 0.4930555522441864, | |
| "rewards/chosen": -2.0182769731036387e-05, | |
| "rewards/margins": 0.0008512942586094141, | |
| "rewards/rejected": -0.0008714771247468889, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.04187385501177702, | |
| "grad_norm": 6.385948578335296, | |
| "learning_rate": 2.0833333333333333e-07, | |
| "logits/chosen": -1.1590700149536133, | |
| "logits/rejected": -1.1333811283111572, | |
| "logps/chosen": -488.711669921875, | |
| "logps/rejected": -505.6346130371094, | |
| "loss": 0.6921, | |
| "rewards/accuracies": 0.5375000238418579, | |
| "rewards/chosen": -0.007256551180034876, | |
| "rewards/margins": 0.002894227858632803, | |
| "rewards/rejected": -0.010150779038667679, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.06281078251766553, | |
| "grad_norm": 5.830212722805484, | |
| "learning_rate": 3.1249999999999997e-07, | |
| "logits/chosen": -1.1879990100860596, | |
| "logits/rejected": -1.2821766138076782, | |
| "logps/chosen": -507.1835021972656, | |
| "logps/rejected": -481.771728515625, | |
| "loss": 0.6858, | |
| "rewards/accuracies": 0.581250011920929, | |
| "rewards/chosen": -0.05630107969045639, | |
| "rewards/margins": 0.009836100041866302, | |
| "rewards/rejected": -0.06613717973232269, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.08374771002355404, | |
| "grad_norm": 6.142811166492132, | |
| "learning_rate": 4.1666666666666667e-07, | |
| "logits/chosen": -1.2522597312927246, | |
| "logits/rejected": -1.2801125049591064, | |
| "logps/chosen": -476.06439208984375, | |
| "logps/rejected": -518.1241455078125, | |
| "loss": 0.6659, | |
| "rewards/accuracies": 0.668749988079071, | |
| "rewards/chosen": -0.24496185779571533, | |
| "rewards/margins": 0.0600227527320385, | |
| "rewards/rejected": -0.30498459935188293, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.10468463752944256, | |
| "grad_norm": 6.8453175062635125, | |
| "learning_rate": 4.999731868769026e-07, | |
| "logits/chosen": -1.34788978099823, | |
| "logits/rejected": -1.364375114440918, | |
| "logps/chosen": -533.0452880859375, | |
| "logps/rejected": -599.4424438476562, | |
| "loss": 0.6196, | |
| "rewards/accuracies": 0.737500011920929, | |
| "rewards/chosen": -0.7288391590118408, | |
| "rewards/margins": 0.277568519115448, | |
| "rewards/rejected": -1.006407618522644, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.12562156503533106, | |
| "grad_norm": 7.830680395352566, | |
| "learning_rate": 4.990353313429303e-07, | |
| "logits/chosen": -1.4588544368743896, | |
| "logits/rejected": -1.4161434173583984, | |
| "logps/chosen": -587.5367431640625, | |
| "logps/rejected": -669.0986328125, | |
| "loss": 0.5822, | |
| "rewards/accuracies": 0.699999988079071, | |
| "rewards/chosen": -1.328367829322815, | |
| "rewards/margins": 0.5138038396835327, | |
| "rewards/rejected": -1.8421716690063477, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.14655849254121958, | |
| "grad_norm": 8.114776579993702, | |
| "learning_rate": 4.967625656594781e-07, | |
| "logits/chosen": -1.4318482875823975, | |
| "logits/rejected": -1.4409352540969849, | |
| "logps/chosen": -666.4456176757812, | |
| "logps/rejected": -729.1796264648438, | |
| "loss": 0.5595, | |
| "rewards/accuracies": 0.6812499761581421, | |
| "rewards/chosen": -1.651793122291565, | |
| "rewards/margins": 0.5167388916015625, | |
| "rewards/rejected": -2.168531894683838, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.16749542004710807, | |
| "grad_norm": 8.8385356864838, | |
| "learning_rate": 4.93167072587771e-07, | |
| "logits/chosen": -1.4754712581634521, | |
| "logits/rejected": -1.5046911239624023, | |
| "logps/chosen": -637.0550537109375, | |
| "logps/rejected": -683.7327270507812, | |
| "loss": 0.5521, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": -1.4783751964569092, | |
| "rewards/margins": 0.6370193362236023, | |
| "rewards/rejected": -2.1153945922851562, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.1884323475529966, | |
| "grad_norm": 7.273331254641903, | |
| "learning_rate": 4.882681251368548e-07, | |
| "logits/chosen": -1.4381468296051025, | |
| "logits/rejected": -1.4646328687667847, | |
| "logps/chosen": -618.9959716796875, | |
| "logps/rejected": -673.4351806640625, | |
| "loss": 0.534, | |
| "rewards/accuracies": 0.699999988079071, | |
| "rewards/chosen": -1.519857406616211, | |
| "rewards/margins": 0.5099424719810486, | |
| "rewards/rejected": -2.0297999382019043, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2093692750588851, | |
| "grad_norm": 7.2496794921936, | |
| "learning_rate": 4.820919832540181e-07, | |
| "logits/chosen": -1.3987996578216553, | |
| "logits/rejected": -1.4214246273040771, | |
| "logps/chosen": -632.5852661132812, | |
| "logps/rejected": -757.4251098632812, | |
| "loss": 0.5269, | |
| "rewards/accuracies": 0.824999988079071, | |
| "rewards/chosen": -1.539336919784546, | |
| "rewards/margins": 0.8474786877632141, | |
| "rewards/rejected": -2.3868155479431152, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.2093692750588851, | |
| "eval_logits/chosen": -1.3850065469741821, | |
| "eval_logits/rejected": -1.3951938152313232, | |
| "eval_logps/chosen": -657.6356201171875, | |
| "eval_logps/rejected": -727.9815063476562, | |
| "eval_loss": 0.5332552194595337, | |
| "eval_rewards/accuracies": 0.699999988079071, | |
| "eval_rewards/chosen": -1.6756343841552734, | |
| "eval_rewards/margins": 0.6563942432403564, | |
| "eval_rewards/rejected": -2.332028865814209, | |
| "eval_runtime": 174.614, | |
| "eval_samples_per_second": 11.454, | |
| "eval_steps_per_second": 0.716, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.23030620256477363, | |
| "grad_norm": 8.10753128188729, | |
| "learning_rate": 4.7467175306295647e-07, | |
| "logits/chosen": -1.2935419082641602, | |
| "logits/rejected": -1.381008505821228, | |
| "logps/chosen": -657.0072021484375, | |
| "logps/rejected": -706.2003784179688, | |
| "loss": 0.529, | |
| "rewards/accuracies": 0.7562500238418579, | |
| "rewards/chosen": -1.6493545770645142, | |
| "rewards/margins": 0.5988022685050964, | |
| "rewards/rejected": -2.248157024383545, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.2512431300706621, | |
| "grad_norm": 8.142470959373666, | |
| "learning_rate": 4.6604720940421207e-07, | |
| "logits/chosen": -1.3212612867355347, | |
| "logits/rejected": -1.3621456623077393, | |
| "logps/chosen": -609.11181640625, | |
| "logps/rejected": -712.6756591796875, | |
| "loss": 0.5053, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": -1.5495909452438354, | |
| "rewards/margins": 0.7342188954353333, | |
| "rewards/rejected": -2.2838101387023926, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.2721800575765506, | |
| "grad_norm": 6.624266912381754, | |
| "learning_rate": 4.5626458262912735e-07, | |
| "logits/chosen": -1.3223083019256592, | |
| "logits/rejected": -1.3553454875946045, | |
| "logps/chosen": -632.750244140625, | |
| "logps/rejected": -708.5543823242188, | |
| "loss": 0.5004, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": -1.4656463861465454, | |
| "rewards/margins": 0.7802583575248718, | |
| "rewards/rejected": -2.2459044456481934, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.29311698508243916, | |
| "grad_norm": 9.216064563940549, | |
| "learning_rate": 4.453763107901675e-07, | |
| "logits/chosen": -1.4132908582687378, | |
| "logits/rejected": -1.4507051706314087, | |
| "logps/chosen": -685.0856323242188, | |
| "logps/rejected": -762.1068115234375, | |
| "loss": 0.5347, | |
| "rewards/accuracies": 0.7437499761581421, | |
| "rewards/chosen": -1.7906442880630493, | |
| "rewards/margins": 0.8574408292770386, | |
| "rewards/rejected": -2.648085117340088, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.31405391258832765, | |
| "grad_norm": 8.861919067605399, | |
| "learning_rate": 4.3344075855595097e-07, | |
| "logits/chosen": -1.3353523015975952, | |
| "logits/rejected": -1.3859806060791016, | |
| "logps/chosen": -644.6119384765625, | |
| "logps/rejected": -741.2134399414062, | |
| "loss": 0.5072, | |
| "rewards/accuracies": 0.831250011920929, | |
| "rewards/chosen": -1.672431230545044, | |
| "rewards/margins": 0.8643969297409058, | |
| "rewards/rejected": -2.5368282794952393, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.33499084009421615, | |
| "grad_norm": 9.292299719932993, | |
| "learning_rate": 4.2052190435769554e-07, | |
| "logits/chosen": -1.2773377895355225, | |
| "logits/rejected": -1.3252099752426147, | |
| "logps/chosen": -646.8428344726562, | |
| "logps/rejected": -767.851806640625, | |
| "loss": 0.5175, | |
| "rewards/accuracies": 0.7749999761581421, | |
| "rewards/chosen": -1.8619630336761475, | |
| "rewards/margins": 0.8954536318778992, | |
| "rewards/rejected": -2.7574167251586914, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.3559277676001047, | |
| "grad_norm": 8.535232197560195, | |
| "learning_rate": 4.0668899744407567e-07, | |
| "logits/chosen": -1.2284882068634033, | |
| "logits/rejected": -1.259841799736023, | |
| "logps/chosen": -643.7806396484375, | |
| "logps/rejected": -719.5083618164062, | |
| "loss": 0.5192, | |
| "rewards/accuracies": 0.668749988079071, | |
| "rewards/chosen": -1.9132611751556396, | |
| "rewards/margins": 0.6574370265007019, | |
| "rewards/rejected": -2.5706982612609863, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.3768646951059932, | |
| "grad_norm": 8.471513794877017, | |
| "learning_rate": 3.920161866827889e-07, | |
| "logits/chosen": -1.3475391864776611, | |
| "logits/rejected": -1.3049246072769165, | |
| "logps/chosen": -670.0532836914062, | |
| "logps/rejected": -781.0071411132812, | |
| "loss": 0.4907, | |
| "rewards/accuracies": 0.731249988079071, | |
| "rewards/chosen": -2.043039083480835, | |
| "rewards/margins": 0.6995586156845093, | |
| "rewards/rejected": -2.742597818374634, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.39780162261188173, | |
| "grad_norm": 7.575885046762926, | |
| "learning_rate": 3.765821230985757e-07, | |
| "logits/chosen": -1.3079270124435425, | |
| "logits/rejected": -1.2785875797271729, | |
| "logps/chosen": -679.8087768554688, | |
| "logps/rejected": -808.7610473632812, | |
| "loss": 0.501, | |
| "rewards/accuracies": 0.71875, | |
| "rewards/chosen": -2.0995121002197266, | |
| "rewards/margins": 0.7744059562683105, | |
| "rewards/rejected": -2.873918056488037, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.4187385501177702, | |
| "grad_norm": 11.938650112523503, | |
| "learning_rate": 3.604695382782159e-07, | |
| "logits/chosen": -1.2764681577682495, | |
| "logits/rejected": -1.2830474376678467, | |
| "logps/chosen": -700.9818115234375, | |
| "logps/rejected": -790.2588500976562, | |
| "loss": 0.5086, | |
| "rewards/accuracies": 0.6812499761581421, | |
| "rewards/chosen": -2.1675782203674316, | |
| "rewards/margins": 0.6838473081588745, | |
| "rewards/rejected": -2.851425886154175, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4187385501177702, | |
| "eval_logits/chosen": -1.2772727012634277, | |
| "eval_logits/rejected": -1.2939157485961914, | |
| "eval_logps/chosen": -699.1298217773438, | |
| "eval_logps/rejected": -787.651123046875, | |
| "eval_loss": 0.504360556602478, | |
| "eval_rewards/accuracies": 0.7039999961853027, | |
| "eval_rewards/chosen": -2.090576171875, | |
| "eval_rewards/margins": 0.8381478786468506, | |
| "eval_rewards/rejected": -2.9287240505218506, | |
| "eval_runtime": 173.7524, | |
| "eval_samples_per_second": 11.511, | |
| "eval_steps_per_second": 0.719, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.4396754776236587, | |
| "grad_norm": 8.791948417829072, | |
| "learning_rate": 3.4376480090239047e-07, | |
| "logits/chosen": -1.2757617235183716, | |
| "logits/rejected": -1.3068811893463135, | |
| "logps/chosen": -736.14794921875, | |
| "logps/rejected": -807.1375732421875, | |
| "loss": 0.5129, | |
| "rewards/accuracies": 0.71875, | |
| "rewards/chosen": -2.278721332550049, | |
| "rewards/margins": 0.7708470225334167, | |
| "rewards/rejected": -3.0495681762695312, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.46061240512954726, | |
| "grad_norm": 7.175329586845978, | |
| "learning_rate": 3.265574537815398e-07, | |
| "logits/chosen": -1.234605312347412, | |
| "logits/rejected": -1.238705039024353, | |
| "logps/chosen": -686.875244140625, | |
| "logps/rejected": -792.4070434570312, | |
| "loss": 0.5007, | |
| "rewards/accuracies": 0.71875, | |
| "rewards/chosen": -2.308474063873291, | |
| "rewards/margins": 0.764443039894104, | |
| "rewards/rejected": -3.0729172229766846, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.48154933263543576, | |
| "grad_norm": 8.60332250684624, | |
| "learning_rate": 3.0893973387735683e-07, | |
| "logits/chosen": -1.2985376119613647, | |
| "logits/rejected": -1.306581974029541, | |
| "logps/chosen": -662.0936279296875, | |
| "logps/rejected": -823.6008911132812, | |
| "loss": 0.5153, | |
| "rewards/accuracies": 0.7749999761581421, | |
| "rewards/chosen": -2.064509868621826, | |
| "rewards/margins": 1.0347161293029785, | |
| "rewards/rejected": -3.0992257595062256, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5024862601413242, | |
| "grad_norm": 7.615515586975465, | |
| "learning_rate": 2.910060778827554e-07, | |
| "logits/chosen": -1.244362711906433, | |
| "logits/rejected": -1.250091552734375, | |
| "logps/chosen": -670.7986450195312, | |
| "logps/rejected": -774.2589721679688, | |
| "loss": 0.488, | |
| "rewards/accuracies": 0.71875, | |
| "rewards/chosen": -2.0815677642822266, | |
| "rewards/margins": 0.7600711584091187, | |
| "rewards/rejected": -2.8416390419006348, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.5234231876472127, | |
| "grad_norm": 8.227460909814754, | |
| "learning_rate": 2.7285261601056697e-07, | |
| "logits/chosen": -1.2554184198379517, | |
| "logits/rejected": -1.2809184789657593, | |
| "logps/chosen": -725.0731201171875, | |
| "logps/rejected": -819.86181640625, | |
| "loss": 0.4948, | |
| "rewards/accuracies": 0.7562500238418579, | |
| "rewards/chosen": -2.1545822620391846, | |
| "rewards/margins": 0.9376426935195923, | |
| "rewards/rejected": -3.0922253131866455, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.5443601151531012, | |
| "grad_norm": 9.508165144084513, | |
| "learning_rate": 2.5457665670441937e-07, | |
| "logits/chosen": -1.2575011253356934, | |
| "logits/rejected": -1.3054656982421875, | |
| "logps/chosen": -716.1454467773438, | |
| "logps/rejected": -812.8175659179688, | |
| "loss": 0.4996, | |
| "rewards/accuracies": 0.737500011920929, | |
| "rewards/chosen": -2.350128650665283, | |
| "rewards/margins": 0.9231952428817749, | |
| "rewards/rejected": -3.2733237743377686, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.5652970426589898, | |
| "grad_norm": 9.240543953923092, | |
| "learning_rate": 2.3627616503391812e-07, | |
| "logits/chosen": -1.2486110925674438, | |
| "logits/rejected": -1.25895094871521, | |
| "logps/chosen": -710.92724609375, | |
| "logps/rejected": -834.6962890625, | |
| "loss": 0.5086, | |
| "rewards/accuracies": 0.7562500238418579, | |
| "rewards/chosen": -2.2452242374420166, | |
| "rewards/margins": 1.0662496089935303, | |
| "rewards/rejected": -3.3114733695983887, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.5862339701648783, | |
| "grad_norm": 8.132372884664997, | |
| "learning_rate": 2.1804923757009882e-07, | |
| "logits/chosen": -1.2164264917373657, | |
| "logits/rejected": -1.3014079332351685, | |
| "logps/chosen": -720.0590209960938, | |
| "logps/rejected": -770.1983642578125, | |
| "loss": 0.4955, | |
| "rewards/accuracies": 0.731249988079071, | |
| "rewards/chosen": -2.291123867034912, | |
| "rewards/margins": 0.673937201499939, | |
| "rewards/rejected": -2.9650609493255615, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.6071708976707668, | |
| "grad_norm": 7.699636946017558, | |
| "learning_rate": 1.9999357655598891e-07, | |
| "logits/chosen": -1.249624252319336, | |
| "logits/rejected": -1.2436301708221436, | |
| "logps/chosen": -687.5323486328125, | |
| "logps/rejected": -815.8274536132812, | |
| "loss": 0.4944, | |
| "rewards/accuracies": 0.7437499761581421, | |
| "rewards/chosen": -2.2351253032684326, | |
| "rewards/margins": 0.8806158900260925, | |
| "rewards/rejected": -3.115741491317749, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.6281078251766553, | |
| "grad_norm": 7.183044496877634, | |
| "learning_rate": 1.8220596619089573e-07, | |
| "logits/chosen": -1.214294672012329, | |
| "logits/rejected": -1.278714656829834, | |
| "logps/chosen": -711.4762573242188, | |
| "logps/rejected": -826.3113403320312, | |
| "loss": 0.4787, | |
| "rewards/accuracies": 0.793749988079071, | |
| "rewards/chosen": -2.1853926181793213, | |
| "rewards/margins": 1.0479013919830322, | |
| "rewards/rejected": -3.2332940101623535, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6281078251766553, | |
| "eval_logits/chosen": -1.2646288871765137, | |
| "eval_logits/rejected": -1.2846323251724243, | |
| "eval_logps/chosen": -719.3385620117188, | |
| "eval_logps/rejected": -811.6696166992188, | |
| "eval_loss": 0.49482667446136475, | |
| "eval_rewards/accuracies": 0.7319999933242798, | |
| "eval_rewards/chosen": -2.292664051055908, | |
| "eval_rewards/margins": 0.8762450218200684, | |
| "eval_rewards/rejected": -3.1689090728759766, | |
| "eval_runtime": 173.013, | |
| "eval_samples_per_second": 11.56, | |
| "eval_steps_per_second": 0.722, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.6490447526825438, | |
| "grad_norm": 11.871889943001767, | |
| "learning_rate": 1.647817538357072e-07, | |
| "logits/chosen": -1.2392758131027222, | |
| "logits/rejected": -1.3158223628997803, | |
| "logps/chosen": -731.3906860351562, | |
| "logps/rejected": -812.5341186523438, | |
| "loss": 0.4937, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -2.279770851135254, | |
| "rewards/margins": 1.1058552265167236, | |
| "rewards/rejected": -3.3856253623962402, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.6699816801884323, | |
| "grad_norm": 8.733519859905128, | |
| "learning_rate": 1.478143389201113e-07, | |
| "logits/chosen": -1.255499005317688, | |
| "logits/rejected": -1.310626745223999, | |
| "logps/chosen": -706.554443359375, | |
| "logps/rejected": -809.0241088867188, | |
| "loss": 0.4781, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": -2.4083428382873535, | |
| "rewards/margins": 0.9482523202896118, | |
| "rewards/rejected": -3.3565948009490967, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.6909186076943209, | |
| "grad_norm": 8.449035915191043, | |
| "learning_rate": 1.3139467229135998e-07, | |
| "logits/chosen": -1.2453768253326416, | |
| "logits/rejected": -1.2541320323944092, | |
| "logps/chosen": -725.0225830078125, | |
| "logps/rejected": -829.4249877929688, | |
| "loss": 0.4839, | |
| "rewards/accuracies": 0.7562500238418579, | |
| "rewards/chosen": -2.4690725803375244, | |
| "rewards/margins": 0.8669673800468445, | |
| "rewards/rejected": -3.336040496826172, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.7118555352002094, | |
| "grad_norm": 9.726561036240426, | |
| "learning_rate": 1.1561076868822755e-07, | |
| "logits/chosen": -1.1633121967315674, | |
| "logits/rejected": -1.2020900249481201, | |
| "logps/chosen": -733.0904541015625, | |
| "logps/rejected": -856.1812744140625, | |
| "loss": 0.499, | |
| "rewards/accuracies": 0.7749999761581421, | |
| "rewards/chosen": -2.551614284515381, | |
| "rewards/margins": 0.9220356941223145, | |
| "rewards/rejected": -3.4736499786376953, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.7327924627060979, | |
| "grad_norm": 8.334278257541456, | |
| "learning_rate": 1.0054723495346482e-07, | |
| "logits/chosen": -1.2510846853256226, | |
| "logits/rejected": -1.2921613454818726, | |
| "logps/chosen": -709.3406982421875, | |
| "logps/rejected": -806.6890869140625, | |
| "loss": 0.461, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": -2.4316649436950684, | |
| "rewards/margins": 0.9165409207344055, | |
| "rewards/rejected": -3.348206043243408, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.7537293902119864, | |
| "grad_norm": 8.447055903339363, | |
| "learning_rate": 8.628481651367875e-08, | |
| "logits/chosen": -1.2806235551834106, | |
| "logits/rejected": -1.3330801725387573, | |
| "logps/chosen": -759.4558715820312, | |
| "logps/rejected": -859.9752807617188, | |
| "loss": 0.5053, | |
| "rewards/accuracies": 0.800000011920929, | |
| "rewards/chosen": -2.384193181991577, | |
| "rewards/margins": 1.1400038003921509, | |
| "rewards/rejected": -3.5241971015930176, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.7746663177178749, | |
| "grad_norm": 7.277711887873546, | |
| "learning_rate": 7.289996455765748e-08, | |
| "logits/chosen": -1.254237174987793, | |
| "logits/rejected": -1.30390465259552, | |
| "logps/chosen": -721.6316528320312, | |
| "logps/rejected": -797.1998291015625, | |
| "loss": 0.4865, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": -2.433364152908325, | |
| "rewards/margins": 0.8884729146957397, | |
| "rewards/rejected": -3.3218369483947754, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.7956032452237635, | |
| "grad_norm": 8.0205493017487, | |
| "learning_rate": 6.046442623320145e-08, | |
| "logits/chosen": -1.2699052095413208, | |
| "logits/rejected": -1.3266394138336182, | |
| "logps/chosen": -743.0064086914062, | |
| "logps/rejected": -852.3271484375, | |
| "loss": 0.4851, | |
| "rewards/accuracies": 0.7562500238418579, | |
| "rewards/chosen": -2.516409397125244, | |
| "rewards/margins": 1.0128896236419678, | |
| "rewards/rejected": -3.529299259185791, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.816540172729652, | |
| "grad_norm": 8.82315393673875, | |
| "learning_rate": 4.904486005914027e-08, | |
| "logits/chosen": -1.2463610172271729, | |
| "logits/rejected": -1.2608873844146729, | |
| "logps/chosen": -727.2433471679688, | |
| "logps/rejected": -831.7511596679688, | |
| "loss": 0.4765, | |
| "rewards/accuracies": 0.731249988079071, | |
| "rewards/chosen": -2.392064332962036, | |
| "rewards/margins": 0.8088263273239136, | |
| "rewards/rejected": -3.200890302658081, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.8374771002355405, | |
| "grad_norm": 8.332606925483372, | |
| "learning_rate": 3.8702478614051345e-08, | |
| "logits/chosen": -1.2610491514205933, | |
| "logits/rejected": -1.3101046085357666, | |
| "logps/chosen": -675.9216918945312, | |
| "logps/rejected": -812.0501708984375, | |
| "loss": 0.4825, | |
| "rewards/accuracies": 0.8062499761581421, | |
| "rewards/chosen": -2.404404401779175, | |
| "rewards/margins": 1.064521074295044, | |
| "rewards/rejected": -3.4689254760742188, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8374771002355405, | |
| "eval_logits/chosen": -1.247671365737915, | |
| "eval_logits/rejected": -1.2644414901733398, | |
| "eval_logps/chosen": -734.7764892578125, | |
| "eval_logps/rejected": -828.874755859375, | |
| "eval_loss": 0.4924214482307434, | |
| "eval_rewards/accuracies": 0.7400000095367432, | |
| "eval_rewards/chosen": -2.4470431804656982, | |
| "eval_rewards/margins": 0.8939173817634583, | |
| "eval_rewards/rejected": -3.3409602642059326, | |
| "eval_runtime": 172.7999, | |
| "eval_samples_per_second": 11.574, | |
| "eval_steps_per_second": 0.723, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.8584140277414289, | |
| "grad_norm": 7.409919398422161, | |
| "learning_rate": 2.9492720416985e-08, | |
| "logits/chosen": -1.2109179496765137, | |
| "logits/rejected": -1.2863398790359497, | |
| "logps/chosen": -763.9716796875, | |
| "logps/rejected": -842.8446044921875, | |
| "loss": 0.4935, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": -2.5099594593048096, | |
| "rewards/margins": 0.8958646059036255, | |
| "rewards/rejected": -3.4058241844177246, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.8793509552473174, | |
| "grad_norm": 7.7696194448634275, | |
| "learning_rate": 2.1464952759020856e-08, | |
| "logits/chosen": -1.207766056060791, | |
| "logits/rejected": -1.2242529392242432, | |
| "logps/chosen": -737.0940551757812, | |
| "logps/rejected": -902.7199096679688, | |
| "loss": 0.4795, | |
| "rewards/accuracies": 0.7875000238418579, | |
| "rewards/chosen": -2.6334424018859863, | |
| "rewards/margins": 1.135379433631897, | |
| "rewards/rejected": -3.768822193145752, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.9002878827532059, | |
| "grad_norm": 8.567338429000005, | |
| "learning_rate": 1.4662207078575684e-08, | |
| "logits/chosen": -1.2356188297271729, | |
| "logits/rejected": -1.275846242904663, | |
| "logps/chosen": -750.3199462890625, | |
| "logps/rejected": -832.4161987304688, | |
| "loss": 0.4693, | |
| "rewards/accuracies": 0.78125, | |
| "rewards/chosen": -2.4712231159210205, | |
| "rewards/margins": 1.0153119564056396, | |
| "rewards/rejected": -3.486534833908081, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.9212248102590945, | |
| "grad_norm": 8.619770650174253, | |
| "learning_rate": 9.12094829893642e-09, | |
| "logits/chosen": -1.26433527469635, | |
| "logits/rejected": -1.3092998266220093, | |
| "logps/chosen": -719.7817993164062, | |
| "logps/rejected": -813.6593627929688, | |
| "loss": 0.4889, | |
| "rewards/accuracies": 0.768750011920929, | |
| "rewards/chosen": -2.478425979614258, | |
| "rewards/margins": 0.9797126054763794, | |
| "rewards/rejected": -3.4581382274627686, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.942161737764983, | |
| "grad_norm": 7.405850569844442, | |
| "learning_rate": 4.8708793644441086e-09, | |
| "logits/chosen": -1.1603410243988037, | |
| "logits/rejected": -1.2475802898406982, | |
| "logps/chosen": -723.2225341796875, | |
| "logps/rejected": -873.5631103515625, | |
| "loss": 0.4742, | |
| "rewards/accuracies": 0.800000011920929, | |
| "rewards/chosen": -2.4385993480682373, | |
| "rewards/margins": 1.1800200939178467, | |
| "rewards/rejected": -3.618619441986084, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.9630986652708715, | |
| "grad_norm": 8.784099959292528, | |
| "learning_rate": 1.9347820230782295e-09, | |
| "logits/chosen": -1.2129580974578857, | |
| "logits/rejected": -1.300582766532898, | |
| "logps/chosen": -732.4578857421875, | |
| "logps/rejected": -833.5997314453125, | |
| "loss": 0.4817, | |
| "rewards/accuracies": 0.78125, | |
| "rewards/chosen": -2.5728919506073, | |
| "rewards/margins": 1.036318063735962, | |
| "rewards/rejected": -3.609210252761841, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.98403559277676, | |
| "grad_norm": 9.493475272094026, | |
| "learning_rate": 3.2839470889836627e-10, | |
| "logits/chosen": -1.2292336225509644, | |
| "logits/rejected": -1.2678720951080322, | |
| "logps/chosen": -725.6412963867188, | |
| "logps/rejected": -856.3084716796875, | |
| "loss": 0.4615, | |
| "rewards/accuracies": 0.7437499761581421, | |
| "rewards/chosen": -2.491250514984131, | |
| "rewards/margins": 1.0373666286468506, | |
| "rewards/rejected": -3.5286173820495605, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.998691442030882, | |
| "step": 477, | |
| "total_flos": 0.0, | |
| "train_loss": 0.5192520058629898, | |
| "train_runtime": 13862.6582, | |
| "train_samples_per_second": 4.41, | |
| "train_steps_per_second": 0.034 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 477, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": false, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |